Sato Lab./Sugano Lab.
Sato Lab./Sugano Lab.
Y. Sato Lab.
Sugano Lab.
News
Publications
Contact
Resources
Internal Wiki
English
日本語
Recent Publications
» List of All Publications
Generative Modeling of Shape-Dependent Self-Contact Human Poses
One can hardly model self-contact of human poses without considering underlying body shapes. For example, the pose of rubbing a belly …
Takehiko Ohkawa
,
Jihyun Lee
,
Shunsuke Saito
,
Jason Saragih
,
Fabian Prada
,
Yichen Xu
,
Shoou-I Yu
,
Ryosuke Furuta
,
Yoichi Sato
,
Takaaki Shiratori
PDF
Cite
Code
AssemblyHands-X: Modeling 3D Hand-Body Coordination for Understanding Bimanual Human Activities
Bimanual human activities inherently involve coordinated movements of both hands and body. However, the impact of this coordination in …
Tatsuro Banno
,
Takehiko Ohkawa
,
Ruicong Liu
,
Ryosuke Furuta
,
Yoichi Sato
PDF
Cite
DOI
Leveraging RGB Images for Pre-Training of Event-Based Hand Pose Estimation
This paper presents RPEP, the first pre-training method for event-based 3D hand pose estimation using labeled RGB images and unpaired, …
Ruicong Liu
,
Takehiko Ohkawa
,
Tze Ho Elden Tse
,
Mingfang Zhang
,
Angela Yao
,
Yoichi Sato
PDF
Cite
EgoInstruct: An Egocentric Video Dataset of Face-to-face Instructional Interactions with Multi-modal LLM Benchmarking
Analyzing instructional interactions between an instructor and a learner who are co-present in the same physical space is a critical …
Yuki Sakai
,
Ryosuke Furuta
,
Juichun Yen
,
Yoichi Sato
PDF
Cite
Affordance-Guided Diffusion Prior for 3D Hand Reconstruction
How can we reconstruct 3D hand poses when large portions of the hand are heavily occluded by itself or by objects? Humans often resolve …
Naru Suzuki
,
Takehiko Ohkawa
,
Tatsuro Banno
,
Jihyun Lee
,
Ryosuke Furuta
,
Yoichi Sato
PDF
Cite
DOI
Cross-View Correspondence Modeling for Joint Representation Learning Between Egocentric and Exocentric Videos
Joint analysis of human action videos from egocentric and exocentric views enables a more comprehensive understanding of human …
Zhehao Zhu
,
Yoichi Sato
PDF
Cite
Code
DOI
EgoExoBench: A Benchmark for First- and Third-person View Video Understanding in MLLMs
Transferring and integrating knowledge across first-person (egocentric) and third-person (exocentric) viewpoints is intrinsic to human …
Yuping He
,
Yifei Huang
,
Guo Chen
,
Baoqi Pei
,
Jilan Xu
,
Tong Lu
,
Jiangmiao Pang
PDF
Cite
Code
DOI
Unveiling Egocentric Reasoning with Spatio-Temporal CoT
Egocentric video reasoning focuses on the unseen, egocentric agent who shapes the scene, demanding inference of hidden intentions and …
Baoqi Pei
,
Yifei Huang
,
Jilan Xu
,
Yuping He
,
Guo Chen
,
Fei Wu
,
Yu Qiao
,
Jiangmiao Pang
PDF
Cite
Code
Vinci: A Real-time Smart Assistant based on Egocentric Vision-language Model for Portable Devices
We present Vinci, a vision-language system designed to provide real-time, comprehensive AI assistance on portable devices. At its core, …
Yifei Huang
,
Jilan Xu
,
Baoqi Pei
,
Yuping He
,
Guo Chen
,
Mingfang Zhang
,
Lijin Yang
,
Zheng Nie
,
Jinyao Liu
,
Guoshun Fan
,
Dechen Lin
,
Fang Fang
,
Kunpeng Li
,
Chang Yuan
,
Xinyuan Chen
,
Yaohui Wang
,
Yali Wang
,
Yu Qiao
,
Limin Wang
PDF
Cite
Code
DOI
Egocentric Action-aware Inertial Localization in Point Clouds with Vision-Language Guidance
This paper presents a novel inertial localization framework named Egocentric Action-aware Inertial Localization (EAIL), which leverages …
Mingfang Zhang
,
Ryo Yonetani
,
Yifei Huang
,
Liangyang Ouyang
,
Ruicong Liu
,
Yoichi Sato
PDF
Cite
Code
See all publications
Cite
×