Sato Lab./Sugano Lab.
Sato Lab./Sugano Lab.
Y. Sato Lab.
Sugano Lab.
News
Publications
Contact
Datasets
Internal Wiki
English
日本語
Paper-Conference
Image Cropping under Design Constraints
Image cropping is essential in image editing for obtaining a compositionally enhanced image. In display media, image cropping is a …
Takumi Nishiyasu
,
Wataru Shimoda
,
Yoichi Sato
PDF
Cite
Code
Proposal-based Temporal Action Localization with Point-level Supervision
Point-level supervised temporal action localization (PTAL) aims at recognizing and localizing actions in untrimmed videos where only a …
Yuan Yin
,
Yifei Huang
,
Ryosuke Furuta
,
Yoichi Sato
PDF
Cite
DeCo : Decomposition and Reconstruction for Compositional Temporal Grounding via Coarse-to-Fine Contrastive Ranking
Understanding dense action in videos is a fundamental challenge towards the generalization of vision models. Several works show that …
Lijin Yang
,
Quan Kong
,
Hsuan-Kung Yang
,
Wadim Kehl
,
Yoichi Sato
,
Norimasa Kobori
PDF
Cite
FineBio: A Fine-Grained Video Dataset of Biological Experiments with Hierarchical Annotations
Takuma Yagi
,
Misaki Ohashi
,
Yifei Huang
,
Ryosuke Furuta
,
Shungo Adachi
,
Toutai Mitsuyama
,
Yoichi Sato
Cite
Structural Multiplane Image: Bridging Neural View Synthesis and 3D Reconstruction
The Multiplane Image (MPI), containing a set of fronto-parallel RGBA layers, is an effective and efficient representation for view …
Mingfang Zhang
,
Jinglu Wang
,
Xiao Li
,
Yifei Huang
,
Yoichi Sato
PDF
Cite
Technical Report for EgoTracks in Ego4D Challenge 2023
Mingfang Zhang
,
Yuan Yin
,
Yifei Huang
,
Yoichi Sato
Cite
Weakly Supervised Temporal Sentence Grounding With Uncertainty-Guided Self-Training
The task of weakly supervised temporal sentence grounding aims at finding the corresponding temporal moments of a language description …
Yifei Huang
,
Lijin Yang
,
Yoichi Sato
PDF
Cite
Fine-grained Affordance Annotation for Egocentric Hand-Object Interaction Videos
Object affordance is an important concept in hand-object interaction, providing information on action possibilities based on human …
Zecheng Yu
,
Yifei Huang
,
Ryosuke Furuta
,
Takuma Yagi
,
Yusuke Gotsu
,
Yoichi Sato
PDF
Cite
Learning Video-independent Eye Contact Segmentation from In-the-Wild Videos
Human eye contact is a form of non-verbal communication and can have a great influence on social behavior. Since the location and size …
Tianyi Wu
,
Yusuke Sugano
PDF
Cite
Code
Background Mixup Data Augmentation for Hand and Object-in-Contact Detection
Detecting the positions of human hands and objects-in-contact (hand-object detection) in each video frame is vital for understanding …
Koya Tango
,
Takehiko Ohkawa
,
Ryosuke Furuta
,
Yoichi Sato
PDF
Cite
«
»
Cite
×