|
Research
I have broad research interests on human-centered vision, including modeling human motion and human-centric interactions for avatars, embodied intelligence, human behavior analysis and biomedical applications.
I'm always open to collaborations. If you're interested in working together, please feel free to reach out. 😊
|
|
Selected Publications & Projects
* denotes equal contribution
|
|
Perceiving and Acting in First-Person: A Dataset and Benchmark for Egocentric Human-Object-Human Interactions
Liang Xu, Chengqun Yang, Zili Lin, Fei Xu, Yifan Liu, Congsheng Xu, Yiyi Zhang, Jie Qin, Xingdong Sheng, Yunhui Liu, Xin Jin, Yichao Yan, Wenjun Zeng, Xiaokang Yang
ICCV, 2025
[Paper] [Project] [Code (Coming Soon)]
We aim to explore a unified first-person framework for modeling and benchmarking human–human, human–object, and human–scene interactions within a vision-language-action framework.
|
|
HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects
Xintao Lv*, Liang Xu*, Yichao Yan, Xin Jin, Congsheng Xu, Shuwen Wu, Yifan Liu, Lincheng Li, Mengxiao Bi, Wenjun Zeng, Xiaokang Yang
ECCV, 2024
[Paper] [Project] [Code]
We investigate the data, paradigms, models, and evaluation protocols for humans interacting with multiple objects.
|
|
Inter-X: Towards Versatile Human-Human Interaction Analysis
Liang Xu, Xintao Lv, Yichao Yan, Xin Jin, Shuwen Wu, Congsheng Xu, Yifan Liu, Yizhou Zhou, Fengyun Rao, Xingdong Sheng, Yunhui Liu, Wenjun Zeng, Xiaokang Yang
CVPR, 2024
[Paper] [Project] [Code]
We propose Inter-X, a large-scale dataset of human-human interactions with 11K interaction sequences and more than 8.1M frames with a diverse set of downstream tasks to systematically evaluate human-human interaction capabilities.
|
|
ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation
Liang Xu*, Ziyang Song*, Dongliang Wang, Jing Su, Zhicheng Fang, Chenjing Ding, Weihao Gan, Yichao Yan, Xin Jin, Xiaokang Yang, Wenjun Zeng, Wei Wu
ICCV, 2023
[Paper] [Project] [Code]
We present a GAN-based Transformer for general action-conditioned 3D human motion generation, including single-person and multi-person interactive actions.
|
|
TeleOpBench: A Simulator-Centric Benchmark for Dual-Arm Dexterous Teleoperation
Hangyu Li*, Qin Zhao*, Haoran Xu, Xinyu Jiang, Qingwei Ben, Feiyu Jia, Haoyu Zhao, Liang Xu, Jia Zeng, Hanqing Wang, Bo Dai, Junting Dong, Jiangmiao Pang
arXiv, 2025
[Paper] [Project] [Code]
We introduce TeleOpBench for benchmarking dual-arm dexterous teleoperation, which integrates motion-capture, VR controllers, upper-body exoskeletons, and vision-only teleoperation pipelines within a single modular framework.
|
|
PaStaNet: Toward Human Activity Knowledge Engine
Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Shiyi Wang, Hao-Shu Fang, Ze Ma, Mingyang Chen, Cewu Lu
CVPR, 2020 TPAMI, 2023
[Paper] [Code] [Project]
We build a large-scale knowledge base PaStaNet with 7M+ PaSta annotations. We infer PaStas first and then reason out the activities based on part-level semantics.
|
|
Experience
Shanghai Jiao Tong University & Eastern Institute of Technology, Ningbo, Shanghai/Ningbo, China
PhD in Computer Science, Sept. 2022 - Present
Shanghai Artificial Intelligence Laboratory, Shanghai, China
Research Intern, Apr. 2025 - Sept. 2025
WeChat, Tencent Inc., Beijing, China
Research Intern, Jan. 2023 - Feb. 2024
SenseTime Technology Development Co., Ltd., Shanghai, China
Computer Vision Researcher, Jun. 2021 - Sept. 2022
Microsoft Research Asia, Beijing, China
Research Intern at Intelligent Multimedia Group, Jul. 2020 - Feb. 2021
Shanghai Jiao Tong University, Shanghai, China
Master of Science and Technology in Computer Science, Sept. 2018 - Mar. 2021
Nanjing University, Nanjing, China
Bachelor of Science in Computer Science, Sept. 2014 - Jun. 2018
|
|
Service
Reviewer: CVPR, ICCV, ECCV, NeurIPS, ICLR, ICML, SIGGRAPH, ACM MM, BMVC, IJCV, TMM, WACV
|
|