Research
I have broad research interests on human-centric vision, including modeling human motion, human-centric interactions and embodied intelligence.
|
Selected Publications & Projects
* denotes equal contribution
|
|
TeleOpBench: A Simulator-Centric Benchmark for Dual-Arm Dexterous Teleoperation
Hangyu Li*, Qin Zhao*, Haoran Xu, Xinyu Jiang, Qingwei Ben, Feiyu Jia, Haoyu Zhao, Liang Xu, Jia Zeng, Hanqing Wang, Bo Dai, Junting Dong, Jiangmiao Pang
arXiv, 2025
[Paper] [Project] [Code]
We introduce TeleOpBench for benchmarking dual-arm dexterous teleoperation, which integrates motion-capture, VR controllers, upper-body exoskeletons, and vision-only teleoperation pipelines within a single modular framework.
|
|
Perceiving and Acting in First-Person: A Dataset and Benchmark for Egocentric Human-Object-Human Interactions
Liang Xu, Chengqun Yang, Zili Lin, Fei Xu, Yifan Liu, Congsheng Xu, Yiyi Zhang, Jie Qin, Xingdong Sheng, Yunhui Liu, Xin Jin, Yichao Yan, Wenjun Zeng, Xiaokang Yang
ICCV, 2025
[Paper] [Project] [Code]
We collect the first large-scale human-object-human interaction dataset called InterVLA with diverse generalist interaction categories and egocentric perspectives.
|
|
HIMO: A New Benchmark for Full-Body Human Interacting with Multiple Objects
Xintao Lv*, Liang Xu*, Yichao Yan, Xin Jin, Congsheng Xu, Shuwen Wu, Yifan Liu, Lincheng Li, Mengxiao Bi, Wenjun Zeng, Xiaokang Yang
ECCV, 2024
[Paper] [Project] [Code]
We build a large-scale dataset of human interacting with multiple objects with 3.3K HOI sequences and 4.08M frames.
|
|
Inter-X: Towards Versatile Human-Human Interaction Analysis
Liang Xu, Xintao Lv, Yichao Yan, Xin Jin, Shuwen Wu, Congsheng Xu, Yifan Liu, Yizhou Zhou, Fengyun Rao, Xingdong Sheng, Yunhui Liu, Wenjun Zeng, Xiaokang Yang
CVPR, 2024
[Paper] [Project] [Code]
We propose Inter-X, a large-scale dataset of human-human interactions with 11K interaction sequences and more than 8.1M frames.
|
|
ActFormer: A GAN-based Transformer towards General Action-Conditioned 3D Human Motion Generation
Liang Xu*, Ziyang Song*, Dongliang Wang, Jing Su, Zhicheng Fang, Chenjing Ding, Weihao Gan, Yichao Yan, Xin Jin, Xiaokang Yang, Wenjun Zeng, Wei Wu
ICCV, 2023
[Paper] [Project] [Code]
We present a GAN-based Transformer for general action-conditioned 3D human motion generation, including single-person actions and multi-person interactive actions.
|
|
PaStaNet: Toward Human Activity Knowledge Engine
Yong-Lu Li, Liang Xu, Xinpeng Liu, Xijie Huang, Yue Xu, Shiyi Wang, Hao-Shu Fang, Ze Ma, Mingyang Chen, Cewu Lu
CVPR, 2020 TPAMI, 2023
[Paper] [Code] [Project]
We build a large-scale knowledge base PaStaNet, which contains 7M+ PaSta annotations. We infer PaStas first and then reason out the activities based on part-level semantics
|
Experience
Shanghai Jiao Tong University & Eastern Institute of Technology, Ningbo, Shanghai/Ningbo, China
PhD in Computer Science, Sept. 2022 - Present
WeChat, Tencent Inc., Beijing, China
Research Intern, Jan. 2023 - Feb. 2024
SenseTime Technology Development Co., Ltd., Shanghai, China
Computer Vision Researcher, Jun. 2021 - Sept. 2022
Microsoft Research Asia, Beijing, China
Research Intern at Intelligent Multimedia Group, Jul. 2020 - Feb. 2021
Shanghai Jiao Tong University, Shanghai, China
Master of Science and Technology in Computer Science, Sept. 2018 - Mar. 2021
Nanjing University, Nanjing, China
Bachelor of Science in Computer Science, Sept. 2014 - Jun. 2018
|
Service
Reviewer
- CVPR'23/24/25
- NeurIPS Datasets and Benchmarks Track'22/23/24/25
- ICCV'25
- ECCV'24
- ICLR'25
- WACV'23/24/25
- ACM Multimedia'25
- IEEE Transactions on Multimedia
- IJCV
|
|