About me

I am currently a 3rd-year Ph.D. student at RCAST, the University of Tokyo, advised by Prof. Tatsuya Harada. Prior to that, I received my M.Eng. from Peking University, where I was advised by Prof. Yasha Wang, and I completed my B.Eng. at Tongji University. For my full CV, please see here.
My research interests focus on 3D human-centric computer vision, specifically in human reconstruction and animation. My research goals are to enable everyone to easily create digital avatars for games and virtual worlds, and to improve machine understanding of humans.

News

✨ [2026.02] Two papers accepted to CVPR 2026!

📜 [2025.08] Starting my internship at Meta Reality Labs Pittsburgh!

📜 [2025.08] Two papers accepted to NeurIPS 2025!

📜 [2025.08] One paper accepted to SIGGRAPH Asia 2025!

📜 [2025.01] One paper accepted to CVPR 2025!

📜 [2024.09] One paper accepted to NeurIPS 2024!

📜 [2024.07] Starting my visiting studies at Princeton University!

📜 [2024.01] One paper accepted to ICLR 2024!

📜 [2023.04] Starting my PhD studies at the University of Tokyo!

📜 [2023.04] One paper accepted to ICCV 2023!

Work and Internship Experience

[2025.08 - 2026.01] Research Scientist Intern at Meta Reality Labs Pittsburgh.

[2025.05 - 2025.10] Student Researcher at Snap Research NYC (remote).

[2024.07 - 2024.09] Visiting Student Researcher at Princeton University, supervised by Prof. Jia Deng.

[2022.12 - 2023.11] Research Intern at International Digital Economy Academy (IDEA).

[2021.03 - 2022.10] Research Engineer at Tencent ARC Lab.

[2020.06 - 2021.02] Research Intern at Microsoft Research Asia.

[2019.01 - 2020.06] Research Intern at Megvii Technology, supervised by Dr. Xiangyu Zhang.

Research

a paper

DyaDiT: A Multi-Modal Diffusion Transformer for Socially-Aware Dyadic Gesture Generation

Yichen Peng, Jyun-Ting Song, Siyeol Jung, Ruofan Liu, Haiyang Liu, Xuangeng Chu, Ruicong Liu, Erwin Wu, Hideki Koike, Kris Kitani

CVPR 2026

DyaDiT generates digital human gestures by processing dyadic audio signals and social context, effectively capturing the mutual dynamics between two speakers.

a paper

I2-NeRF: Learning Neural Radiance Fields Under Physically-Grounded Media Interactions

Shuhong Liu, Lin Gu, Ziteng Cui, Xuangeng Chu, Tatsuya Harada

NeurIPS 2025

I2-NeRF is a novel neural radiance field framework that enhances isometric and isotropic metric perception under media degradation.

a paper

Intend to Move: A Dataset for Intention and Scene Aware Human Motion Prediction

Ryo Umagami, Liu Yue, Xuangeng Chu, Ryuto Fukushima, Tetsuya Narita, Yusuke Mukuta, Tomoyuki Takahata, Jianfei Yang, Tatsuya Harada

NeurIPS Datasets & Benchmarks Track 2025

Intend to Move is a new dataset for embodied AI, focusing on intention-aware long-term human motion in real-world environments.

a paper

ARTalk: Speech-Driven 3D Head Animation via Autoregressive Model

Xuangeng Chu, Nabarun Goswami, Ziteng Cui, Hanqin Wang, Tatsuya Harada

SIGGRAPH Asia 2025

ARTalk generates realistic 3D head motions (lip sync, blinking, expressions, head poses) from audio in real-time.

a paper

Luminance-GS: Adapting 3D Gaussian Splatting to Challenging Lighting Conditions with View-Adaptive Curve Adjustment

Ziteng Cui, Xuangeng Chu, Tatsuya Harada

CVPR 2025

A simple multi-view curve adjustment method for novel view synthesis under challenging lighting conditions, including low-light, overexposure, and varying exposure.

a paper

Generalizable and Animatable Gaussian Head Avatar

Xuangeng Chu, Tatsuya Harada

NeurIPS 2024

The first generalizable 3DGS framework for head avatars that can reconstruct realistic avatars from a single image and achieve real-time reenactment.

a paper

GPAvatar: Generalizable and Precise Head Avatar from Image(s)

Xuangeng Chu, Yu Li, Ailing Zeng, Tianyu Yang, Lijian Lin, Yunfei Liu, Tatsuya Harada

ICLR 2024

A framework to reconstructs 3D head avatars from one or several images in a single forward pass.

a paper

Real-time High-resolution View Synthesis of Complex Scenes with Explicit 3D Visibility Reasoning

Tiansong Zhou, Yebin Liu, Xuangeng Chu, Chengkun Cao, Changyin Zhou, Fei Yu, Yu Li

TVCG 2024

A real-time high-resolution novel view synthesis method from sparse view inputs.

a paper

Accurate 3D Face Reconstruction with Facial Component Tokens

Tianke Zhang, Xuangeng Chu, Yunfei Liu, Lijian Lin, Zhendong Yang, Zhengzhuo Xu, Chengkun Cao, Fei Yu, Changyin Zhou, Chun Yuan, Yu Li

ICCV 2023

A framework for 3D face reconstruction from monocular images based on transformers.

a paper

Detection in Crowded Scenes: One Proposal, Multiple Predictions

Xuangeng Chu*, Anlin Zheng*, Xiangyu Zhang*, Jian Sun

CVPR 2020 (Oral)

A nearly cost-free method to improve the detection performance in crowded scenes.