I am a final-year Ph.D. candidate in the Department of Electronic Engineering and Information Science, University of Science and Technology of China, supervised by Prof. Hongtao Xie. Before that, I graduated from the School of Artificial Intelligence, Xidian University with a bachelor’s degree. My research interest includes multimodal digital human synthesis, motion generation, and face-swapping.

πŸ”₯ News

  • 2025.06: Β πŸŽ‰πŸŽ‰ Two papers is accepted by ICCV 2025.
  • 2025.02: Β πŸŽ‰πŸŽ‰ Two papers is accepted by CVPR 2025.
  • 2025.01: Β πŸŽ‰πŸŽ‰ One paper is accepted by Computer Graphics Forum 2025.
  • 2024.09: Β πŸŽ‰πŸŽ‰ One paper is accepted by NeurIPS 2024.
  • 2024.09: Β πŸŽ‰πŸŽ‰ One paper is accepted by TMM 2024.
  • 2024.08: Β πŸŽ‰πŸŽ‰ One paper is accepted by CVIU 2024.
  • 2024.07: Β πŸŽ‰πŸŽ‰ One paper is accepted by SIGGRAPH Asia 2024.

πŸ“ Publications

ICCV 2025
sym

GestureHYDRA: Semantic Co-speech Gesture Synthesis via Hybrid Modality Diffusion Transformer and Cascaded-Synchronized Retrieval-Augmented Generation

Quanwei Yang*, Luying Huang*, Kaisiyuan Wang, Jiazhi Guan, Shengyi He, Fengguo Li, Lingyun Yu, Yingying Li, Haocheng Feng, Hang Zhou, Hongtao Xie.(*Equal contribution)

Project Page

NeurIPS 2024
sym

ShowMaker: Creating High-Fidelity 2D Human Video via Fine-Grained Diffusion Modeling

Quanwei Yang, Jiazhi Guan, Kaisiyuan Wang, Lingyun Yu, Wenqing Chu, Hang Zhou, ZhiQiang Feng, Haocheng Feng, Errui Ding, Jingdong Wang, Hongtao Xie.

Project Page

CVPR 2025
sym

Re-HOLD: Video Hand Object Interaction Reenactment via adaptive Layout-instructed Diffusion Model

Yingying Fan, Quanwei Yang, Kaisiyuan Wang, Hang Zhou, Yingying Li, Haocheng Feng, Errui Ding, Yu Wu, Jingdong Wang

Project Page

SIGGRAPH Asia 2024
sym

TALK-Act: Enhance Textural-Awareness for 2D Speaking Avatar Reenactment with Diffusion Model

Jiazhi Guan, Quanwei Yang, Kaisiyuan Wang, Hang Zhou, Shengyi He, Zhiliang Xu, Haocheng Feng, Errui Ding, Jingdong Wang, Hongtao Xie, Youjian Zhao, Ziwei Liu.

Project Page

  • [TMM 2024] High Fidelity Face Swapping via Facial Texture and Structure Consistency Mining. Fengyuan Liu, Lingyun Yu, Quanwei Yang, Meng Shao, Hongtao Xie.
  • [CGF 2025] THGS: Lifelike Talking Human Avatar Synthesis from Monocular Video via 3D Gaussian Splatting.Chuang Chen, Lingyun Yu, Quanwei Yang, Aihua Zheng, Hongtao Xie.
  • [ICCV 2025] Forensic-MoE: Exploring Comprehensive Synthetic Image Detection Traces with Mixture of Experts. Mingqi Fang, Ziguang Li, Lingyun Yu, Quanwei Yang, Hongtao Xie, Yongdong Zhang.
  • [CVPR 2025] AudCast: Audio-Driven Human Video Generation by Cascaded Diffusion Transformers. Jiazhi Guan, Kaisiyuan Wang, Zhiliang Xu, Quanwei Yang, Yasheng SUN, Shengyi He, Borong Liang, Yukang Cao, Yingying Li, Haocheng Feng, Errui Ding, Jingdong Wang, Youjian Zhao, Hang Zhou, Ziwei Liu. Project Page
  • [ACM MM 2023] High Fidelity Face Swapping via Semantics Disentanglement and Structure Enhancement. Fengyuan Liu, Lingyun Yu, Hongtao Xie, Chuanbin Liu, Zhiguo Ding, Quanwei Yang, Yongdong Zhang.

πŸŽ– Honors and Awards

  • 2025.07, Baidu Outstanding intern
  • 2022.09, Longhu scholarship of University of Science and Technology of China
  • 2020.07, Outstanding Graduate of Xidian University
  • 2019.08, Meritorious Winner in the Mathematical Contest in Modeling for College students, USA
  • 2018.10, CASC scholarship of China Aerospace Science and Technology Corporation

πŸ“– Educations

  • 2020.09 - 2026.06 (now), Ph.D., Information and Communication Engineering, University of Science and Technology of China.
  • 2016.08 - 2020.06, Bachelor’s degree, Intelligent Science and Technology. Xidian University,

πŸ’» Internships

  • 2023.11 - now, VIS, Baidu, China.
  • 2021.10 - 2023.04, JD Explore Academy, China.
  • 2020.07 - 2021.07, MMU, Kuaishou, China.