I’m Kangwei Liu (刘康威), a Ph.D. candidate in Institute of Information Engineering, Chinese Academy of Sciences (IIE, CAS), under the tutelage of Prof. Xianfeng Zhao and Associate Prof. Yun Cao. My research interests include Audio-driven Talking Face Generation and Facial Animation.

I received my B.E. degree from College of Computer Science and Engineering, Northeastern University and was mentored by Prof. Peng Cao at NEU.

I am open for collaborations in research, especially in the fields of Audio-driven Talking Face Generation and Facial Animation. Please feel free to contact me.

🔥 News

  • 2025.03:  🎉🎉 Two first-author papers accepted by ICME 2025 (CCF B)!
  • 2025.03:  🎉🎉 One first-author paper accepted by IJCNN 2025 (CCF C)!
  • 2024.03:  🎉🎉 One first-author paper accepted by ICME 2024 (CCF B, oral)!
  • 2023.05:  🎉🎉 Won 2nd place in “天马杯” National College Science and Technology Innovation Competition (20,000 RMB Award).

📝 Publications

ICME 2024 Oral
ProDub Framework

ProDub: Progressive Growing of Facial Dubbing Networks for Enhanced Lip Sync and Fidelity

Kangwei Liu, Xiaowei Yi, Xianfeng Zhao

Demo Video

  • TL;DR: A progressive framework for facial dubbing that disentangles mouth shape and texture for better lip-sync, while enhancing visual quality through a lip-aware temporal network. Improves lip-sync by 10.7% and visual quality by 28.5% over SOTA methods.
ICME 2025
3D Animation Framework

Controllable Expressive 3D Facial Animation via Diffusion in a Unified Multimodal Space

Kangwei Liu, Junwu Liu, Xiaowei Yi, Jinlin Guo, Yun Cao

Project Page

  • TL;DR: A diffusion-based framework for 3D facial animation that unifies multiple control signals (text, audio, emotion labels) and generates diverse, temporally coherent facial expressions, achieving 21.6% improvement in emotion similarity while maintaining natural dynamics.
ICME 2025
DisentTalk Framework

DisentTalk: Cross-lingual Talking Face Generation via Semantic Disentangled Diffusion Model

Kangwei Liu, Junwu Liu, Yun Cao, Jinlin Guo, Xiaowei Yi

Project Page

  • TL;DR: A novel approach that decomposes 3DMM expression parameters into meaningful subspaces for fine-grained facial control, enabling cross-lingual talking face generation with superior lip synchronization, expression quality, and temporal consistency.

🎖 Honors and Awards

Graduate Studies

  • 2024: Outstanding Student Cadre of University of Chinese Academy of Sciences (UCAS)
  • 2023: 2nd Place in “天马杯” National College Science and Technology Innovation Competition (20,000 RMB Award)
  • 2022-2024: Outstanding Student of UCAS
  • 2021: UCAS Scholarship

Undergraduate Studies

  • 2020.12: China National Scholarship, Ministry of Education of the People’s Republic of China (Top 1.5%), NEU
  • 2020: Mayor’s Scholarship of ShenYang, China (Ranked 6 out of 4000+)
  • 2018-2020: Northeastern University First-Class Scholarship (Ranked 1/268, 3/268, 2/317)
  • 2018-2019: Northeastern University “Outstanding Student Model” (Top 1%)
  • National Second Prize in China Undergraduate Mathematical Contest
  • Second Prize in International Mathematical Modeling Contest (MCM)

📖 Education

  • 2021.09 - 2026.06 (expected), School of Cyberspace Security, University of Chinese Academy of Sciences, China.
    Ph.D. Candidate

  • 2017.09 - 2021.06, College of Computer Science and Engineering, Northeastern University, China.
    Bachelor of Engineering
    Ranking: 14/268

💬 Invited Talks & Teaching

  • 2024-2025: Delivered 9 science popularization lectures on facial deepfake technology to primary and secondary schools in Beijing (Zhongguancun No.1 Primary School, Beijing Foreign Studies University Affiliated School, CAS Qiyuan School, etc.)
  • 2024.07.18: Gave an oral presentation at ICME 2024 conference in Niagara
  • 2023.11: Presented deepfake content generation technology at the National Defense University Academic Forum
  • 2023.06: Presented deepfake content generation technology at the Military Science Academy Academic Forum