SMIIP Lab

Speech and Multimodal Intelligent Information Processing (SMIIP) Lab at the Chinese University of Hong Kong, Shenzhen

photo_mingli.jpg

Prof. Ming Li, Ph.D.

Chinese University of Hong Kong, Shenzhen

Shenzhen, Guangdong, China

email Google Scholar

Welcome to the SMIIP Lab!

Our research interests lie in the areas of intelligent speech processing as well as multimodal behavior signal analysis and interpretation.

Intelligent Speech Processing:

  • Multi-channel, Multi-speaker, Multi-lingual, Multi-modal, Multi-task, Online Speech Perception
  • Generalized Target Speaker Extraction in Complex Scenarios
  • Manipulate, Generate and Hide Speaker’s Timbre Attributes
  • Speech Generation, Watermarking and Anti-Spoofing Countermeasure

Multimodal Behavior Signal Analysis and Interpretation:

  • Audio-Visual Behavior Understanding in ASD Diagnose and Assessment
  • AI Agent for ASD Behavior Understanding and Intervention Planning
  • Audio-Visual Laryngoscopy Data Analysis

Spoken Dialog System:

  • Audio-Visual Emotion-aware Spoken Dialog System
  • AI Agent-Based Spoken Dialog System



Prof. Ming Li’s Bio:

Ming Li received his Ph.D. in Electrical Engineering from University of Southern California in 2013. He is currently a Full Professor in the School of Artificial Intelligence at the Chinese University of Hong Kong, Shenzhen. He is also an Adjunct Professor in the School of Artificial Intelligence at Wuhan University. He was a faculty member at Duke Kunshan University from 2018 to 2026, a research scholar in the Department of Electrical and Computer Engineering at Duke University from 2018 to 2023 and an adjunct professor in the Department of Electrical and Computer Engineering at Carnegie Mellon University from 2015 to 2018. His research interests are in the areas of audio, speech and language processing as well as multimodal behavior signal analysis and interpretation. He has published more than 200 papers and served as the member of IEEE speech and language technical committee, CCF speech dialogue and auditory processing technical committee, CAAI affective intelligence technical committee, APSIPA speech and language processing technical committee. He was an area chair or meta reviewer for many Interspeech, ICASSP, SLT, ASRU conferences. He is the technical program co-chair at Odyssey 2022 and ASRU 2023. He is an editorial member of IEEE Transactions on Audio, Speech and Language Processing and Computer Speech and Language. Works co-authored with his colleagues have won first prize awards at Interspeech Computational Paralinguistic Challenges 2011, 2012 and 2019, ASRU 2019 MGB-5 ADI Challenge, Interspeech 2020 and 2021 Fearless Steps Challenges, VoxSRC 2021, 2022 and 2023 Challenges, ICASSP 2022 M2MeT Challenge, IJCAI 2023 ADD challenge, ICME 2024 ChatCLR challenge, Interspeech 2025 MISP challenge, Interspeech 2024 and 2025 AVSE challenges. As a co-author, he has won the best paper award in DCOSS 2009 and ISCSLP 2014 as well as the best paper shortlist in Interspeech 2024. He received the IBM faculty award in 2016, the ISCA Computer Speech and Language 5-years best journal paper award in 2018 and the youth achievement award of outstanding scientific research achievements of Chinese higher education in 2020. He is a senior member of IEEE.

news

No news so far...