π§βπ I am a bachelor-straight-to-doctorate student at Academy for Engineering and Technology of Fudan University. As a part of the Cognition and Intelligent Technology Laboratory, I am advised by Prof. Lihua Zhang . Prior to this, I received the B.E. degree in Communication Engineering from the joint training of Yunnan University and the Chinese People's Armed Police (PAP), Kunming, China, in 2020.
π My research interests include multimodal learning, sentiment analysis/emotion recognition, causal inference/unbiased estimation, and driving perception/multi-agent collaborative perception in autonomous driving. I have published 20+ papers at the top international AI conferences such as NeurIPS, CVPR, ICCV, and ECCV.
π I won the National Scholarship (postgraduate student, Top 1%), National Scholarship (doctoral student, Top 1%), Huatai Securities Technology Scholarship (Top 1%), Excellent Student Cadre, Outstanding League Cadre, and Excellent Student (twice) from 2020 to 2023.
π My academic service experience includes serving as a PC member for AAAI, ICCV, CVPR, ECCV, ACM MM and as a journal reviewer for IEEE TPAMI, TIP, and TCSVT.
π₯ News
- 2023.10: ππ 1 paper accepted to IEEE SPL 2023
- 2023.09: ππ 1 paper accepted to NeurIPS 2023
- 2023.07: ππ 2 paper accepted to ICCV 2023
- 2023.07: ππ 1 paper accepted to ACM MM 2023
π Selected Publications
Equal contribution Corresponding author
Driving/Collaborative Perception in Autonomous Driving
How2comm: Communication-Efficient and Collaboration-Pragmatic Multi-Agent Perception
Dingkang Yang, Kun Yang, Yuzheng Wang, ..., Peng Zhai, Lihua Zhang
- Multi-agent collaborative perception has recently received widespread attention as an emerging application in driving scenarios. We propose How2comm, a collaborative perception framework that seeks a trade-off between perception performance and communication bandwidth.
What2comm: Towards Communication-efficient Collaborative Perception via Feature Decoupling
Kun Yang, Dingkang Yang, Jingyu Zhang, ...
- We propose What2comm, a communication-efficient multiagent collaborative perception framework. Our framework outperforms previous approaches on real-world and simulated datasets by addressing various collaboration interferences, including communication noises, transmission delay, and localization errors, in an end-to-end manner.
Dingkang Yang, Shuai Huang, Zhi Xu, ..., Lihua Zhang
- We propose an AssIstive Driving pErception dataset (AIDE) to facilitate further research on the vision-driven driver monitoring systems. AIDE captures rich information inside and outside the vehicle from several drivers in realistic driving conditions.
Spatio-Temporal Domain Awareness for Multi-Agent Collaborative Perception
Kun Yang, Dingkang Yang, Jingyu Zhang, ...
- Multi-agent collaborative perception as a potential application for vehicle-to-everything communication could significantly improve the perception performance of autonomous vehicles over single-agent perception. We propose SCOPE, a novel collaborative perception framework that aggregates the spatio-temporal awareness characteristics across on-road agents in an end-to-end manner.
Visual Emotion Recognition
Context De-confounded Emotion Recognition
Dingkang Yang, Zhaoyu Chen, Yuzheng Wang, ..., Lihua Zhang
- We are the first to investigate the adverse context bias of the datasets in the context-aware emotion recognition task from the causal inference perspective and identify that such bias is a confounder, which misleads the models to learn the spurious correlation. In this case, we propose a contextual causal intervention module based on the backdoor adjustment to de-confound the confounder and exploit the true causal effect for model training.
Emotion Recognition for Multiple Context Awareness
Dingkang Yang, Shuai Huang, Shunli Wang, ..., Lihua Zhang
Project | Supplementary | Data
- We present a context-aware emotion recognition framework that combines four complementary contexts.
Multimodal Learning and Sentiment Analysis
Towards Robust Multimodal Sentiment Analysis under Uncertain Signal Missing
Mingcheng Li, Dingkang Yang, Lihua Zhang
- We propose a robust multimodal missing signal framework to handle the problem of uncertain signal missing for the multimodal sentiment analysis task and can be generalized to other multimodal patterns.
Dingkang Yang, Yang Liu, Can Huang, ..., Peng Zhai, Lihua Zhang
- Inspired by the human perception paradigm, we propose a target and source modality co-reinforcement approach to achieve sufficient crossmodal interaction and fusion at different granularities.
Disentangled Representation Learning for Multimodal Emotion Recognition
Dingkang Yang, Shuai Huang, Haopeng Kuang, Yangtao Du, Lihua Zhang
- We propose a feature-disentangled multimodal emotion recognition method, which learns the common and private feature representations for each modality.
Dingkang Yang, Haopeng Kuang, Shuai Huang, Lihua Zhang
- We propose a multimodal fusion approach for learning modality-specific and modality-agnostic representations to refine multimodal representations and leverage the complementarity across different modalities.
Contextual and Cross-modal Interaction for Multi-modal Speech Emotion Recognition
Dingkang Yang, Shuai Huang, Yang Liu, Lihua Zhang
- We propose a multimodal speech emotion recognition method based on interaction awareness.