πŸ§‘β€πŸŽ“ I am a bachelor-straight-to-doctorate student at Academy for Engineering and Technology of Fudan University. As a part of the Cognition and Intelligent Technology Laboratory, I am advised by Prof. Lihua Zhang . Prior to this, I received the B.E. degree in Communication Engineering from the joint training of Yunnan University and the Chinese People's Armed Police (PAP), Kunming, China, in 2020.

πŸ“– My research interests include multimodal learning, sentiment analysis/emotion recognition, causal inference/unbiased estimation, and driving perception/multi-agent collaborative perception in autonomous driving. I have published 20+ papers at the top international AI conferences such as NeurIPS, CVPR, ICCV, and ECCV.

πŸŽ– I won the National Scholarship (postgraduate student, Top 1%), National Scholarship (doctoral student, Top 1%), Huatai Securities Technology Scholarship (Top 1%), Excellent Student Cadre, Outstanding League Cadre, and Excellent Student (twice) from 2020 to 2023.

πŸ”ˆ My academic service experience includes serving as a PC member for AAAI, ICCV, CVPR, ECCV, ACM MM and as a journal reviewer for IEEE TPAMI, TIP, and TCSVT.

πŸ”₯ News

  • 2023.10:  πŸŽ‰πŸŽ‰ 1 paper accepted to IEEE SPL 2023
  • 2023.09:  πŸŽ‰πŸŽ‰ 1 paper accepted to NeurIPS 2023
  • 2023.07:  πŸŽ‰πŸŽ‰ 2 paper accepted to ICCV 2023
  • 2023.07:  πŸŽ‰πŸŽ‰ 1 paper accepted to ACM MM 2023

πŸ“ Selected Publications

Equal contribution⋆ Corresponding author†

Driving/Collaborative Perception in Autonomous Driving

NuerIPS 2023
sym

How2comm: Communication-Efficient and Collaboration-Pragmatic Multi-Agent Perception

Dingkang Yang⋆, Kun Yang⋆, Yuzheng Wang, ..., Peng Zhai†, Lihua Zhang†

Project

  • Multi-agent collaborative perception has recently received widespread attention as an emerging application in driving scenarios. We propose How2comm, a collaborative perception framework that seeks a trade-off between perception performance and communication bandwidth.
ACM MM 2023
sym

What2comm: Towards Communication-efficient Collaborative Perception via Feature Decoupling

Kun Yang⋆, Dingkang Yang⋆, Jingyu Zhang, ...

  • We propose What2comm, a communication-efficient multiagent collaborative perception framework. Our framework outperforms previous approaches on real-world and simulated datasets by addressing various collaboration interferences, including communication noises, transmission delay, and localization errors, in an end-to-end manner.
ICCV 2023
sym

AIDE: A Vision-Driven Multi-View, Multi-Modal, Multi-Tasking Dataset for Assistive Driving Perception

Dingkang Yang, Shuai Huang, Zhi Xu, ..., Lihua Zhang†

Project | ArXiv

  • We propose an AssIstive Driving pErception dataset (AIDE) to facilitate further research on the vision-driven driver monitoring systems. AIDE captures rich information inside and outside the vehicle from several drivers in realistic driving conditions.
ICCV 2023
sym

Spatio-Temporal Domain Awareness for Multi-Agent Collaborative Perception

Kun Yang⋆, Dingkang Yang⋆, Jingyu Zhang, ...

Project | ArXiv

  • Multi-agent collaborative perception as a potential application for vehicle-to-everything communication could significantly improve the perception performance of autonomous vehicles over single-agent perception. We propose SCOPE, a novel collaborative perception framework that aggregates the spatio-temporal awareness characteristics across on-road agents in an end-to-end manner.

Visual Emotion Recognition

CVPR 2023
sym

Context De-confounded Emotion Recognition

Dingkang Yang, Zhaoyu Chen, Yuzheng Wang, ..., Lihua Zhang†

Project | ArXiv

  • We are the first to investigate the adverse context bias of the datasets in the context-aware emotion recognition task from the causal inference perspective and identify that such bias is a confounder, which misleads the models to learn the spurious correlation. In this case, we propose a contextual causal intervention module based on the backdoor adjustment to de-confound the confounder and exploit the true causal effect for model training.
ECCV 2022
sym

Emotion Recognition for Multiple Context Awareness

Dingkang Yang⋆, Shuai Huang⋆, Shunli Wang, ..., Lihua Zhang†

Project | Supplementary | Data

  • We present a context-aware emotion recognition framework that combines four complementary contexts.

Multimodal Learning and Sentiment Analysis

IEEE SPL 2023
sym

Towards Robust Multimodal Sentiment Analysis under Uncertain Signal Missing

Mingcheng Li, Dingkang Yang, Lihua Zhang†

  • We propose a robust multimodal missing signal framework to handle the problem of uncertain signal missing for the multimodal sentiment analysis task and can be generalized to other multimodal patterns.
KBS 2023
sym

Target and Source Modality Co-reinforcement for Emotion Understanding from Asynchronous Multimodal Sequences

Dingkang Yang, Yang Liu, Can Huang, ..., Peng Zhai†, Lihua Zhang†

  • Inspired by the human perception paradigm, we propose a target and source modality co-reinforcement approach to achieve sufficient crossmodal interaction and fusion at different granularities.
ACM MM 2022
sym

Disentangled Representation Learning for Multimodal Emotion Recognition

Dingkang Yang, Shuai Huang, Haopeng Kuang, Yangtao Du, Lihua Zhang†

  • We propose a feature-disentangled multimodal emotion recognition method, which learns the common and private feature representations for each modality.
ACM MM 2022
sym

Learning Modality-specific and -agnostic Representations for Asynchronous Multimodal Language Sequences

Dingkang Yang, Haopeng Kuang, Shuai Huang, Lihua Zhang†

  • We propose a multimodal fusion approach for learning modality-specific and modality-agnostic representations to refine multimodal representations and leverage the complementarity across different modalities.
IEEE SPL 2022
sym

Contextual and Cross-modal Interaction for Multi-modal Speech Emotion Recognition

Dingkang Yang, Shuai Huang, Yang Liu, Lihua Zhang†

  • We propose a multimodal speech emotion recognition method based on interaction awareness.