My research interests lie in Multimodal Machine Learning. The central goal of my research is to design scalable inference and learning algorithms to connect language, perception, and control for robust multimodal learning.
My current research lies in the foundations of multimodal learning with applications in multimedia, computer vision, natural language processing, healthcare, and embodied AI.
I received my Ph.D. from the Computer Science Department at School of Computer Science, Carnegie Mellon University. I was extremely fortunate to be advised by Prof. Lei Li and Prof. Christos Faloutsos.
Before that, I received my B.Eng. from Shanghai Jiao Tong University, advised by Prof. Bao-Liang Lu.
I've worked as a research intern at Google, Meta, Microsoft, Amazon Web Services, and Adobe. My research was generously supported by CMU CSD fellowships and fundings from DARPA, NSF, Adobe, Allegheny Health Network, and Cleveland Clinic.
News
[2025-04] We release Higgs-Audio, a powerful model for audio understanding and generation.
[2024-11] MMWatermark Robustness gets accepted by Journal of Data-centric Machine Learning Research (DMLR) 2024.
[2024-09] SnapNTell gets accepted by EMNLP 2024 Findings.
[2024-04] MMSum dataset gets accepted by CVPR 2024 as Poster Highlight (Top 11.9%). Check our MMSum dataset!
[2024-03] Embodied Policy Learning with Language-based Scene Summarization gets accepted by NAACL 2024.
[2024-01] MMRobustness gets accepted as the very first paper at Journal of Data-centric Machine Learning Research (DMLR) 2024. Check our MMRobustness benchmark!
[2023-11] One paper about Cardiovascular record retrieval gets accepted by PMLR ML4H 2023.
[2023-10] Start a research internship at Google.
[2023-10] One paper about human languages and brain signals gets accepted by EMNLP Findings 2023.
[2023-06] One paper accepted as spotlight by ICML 2023 Workshop on Interactive Learning with Implicit Human Feedback.
[2023-06] Two papers accepted by ICML 2023 Workshop on Machine Learning for Multimodal Healthcare Data.
[2023-05] Start a research internship at Meta.
[2023-05] One paper about multimodal summarization by Optimal Transport gets accepted by ACL Findings 2023.
[2023-04] One paper about data augmentation on Geodesics gets accepted by ICML 2023.
[2023-04] Invited talk at Microsoft Research Cambridge.
[2023-02] One paper accepted by CVPR 2023.
[2023-02] One paper accepted by ICASSP 2023.
[2023-01] Start a research internship at Microsoft.
[2023-01] One paper accepted by EACL Findings 2023.
[2023-01] One paper accepted by AISTATS 2023.
[2022-10] One paper accepted by WACV 2023.
[2022-10] One paper accepted by NeurIPS 2022 Workshop on Distribution Shifts.
[2022-10] Top Reviewers in NeurIPS 2022.
[2022-06] One paper accepted by MLHC 2022.
[2022-05] Start a research internship at AWS AI.
[2022-05] One paper accepted by ICML 2022 workshop on Principles of Distribution Shift.
[2022-04] One paper accepted by ICLR 2022 Workshop on Socially Responsible Machine Learning.
[2021-09] Receive a gift funding from Adobe. Thanks, Adobe!
[2021-05] Start a research internship at Adobe research.
Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition Wei Liu,
Jielin Qiu,
Wei-Long Zheng,
Bao-Liang Lu
IEEE Transactions on Cognitive and Developmental Systems 2021
[paper][code]
Visual Sequence Learning in Hierarchical Prediction Networks and Primate Visual Cortex Jielin Qiu,
Ge Huang,
Tai Sing Lee
NeurIPS 2019
[paper]
Journal Reviewer: Transactions on Pattern Analysis and Machine Intelligence (TPAMI), Transactions on Machine Learning Research (TMLR), Journal of Data-centric Machine Learning Research (DMLR), IEEE Transactions on Neural Networks and Learning Systems.