Photo of Mengzhou Xia

Mengzhou Xia chinese name

Email: [email protected]

I am currently a Member of Technical Staff at OpenAI and an incoming Assistant Professor in the Machine Learning Department at Carnegie Mellon University.

I recently completed my Ph.D. in Computer Science at Princeton University as part of Princeton NLP and Princeton Language Intelligence (PLI), advised by Prof. Danqi Chen. Prior to this, I obtained my Master's degree at Carnegie Mellon University, where I was advised by Prof. Graham Neubig, and my Bachelor's degree in Data Science and Software Engineering from Fudan University in China.

My research focuses on developing algorithms that enable the training of capable language models with less computational resources. I am fascinated by simple, general, and scalable approaches. Specifically, my work spans across:

Please find me on Google Scholar, Semantic Scholar, Github, X, and here is my updated CV.

Selected Publications and Preprints

For a full list of publications, please refer to this page.

  • SimPO: Simple Preference Optimization with a Reference-Free Reward
    Yu Meng*, Mengzhou Xia*, Danqi Chen
    NeurIPS 2024; [arXiv] [Code]

  • CharXiv: Charting Gaps in Realistic Chart Understanding in Multimodal LLMs
    Zirui Wang, Mengzhou Xia, Luxi He, Howard Chen, Yitao Liu, Richard Zhu, Kaiqu Liang, Xindi Wu, Haotian Liu, Sadhika Malladi, Alexis Chevalier, Sanjeev Arora, Danqi Chen
    NeurIPS 2024 Datasets and Benchmarks Track; [arXiv] [Code] [Project Page]

  • LESS: Selecting Influential Data for Targeted Instruction Tuning
    Mengzhou Xia*, Sadhika Malladi*, Suchin Gururangan, Sanjeev Arora, Danqi Chen
    ICML 2024; [arXiv] [Code] [Blog]

  • What is in Your Safe Data? Identifying Benign Data that Breaks Safety
    Luxi He*, Mengzhou Xia*, Peter Henderson
    COLM 2024; DPFM Workshop@ICLR 2024 (Best Paper); [arXiv] [Code]

  • Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
    Mengzhou Xia, Tianyu Gao, Zhiyuan Zeng, Danqi Chen
    ICLR 2024; [arXiv] [Code] [Blog]