About me

I am a PhD student at Berkeley Artificial Intelligence Research Lab (BAIR), advised by Prof. Kurt Keutzer. Previously, I completed my undergrad in Tsinghua University. During this time, I was fortunate to be advised by Prof. Jianfei Chen and Prof. Jun Zhu in TSAIL, and Prof. Tuo Zhao in Georgia Tech.

Research Interests

My research interests spans a couple of topics in efficient machine learning, particularly efficient training and inference. A list of my interests and some related projects:

  • Diffusion Language Model
  • Speculative Decoding
  • Large Reasoning Model
  • Dynamic Sparse Training

Selected Publications

  • Residual Context Diffusion Language Models [arXiv] [Code] [Project page] [Models]

    Yuezhou Hu*, Harman Singh*, Monishwaran Maheswaran*, Haocheng Xi, Coleman Hooper, Jintao Zhang, Aditya Tomar, Michael W. Mahoney, Sewon Min, Mehrdad Farajtabar, Kurt Keutzer, Amir Gholami, Chenfeng Xu

    Preprint

  • Arbitrage: Efficient Reasoning via Advantage-Aware Speculation [arXiv] [Project page]

    Monishwaran Maheswaran*, Rishabh Tiwari*, Yuezhou Hu*, Kerem Dilmen, Coleman Hooper, Haocheng Xi, Nicholas Lee, Mehrdad Farajtabar, Michael Mahoney, Kurt Keutzer, Amir Gholami

    Preprint

  • AdaSPEC: Selective Knowledge Distillation for Efficient Speculative Decoders [arXiv] [OpenReview] [Project page]

    Yuezhou Hu*, Jiaxin Guo*, Xinyu Feng, Tuo Zhao

    Neural Information Processing Systems (NeurIPS), 2025 (Spotlight)

  • S-STE: Continuous Pruning Function for Efficient 2:4 Sparse Pre-training [arXiv] [OpenReview] [Project page]

    Yuezhou Hu, Jun Zhu, Jianfei Chen

    Neural Information Processing Systems (NeurIPS), 2024

  • Accelerating Transformer Pre-training with 2:4 Sparsity [arXiv] [OpenReview] [PDF] [Project page]

    Yuezhou Hu, Kang Zhao, Weiyu Huang, Jianfei Chen, Jun Zhu

    International Conference on Machine Learning (ICML), 2024

Contact

Feel free to contact me via email or ask me any questions at my Askbox.