Can Jin

Ph.D. Candidate in Computer Science, Rutgers University

personal3.jpg
CBIM, Busch Campus [email protected]

I am a Ph.D. Candidate in Computer Science at Rutgers University, New Brunswick, advised by Professor Dimitris N. Metaxas. My research interests include Pre-training/Post-training/Inference of Large Foundation Models, Efficient AI, and AI Agents.

I earned my M.S. and B.S. degrees in Mathematics from the University of Science and Technology of China (USTC). Before my doctoral studies, I worked as a Machine Learning Engineer at Meituan Dianping Corporation, developing forecasting models for logistics and supply chain management. Recently, as a Research Intern at Adobe Research, I focused on the efficient pre-training of large foundation models (LLMs and DiTs) via dynamic routing Mixture-of-Experts (MoE).

I am open to collaboration on related projects. Please feel free to reach out via email if you share similar interests.

I am actively seeking research internship opportunities for Summer 2026, focusing on Pre-training/Post-training/Inference of Large Foundation Models, Efficient AI, and AI Agents. You can find my CV here.

Research

Large Foundation Models Pre-training, Post-training, and Inference
  • Focus: Enhancing the effectiveness of pre-training and post-training for generalization and reasoning capabilities.
  • Approach: Investigating advanced training techniques including Mixture-of-Experts, supervised fine-tuning, reinforcement learning, and inference-time scaling methods such as self-refinement and tree search.
  • Outcomes: NeurIPS 2024, WWW 2025 (RelWeb), arXiv 2025.
Efficient AI
  • Focus: Improving the efficiency of models’ training and inference.
  • Approach: Exploring methods such as model distillation, pruning, and prompting etc.
  • Outcomes: ICLR 2025, ICML 2025, NeurIPS 2025, AAAI 2025, and WWW 2025 (RelWeb).
AI Agents
  • Focus: Enhancing the performance of AI agents in multi-agent systems.
  • Approach: Utilizing reinforcement learning, supervised fine-tuning, and AI data generation pipelines.
  • Outcomes: NeurIPS 2025 (SEA), arXiv 2025.

Academic Services

Teaching Assistant
  • Rutgers University:
    • CS534: Computer Vision (Spring 2025)
    • CS210: Data Management for Data Science (Fall 2024)
    • CS211: Computer Architecture (Fall 2025)
Peer Review
  • Conference: NeurIPS 2025, CVPR 2025/2026, ICLR 2025, AAAI 2026, ICML 2024
  • Journal: Alexandria Engineering Journal, Information Fusion, Pattern Recognition, Signal Processing

selected publications

  1. Sparsity-Controllable Dynamic Top-p MoE for Large Foundation Model Pre-training
    Can Jin*Hongwu Peng*, Mingcan Xiang, Qixin Zhang, Xiangchi Yuan, Amit Hasan, Ohiremen Dibua, Yifan Gong, Yan Kang, and Dimitris N. Metaxas
    arXiv preprint arXiv:2512.13996, 2025
  2. Your reward function for RL is your best PRM for search: Unifying RL and search-based TTS
    arXiv preprint arXiv:2508.14313, 2025
  3. NeurIPS-SEA
    MAS-TTS.jpg
    Two Heads are Better Than One: Test-time Scaling of Multi-agent Collaborative Reasoning
    In Workshop on Scaling Environments for Agents, 2025
  4. Learning from Teaching Regularization: Generalizable Correlations Should be Easy to Imitate
    In Advances in Neural Information Processing Systems, 2024
  5. LoR-VP: Low-Rank Visual Prompting for Efficient Vision Model Adaptation
    Can Jin , Ying Li , Mingyu Zhao, Shiyu ZhaoZhenting WangXiaoxiao HeLigong HanTong Che, and Dimitris N. Metaxas
    In The Thirteenth International Conference on Learning Representations, 2025
  6. Visual Prompting Upgrades Neural Network Sparsification: A Data-Model Perspective
    Can Jin*Tianjin Huang* , Yihua Zhang, Mykola PechenizkiySijia LiuShiwei Liu, and Tianlong Chen
    In Proceedings of the AAAI Conference on Artificial Intelligence, 2025
  7. WWW
    APEER.png
    APEER: Automatic Prompt Engineering Enhances Large Language Model Reranking (Best Paper Award @ RelWeb)
    In Companion Proceedings of the ACM Web Conference 2025, Sydney, NSW, Australia, 2025
  8. WWW
    RankFlow.png
    RankFlow: A Multi-Role Collaborative Reranking Workflow Utilizing Large Language Models
    Can Jin*Hongwu Peng* , Anxiang Zhang , Nuo Chen , Jiahui Zhao, Xi Xie , Kuangzheng Li, Shuya Feng, Kai ZhongCaiwen Ding, and Dimitris N Metaxas
    In Companion Proceedings of the ACM Web Conference 2025, Sydney, NSW, Australia, 2025
  9. ACM MM
    Graph3D.jpg
    Graph Canvas for Controllable 3D Scene Generation
    Libin Liu , Shen Chen, Sen Jia, Jingzhe Shi, Can Jin, Zongkai Wu, Jenq-Neng Hwang , and Lei Li
    In Proceedings of the 33rd ACM International Conference on Multimedia, Dublin, Ireland, 2025
  10. Effective Policy Learning for Multi-Agent Online Coordination Beyond Submodular Objectives
    Qixin Zhang, Yan Sun, Can Jin , Xikun ZHANG, Yao Shu , Puning Zhao, Li Shen, and Dacheng Tao
    In The Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025
  11. ICML
    SCG.jpg
    Multinoulli Extension: A Lossless Yet Effective Probabilistic Framework for Subset Selection over Partition Constraints
    Qixin Zhang , Wei Huang, Can Jin , Puning Zhao, Yao Shu, Li Shen, and Dacheng Tao
    In Forty-second International Conference on Machine Learning, 2025