I'm interested in making AI systems easy to use for all humans --- can we communicate our objectives intuitively and can we train agents that are aligned with those objectives? My research lies in the intersection of reinforcement learning and human-AI interaction with a recent focus on Foundation Models. My research also has a heavy emphasis on human studies and learning from human data.
Our paper on Auto-Aligning Multiagent Incentives with Global Objectives received a Long Talk and was a finalist for Best Paper at the ALA Workshop at AAMAS!
[Jan 2023]
Our paper on Reward Design with LMs was accepted to ICLR 2023.
[Aug 2022]
I will be starting an internship at DeepMind!
Publications
For the most up-to-date list of publications, please see google scholar.
* indicates equal contribution and co-authorship.
Toward Grounded Social Reasoning Minae Kwon, Hengyuan Hu, Vivek Myers, Siddharth Karamcheti, Anca Dragan, Dorsa Sadigh
Preprint, 2023
Evaluating Human-Language Model Interaction
Mina Lee, Megha Srivastava, Amelia Hardy, John Thickstun, Esin Durmus, Ashwin Paranjape, Ines Gerard-Ursin, Xiang Lisa Li, Faisal Ladhak, Frieda Rong, Rose E. Wang, Minae Kwon, Joon Sung Park, Hancheng Cao, Tony Lee,Rishi Bommasani, Michael Bernstein, Percy Liang
arXiv, 2022
Continual Adaptation for Efficient Machine Communication
Robert D. Hawkins, Minae Kwon, Dorsa Sadigh, Noah D. Goodman
ICML Adaptive and Multitask Learning Workshop, 2019
Proceedings of the 24rd Conference on Computational Natural Language Learning (CoNLL), 2020
(Best Paper Award at ICML Adaptive and Multitask Learning Workshop)