I am a final-year PhD student at MIT CSAIL advised by Professor Daniela Rus. My research focus is on sample-efficient learning for robot control based on concepts from model-based exploration, architecture simplification, and compositionality. Before joining MIT, I obtained a BSc in Mechanical Engineering and an MSc in Robotics and Control from ETH Zurich under supervision of Professor Marco Hutter. Throughout my studies, I had the opportunity of various research stays and internships in legged robotics as well as learning control, and was fortunate to learn from many brilliant mentors.
.
Education
MIT CSAIL
PhD student
09/2018-Present
ETH Zurich
MSc Robotics
09/2015-05/2018
University of Tokyo
Visiting student
03/2016-08/2016
ETH Zurich
BSc MechE
09/2011-08/2014
DeepMind
Research Scientist Intern
06/2022-09/2022
IHMC Robotics
Research Intern
03/2017-09/2017
PAL Robotics
Research Intern
04/2015-07/2015
DLR Robotics
Research Intern
10/2014-04/2015
Research
I'm interested in reinforcement learning, motion planning, and control. I'm particularly excited about application in (legged) robotics, leveraging learning-based approaches to unlock capabilities that prove challenging for traditional methods.
Summary: we combine the benefits of coarse exploration during learning and smooth control at convergence by growing action spaces via decoupled Q-learning.
Summary: we leverage a hierarchical policy design for multi-team dynamic flight control, enabling high-level strategic coordination via distributional decoupled RL.
Summary: we present Gigastep, a fully vectorized JAX-based MARL environment that features high-dimensional 3D dynamics, heterogeneous agent types, stochasticity, and partial observation.
Summary: we deploy a hierarchical RL agent in multi-team racing scenarios on scale-car hardware, combining decoupled SARSA for team-centric strategic reasoning with continuous PPO for ego-centric low-level decision making.
Solving Continuous Control via Q-learning Tim Seyde, Peter Werner, Wilko Schwarting, Igor Gilitschenski, Martin Riedmiller, Daniela Rus, Markus Wulfmeier
ICLR, 2023
Summary: we show that DQN combined with value decomposition and bang-bang action space discretization yields performance competitive with recent model-free and model-based actor critic algorithms when training from features or raw pixels.
Summary: we analyze how decision trees based on logic programs extracted from a compact bio-inspired model architecture can help interpretable decision making.
Summary: we show that several recent actor critic algorithms yield competitive performance when only considering bang-bang policy heads and discuss implications for agent and benchmark design.
Summary: we leverage a hierarchical model over diverse low-level policy architectures to transfer the burden of hyperparameter selection from the engineer to the agent.
Summary: we leverage a latent model ensemble to compute an upper confindence bound objective over predicted returns to guide exploration in continuous control from pixels.
Summary: we learn a value function that represents an upper confidence bound over expected returns to guide exploration in continuous control from features.
Summary: we solve a mixed-integer gait planning problem for a single-legged hopper by learning footstep selection with a Gaussian process, and using this to constrain a low-level trajectory planner.
Summary: we augment a capture-point based walking controller to account for swing-leg angular momentum during reference trajectory planning and show improved locomotion capabilities with an Atlas humanoid robot.
Summary: we implement a bio-inspired modular posture controller and compare to model-based control on disturbance compensations tasks with the TORO humanoid.