About me
I’m co-organizing an open invited track at the IFAC world congress 2026 on learning interpretable control policies. Find details and submission code here!
I’m a postdoc in the Mesbah Lab at UC Berkeley. My works is driven by the need for built-in safety of the architectures and algorithms governing decision-making agents. In that vein, my recent work develops a complementary framework inspired by deep reinforcement learning and model predictive control. These two areas are often viewed as opposites, but really they have a common core in dynamic programming and Markov decision processes. Essentially, RL represents one branch that solves decision-making tasks through trial and error and function approximators, while MPC is another branch that is based on dynamics, constraints, and optimization. RL has proved to be incredibly versatile but is not necessarily safe by-design; MPC puts safety and robustness at the forefront, but it can be difficult to design toward high performance. A more unified perspective of these two areas would make RL more appealing for real-world applications while also making MPC more flexible and scalable under general learning algorithms. Please see my recent works on RL and MPC and don’t hesitate to get in touch.
Publications
Most of my papers are linked to arXiv or a DOI. You can also find my articles on my Google Scholar profile.
- Journal Papers
- A view on learning robust goal-conditioned value functions: Interplay between RL and MPC
arXiv. 2025.
[Link] [arXiv] - Stabilizing reinforcement learning control: A modular framework for optimizing over all stable behavior
Automatica. 2024.
[Link] [arXiv] - Machine learning techniques for industrial sensing and control: A survey and practical perspective
Control Engineering Practice. 2024.
[Link] [arXiv] - Automated deep reinforcement learning for real-time scheduling strategy of multi-energy system integrated with post-carbon and direct-air carbon captured system
Applied Energy. 2023.
[Link] [arXiv] - Meta-reinforcement learning for the tuning of PI controllers: An offline approach
Journal of Process Control. 2022.
[Link] [arXiv] - Deep reinforcement learning with shallow controllers: An experimental application to PID tuning
Control Engineering Practice. 2022.
[Link] [arXiv] - Toward self‐driving processes: A deep reinforcement learning approach to control
AIChE Journal. 2019.
[Link] [arXiv]
- A view on learning robust goal-conditioned value functions: Interplay between RL and MPC
- Conference Proceedings
- MPCritic: A plug-and-play MPC architecture for reinforcement learning
arXiv. 2025.
[Link] [arXiv] - Local-Global Learning of Interpretable Control Policies: The Interface between MPC and Reinforcement Learning
arXiv. 2025.
[Link] [arXiv] - Guiding reinforcement learning with incomplete system dynamics
IROS. 2024.
[arXiv] - Deep Hankel matrices with random elements
Learning for Dynamics & Control Conference. 2024.
[Link] [arXiv] - Reinforcement learning with partial parametric model knowledge
IFAC World Congress. 2023.
[Link] [arXiv] - A modular framework for stabilizing deep reinforcement learning control
IFAC World Congress. 2023.
[Link] [arXiv] - Meta-reinforcement learning for adaptive control of second order systems
AdCONIP. 2022.
[Link] [arXiv] - A meta-reinforcement learning approach to process control
AdCHEM. 2021.
[Link] [arXiv] | Keynote Paper - Reinforcement learning based design of linear fixed structure controllers
IFAC World Congress. 2020.
[Link] [arXiv] - Optimal PID and antiwindup control design as a reinforcement learning problem
IFAC World Congress. 2020.
[Link] [arXiv] - Almost Surely Stable Deep Dynamics
NeurIPS. 2020.
[Link] [arXiv] | NeurIPS Spotlight - Modern machine learning tools for monitoring and control of industrial processes: A survey
IFAC World Congress. 2020.
[Link] [arXiv]
- MPCritic: A plug-and-play MPC architecture for reinforcement learning
- Theses
