This series of Control & Pizza (Co-PI) seminars focuses on recent advances in optimization, control, and learning. Invited speakers will present their current research work. We will provide free pizza/food for attendees.
This is jointly organized by Prof. Sylvia Herbert (MAE), Prof. Jorge Poveda (ECE), Prof. Yuanyuan Shi (ECE), Prof. Behrouz Touri (ECE), and Prof. Yang Zheng (ECE).
For the Winter 2025, the seminars will be held biweekly on Wednesdays from 12:00 p.m. to 1:00 p.m. at FAH 4002. Check a full schedule here.
[TBD]
[2025.06.11]
Learning Neural Controllers with Optimality and Stability Guarantees Using Input-Output Dissipativity
Speaker: Keyan Miao [Slides]
Abstract: Deep learning methods have demonstrated significant potential for addressing complex nonlinear control problems. For real-world safety-critical tasks, however, it is crucial to provide formal stability guarantees for the designed controllers. In this talk, a new framework is presented for designing neural controllers that achieve both stability and optimality with respect to certain functions. The key idea is to exploit the concept of input-output dissipativity of nonlinear systems by learning neural storage functions and supply rate functions. As a generalization of Lyapunov theory, dissipativity theory provides a natural connection to optimal control theory, offering both stability guarantees and meaningful optimality certificates. The neural controllers can be directly derived from the learned supply rate functions and guarantee closed-loop stability while inheriting optimality properties that can be shaped towards user-defined control objectives. Extensive numerical experiments demonstrate the effectiveness of the proposed approach.
Short Bio: Keyan Miao is currently a doctoral candidate in the Control Group at the University of Oxford, supervised by Prof. Antonis Papachristodoulou and Dr. Konstantinos Gatsis, and supported by EPSRC and the Oxford-Ashton Memorial Graduate Scholarship. She earned her master’s degree from Imperial College London under the supervision of Prof. Richard Vinter. Keyan is currently a visiting student at ETH Zürich’s AI Center, advised by Prof. Andreas Krause’s group through the support of an NCCR Automation Fellowship. Her research interests lie at the intersection of learning for control and control for learning, particularly focusing on neural ordinary differential equations (Neural ODEs). Keyan seeks to develop innovative frameworks that effectively integrate machine learning techniques with control theory, aiming to enhance the performance and interpretability of control systems in practical scenarios. She has presented her work at leading international conferences and journals, including ICML and L4DC.
[2025.5.07]
Incentive Aligned and Robust Distributed Learning Methods
Abstract: Distributed and federated learning enables machine learning algorithms to be trained over decentralized edge devices without requiring the exchange of local datasets. We consider two scenarios in this talk. In the first scenario, we have mostly cooperative agents running distributed optimization methods. We analyze how the distribution of data affects agents' incentives to voluntarily participate and obediently follow traditional federated learning algorithms. We design a Faithful Federated Learning (FFL) mechanism based on FedAvg method and VCG mechanism which achieves (probably approximate) optimality, faithful implementation, voluntary participation, and balanced budget. We then analyze an alternative approach to align individual agent’s incentive to participate by allowing them to opt in or out. We propose a game theoretic framework and study the equilibrium properties with both rational and bounded rational agents. In the second scenario, we turn to a game theoretic formulation, where the agents may be under attack. We characterize the tradeoffs between convergence speed and robustness of learning dynamics.
Short Bio: Ermin Wei is an Associate Professor at the Electrical and Computer Engineering Department and Industrial Engineering and Management Sciences Department of Northwestern University. She completed her PhD studies in Electrical Engineering and Computer Science at MIT in 2014, advised by Professor Asu Ozdaglar, where she also obtained her M.S. She received her undergraduate triple degree in Computer Engineering, Finance and Mathematics with a minor in German, from University of Maryland, College Park. Her team won 2nd place in the GO-competition Challenge 1, an electricity grid optimization competition organized by the Department of Energy. Wei's research interests include distributed optimization methods, convex optimization and analysis, smart grid, communication systems and energy networks and market economic analysis.
[2025.4.23]
Pick-to-Learn for Systems and Control: Data-driven design with state-of-the art safety guarantees
Speaker: Dario Paccagnan [Slides]
Abstract: Data-driven methods have become powerful tools for tackling increasingly complex problems in Systems and Control. However, deploying these methods in real-world settings — especially safety-critical ones — requires rigorous safety and performance guarantees. This need has motivated much recent work at the interface of Statistical Learning and Control, aiming to integrate formal guarantees with data-driven design methods. However, many existing approaches achieve this only by sacrificing valuable data for testing/calibration or by restricting the design space, thus leading to suboptimal performances.
Against this backdrop, in this talk I will introduce Pick-to-Learn (P2L) for Systems and Control, a novel framework designed to equip any data-driven control method with state-of-the-art safety and performance guarantees. Crucially, P2L enables the use of all available data to jointly synthesize and certify the design, eliminating the need to set aside data for calibration or validation purposes. I will then demonstrate how, as a result, P2L delivers designs and certificates that outperforms existing methods across a range of core problems including optimal control, reachability analysis, safe synthesis, and robust control.
Short Bio: Dario Paccagnan is a Senior Lecturer (US Associate Professor) at the Department of Computing, Imperial College London where he joined in the Fall 2020. Before that, he was a postdoctoral fellow with the Center for Control, Dynamical Systems and Computation, University of California, Santa Barbara. He obtained his PhD from the Automatic Control Laboratory, ETH Zurich, Switzerland, in 2018. He received a B.Sc. and M.Sc. in Aerospace Engineering from the University of Padova, Italy, in 2011 and 2014, and a M.Sc. in Mathematical Modelling and Computation from the Technical University of Denmark in 2014; all with Honors. Dario's interests are at the interface of game theory, control theory, and learning theory with a focus on tackling societal-scale challenges. Dario was a finalist for the 2019 EECI best PhD thesis award and was recognized with the SNSF Early Postdoc Mobility Fellowship, the SNSF Doc Mobility Fellowship, and the ETH medal for his doctoral work.
[2025.2.26]
Convex Constrained Controller Synthesis for Evolution Equations
Speaker: Lauren Conger [Slides]
Abstract: We propose a convex controller synthesis framework for a large class of constrained linear systems, including those described by partial differential equations and integral equations, commonly used in fluid dynamics, thermo-mechanical systems, quantum control, or transportation networks. Most existing control techniques rely on a (finite-dimensional) discrete description of the system, via ordinary differential equations. Here, we work instead with more general (infinite-dimensional) Hilbert spaces. This enables the discretization to be applied after the optimization (optimize-then-discretize). Using output-feedback SLS, we formulate the controller synthesis as a convex optimization problem. Structural constraints like sensor and communication delays, and locality constraints, are incorporated while preserving convexity, allowing parallel implementation and extending key SLS properties to infinite dimensions. The proposed approach and its benefits are demonstrated on a linear Boltzmann equation.
Short Bio: Lauren Conger is a PhD candidate in Control and Dynamical Systems at Caltech. She is advised by Professors Eric Mazumdar, Franca Hoffmann and John Doyle. Her research is on control theory and analysis of game dynamics using PDEs. Before Caltech, Lauren earned her Bachelor's degree in Electrical and Computer Engineering with a physics minor from Cornell University in 2018.
[2025.2.19]
Optimization Algorithm Design via Electric Circuits
Speaker: Ernest Ryu [Slides]
Abstract: We present a novel methodology for convex optimization algorithm design using ideas from electric RLC circuits. Given an optimization problem, the first stage of the methodology is to design an appropriate electric circuit whose continuous-time dynamics converge to the solution of the optimization problem at hand. Then, the second stage is an automated, computer-assisted discretization of the continuous-time dynamics, yielding a provably convergent discrete-time algorithm. Our methodology recovers many classical (distributed) optimization algorithms and enables users to quickly design and explore a wide range of new algorithms with convergence guarantees.
Short Bio: Professor Ryu received a B.S. degree in Physics and Electrical engineering with honors at the California Institute of Technology in 2010 and an M.S. in Statistics and a Ph.D. in Computational and Mathematical Engineering with the Gene Golub Best Thesis Award at Stanford University in 2016. In 2016, he joined the Department of Mathematics at UCLA, as an Assistant Adjunct Professor. In 2020, he joined the Department of Mathematical Sciences at Seoul National University as a tenure-track faculty. In 2024, returned to UCLA as an assistant professor.
[2025.2.12]
Closing the loop between optimal nonlinear control and learning-based optimization
Speaker: Luca Furieri [Slides]
Abstract: The increasing complexity of modern engineering systems—ranging from maintaining power grid stability in the face of renewable energy fluctuations to ensuring safe and efficient automated traffic—demands new approaches to control and optimization. While traditional methods offer theoretical guarantees, they often struggle with scalability and adaptation to real-world uncertainties. Conversely, machine learning-based techniques achieve remarkable empirical performance but lack formal guarantees of stability and convergence. This talk introduces a recent unified approach towards (1) neural-network control with stability guarantees and (2) learning convergent algorithms for non-convex optimization. First, we introduce a parametrization of all and only those control policies that can stabilize a given time-varying nonlinear system. The main insight is that we can learn over a stable neural-network operator, in order to capture exclusively the stabilizing nonlinear control policies for a wide class of nonlinear systems — even under model uncertainty. Second, we turn to the problem of optimizing highly non-convex objective functions. While systems theory has provided optimal worst-case convergence rates for convex functions, a recent trend in machine learning named Learning to Optimize (L2O) uses neural networks to discover update rules that excel in non-convex scenarios - the catch being, formal convergence guarantees are generally not available. We bridge these two paradigms by developing an unconstrained parametrization of all convergent algorithms for smooth, non-convex functions. We showcase the developed methods on optimal control benchmarks inspired by collision avoidance problems and optimization benchmarks emerging in image classification.
Short Bio: Luca Furieri is a Principal Investigator at EPF Lausanne since 2023. He will join the University of Oxford as an Associate Professor in June 2025. His research focuses on optimal control and optimization for distributed decision-making and large-scale cyber-physical systems. Previously, he has been a Postdoctoral researcher at the Automatic Control Laboratory, EPFL. In 2020, he has been awarded a Ph.D. degree in Control and Optimization from ETH - Zurich. He has received the SNSF Ambizione career grant in 2022, the IEEE Transactions on Control of Network Systems Best Paper Award in 2022, and the American Control Conference O. Hugo Schuck Best Paper Award in 2018.
[2025.1.29]
On the level set theorems of Hamilton-Jacobi Reachability
Speaker: Dylan Hirsch [Slides]
Abstract: Hamilton-Jacobi Reachability (HJR) is an exciting framework for controlling safety-critical systems, such as aircraft, vehicles, and robots. To control such systems in a certifiably safe manner, even under uncertainty, HJR builds upon the theory of differential games using modern numerical capabilities and recent theoretical advancements. In this talk, I will give a rapid introduction to HJR, and then I will dive into detail regarding complications in two of the canonical results in this field. These results relate the games of kind on which HJR focuses to the games of degree which are central to the classical differential game literature. Unfortunately, there are seemingly simple scenarios in which both canonical results fail to hold. I will explain the origin of these technical difficulties, and then rigorously demonstrate how one can resolve these complications. To do so, I will leverage techniques from general topology and introduce mathematical concepts more broadly relevant to those interested in the theory behind HJR.
Short Bio: Dylan is a Ph.D. student in Mechanical and Aerospace Engineering at UCSD. He is interested in developing methods for safe control and reduced-order modeling, with applications in cyber-physical and biological systems. He received his B.S. in Biomedical Engineering from Johns Hopkins University and his S.M. in Biological Engineering from MIT. Through his master's research, he became excited about systems and control theory, inspiring his transition across fields. In his free time, Dylan enjoys running, hiking, and watching movies.
[2025.1.15]
Toward Resilient and Scalable Distributed Perception: Algorithms and Systems
Speaker: Yulun Tian [Slides]
Abstract: Collaborative perception, which enables multiple agents to construct globally consistent “mental models” of the environment from local noisy observations, forms the sensory foundation of multi-agent systems in many large-scale applications. However, achieving reliable collaborative perception in the real world is hard, due to computational challenges associated with the underlying optimization problems and operational constraints imposed by practical communication networks. To address these challenges, I will present a paradigm based on distributed Riemannian optimization and discuss how the proposed framework achieves theoretical guarantees such as certifiable optimality and convergence under asynchronous communication. Building on this algorithmic framework, I will present a fully distributed system for multi-agent metric-semantic mapping. I will demonstrate the system by showing results from recent large-scale experiments where a fleet of robots collaboratively map a university campus using only onboard sensors, computers, and communication. I will conclude the talk by returning to the big picture of collaborative perception and discuss open challenges for future research.
Short Bio: Dr. Yulun Tian is a postdoctoral researcher in the Existential Robotics Lab at UCSD and an incoming assistant professor at the University of Michigan. He received the Ph.D. degree in autonomous systems from MIT in 2023. His work applies tools from nonlinear and distributed optimization, machine learning, and graph theory to develop principled algorithms with theoretical guarantees and real-world systems for multi-agent perception and navigation. His research has received several awards, including the 2024 Best Dissertation Award from IEEE RAS TC for Multi-Robot Systems, and the 2022 King-Sun Fu Memorial Best Paper Award from the IEEE Transactions on Robotics.
[2024.12.06]
CDC Practice Talks
Talk 1: State-Augmented Linear Games with Antagonistic Error for High-Dimensional, Nonlinear Hamilton-Jacobi Reachability
Speaker: Will Sharpless [Slides]
Abstract: Hamilton-Jacobi Reachability (HJR) is a popular method for analyzing the liveness and safety of a dynamical system with bounded control and disturbance. The corresponding HJ value function offers a robust controller and characterizes the reachable sets, but is traditionally solved with Dynamic Programming (DP) and limited to systems of dimension less than six. Recently, the space-parallelizeable, generalized Hopf formula has been shown to also solve the HJ value with a nearly three-log increase in dimension limit, but is limited to linear systems. To extend this potential, we demonstrate how state-augmented (SA) spaces, which are well-known for their improved linearization accuracy, may be used to solve tighter, conservative approximations of the value function with any linear model in this SA space. Namely, we show that with a representation of the true dynamics in the SA space, a series of inequalities confirms that the value of a SA linear game with antagonistic error is a conservative envelope of the true value function. It follows that if the optimal controller for the HJ SA linear game with error may succeed, it will also succeed in the true system. Unlike previous methods, this result offers the ability to safely approximate reachable sets and their corresponding controllers with the Hopf formula in a non-convex manner. Finally, we demonstrate this in the slow manifold system for clarity, and in the controlled Van der Pol system with different lifting functions.
Talk 2: Filterless Fixed-Time Extremum Seeking for Scalar Quadratic Maps
Speaker: Michael Tang [Slides]
Abstract:
Talk 3: Inexact Augmented Lagrangian Methods for Conic Optimization: Quadratic Growth and Linear Convergence
Speaker: Feng-Yi Liao [Slides]
Abstract: Augmented Lagrangian Methods (ALMs) are widely employed in solving constrained optimizations, and some efficient solvers are developed based on this framework. Under the quadratic growth assumption, it is known that the dual iterates and the Karush–Kuhn–Tucker (KKT) residuals of ALMs applied to semidefinite programs (SDPs) converge linearly. In contrast, the convergence rate of the primal iterates has remained elusive. In this paper, we resolve this challenge by establishing new quadratic growth and error bound properties for primal and dual SDPs under the strict complementarity condition. Our main results reveal that both primal and dual iterates of the ALMs converge linearly contingent solely upon the assumption of strict complementarity and a bounded solution set. This finding provides a positive answer to an open question regarding the asymptotically linear convergence of the primal iterates of ALMs applied to semidefinite optimization.
Talk 4: Safe Returning FaSTrack With Robust Control Lyapunov-Value Functions
Speaker: Boyang Li [Slides]
Abstract: Real-time navigation in a priori unknown environment remains a challenging task, especially when an unexpected (unmodeled) disturbance occurs. In this letter, we propose the framework Safe Returning Fast and Safe Tracking (SR-F) that merges concepts from 1) Robust Control Lyapunov-Value Functions (R-CLVF) 1, and 2) the Fast and Safe Tracking (FaSTrack) framework 2. The SR-F computes an R-CLVF offline between a model of the true system and a simplified planning model. Online, a planning algorithm is used to generate a trajectory in the simplified planning space, and the R-CLVF is used to provide a tracking controller that exponentially stabilizes to the planning model. When an unexpected disturbance occurs, the proposed SR-F algorithm provides a means for the true system to recover to the planning model. We take advantage of this mechanism to induce an artificial disturbance by “jumping” the planning model in open environments, forcing faster navigation. Therefore, this algorithm can both reject unexpected true disturbances and accelerate navigation speed. We validate our framework using a 10D quadrotor system and show that SR-F is empirically 20% faster than the existing works while maintaining safety.
Talk 5: Stability-Constrained Learning for Frequency Regulation in Power Grids with Variable Inertia
Speaker: Jie Feng [Slides]
Abstract: The increasing penetration of converter-based renewable generation has resulted in faster frequency dynamics, and low and variable inertia. As a result, there is a need for frequency control methods that are able to stabilize a disturbance in the power system at timescales comparable to the fast converter dynamics. This letter proposes a combined linear and neural network controller for inverter-based primary frequency control that is stable at time-varying levels of inertia. We model the time-variance in inertia via a switched affine hybrid system model. We derive stability certificates for the proposed controller via a quadratic candidate Lyapunov function. We test the proposed control on a 12-bus 3-area test network, and compare its performance with a base case linear controller, optimized linear controller, and finite-horizon Linear Quadratic Regulator (LQR). Our proposed controller achieves faster mean settling time and over 50% reduction in average control cost across 100 inertia scenarios compared to the optimized linear controller. Unlike LQR which requires complete knowledge of the inertia trajectories and system dynamics over the entire control time horizon, our proposed controller is real-time tractable, and achieves comparable performance to LQR.
[2024.12.03]
Mitigating Bias in Decision-Making Systems: a Control Systems Perspective
Speaker: Giulia De Pasquale, [Slides]
Abstract: Prediction-based decision-making systems are becoming increasingly prevalent in various domains. Previous studies have demonstrated that such systems are vulnerable to runaway feedback loops which exacerbate existing biases. The automated decisions have dynamic feedback effects on the system itself. In this talk we will show how existence of feedback loops in the machine learning-based decision-making pipeline can perpetuate and reinforce machine learning biases and propose control strategies to counteract their undesired effects.
Short Bio: Giulia De Pasquale is currently a Postdoc at ETH Zürich, Switzerland, in Professor Florian Dorfler’s group. Starting from January 2025 she will join the Control Systems group at Eindhoven University of Technology as an Assistant Professor. She received her PhD in Control and Systems Engineering at the University of Padova, Italy, in 2023. She obtained both the Master Degree in Control Engineering and the Bachelor Degree in Information Engineering from the University of Padova, in 2019 and 2017 respectively. Her current research interests include modeling, analysis, and control of networked socio-technical systems.
[2024.11.20]
Controllability Metrics, Limitations and Algorithms for Complex Networks
Speaker: Xinran Zheng, [Slides]
Abstract: N/A
Short Bio: Xinran Zheng is a third-year PhD student in the ECE Department at UC San Diego working with Prof. Tara Javidi and Prof. Behrouz Touri. Her research interest lies in zeroth-order distributed optimization and its application in control, federated learning, etc. Prior to studying at UCSD, she received a BSc degree in Mathematics and Physics from Tsinghua University in 2020.
[2024.11.06]
Characterizing Tie Strength with An Algebraic Topological Stochastic Process
Speaker: Arnab Kumar Sarkar, [Slides]
Abstract: N/A
Short Bio: Arnab is a final stage PhD candidate in Social & Engineering Systems (SES) and Statistics at MIT, advised by Prof. Ali Jadbabaie. Prior to MIT, he graduated from the University of Pennsylvania with a B.S.E. in Networked and Social Systems Engineering and an M.S.E. in Data Science. His research uses tools from statistics, machine learning, and algebraic topology to uncover how micro-level interactions between people shape macro-level social and organizational outcomes.
[2024.10.23]
Policy optimization for mixed H2/H∞ control: benign nonconvexity and global optimality
Speaker: Chih-Fan (Rich) Pai, [Slides]
Abstract: Recent advances in reinforcement learning have renewed interest in the theoretical foundations of policy optimization (PO) for continuous control tasks. Despite the inherent nonconvexity of PO, classical optimal and robust control problems such as LQR, LQG, and H∞ control are known to exhibit favorable landscape properties, which are crucial for establishing global convergence. In this talk, we extend this insight to mixed H2/H∞ control, exploring its optimization landscape and revealing an underlying convex structure. We show that there exist no spurious stationary points, by utilizing a unified Extended Convex Lifting framework that closely connects with convex reformulations of the corresponding control problem via linear matrix inequalities. This analysis offers new perspectives on the benign nonconvexity of mixed H2/H∞ control and its implications for robust control synthesis.
Short Bio: Rich Pai is currently pursuing a PhD in Electrical and Computer Engineering at the University of California, San Diego. He holds an MS in Communication Engineering from the National Taiwan University and a BS in Electrical and Computer Engineering from the National Yang Ming Chiao Tung University. His research interest spans optimization theory and sequential decision-making under uncertainty, with applications in stochastic, robust, and nonstochastic control.
[2024.10.09]
Dynamic Gains for Shaping Transient Behavior in Hybrid Systems: A Time Deformation Approach
Speaker: Daniel E. Ochoa, [Slides]
Abstract: This presentation introduces a framework for modulating the transient behavior of nonlinear dynamical systems, including those with hybrid dynamics. Our approach interconnects the original system with an exogenous dynamic gain, generating continuous-time deformations of hybrid time domains. We establish sufficient conditions for preserving stability properties through these time deformations, enabling a spectrum of transient behaviors ranging from constant to prescribed-time scalings. The framework leverages tools from hybrid dynamical systems theory and formulates a bijective map relating solution sets between original and interconnected systems. We demonstrate the versatility of our approach through applications to gradient flow systems and momentum-based optimization techniques with resets, showcasing how it can be used to customize convergence rates for strictly convex objective functions.
[2024.06.27]
Hybrid Dynamical Seeking Systems: Model-Free Feedback Decision-Making and Control
Speaker: Jorge Poveda, [Slides]
Abstract: The convergence of physical and digital systems in modern engineering applications has inevitably led to closed-loop systems that exhibit both continuous-time and discrete-time dynamics. These closed-loop architectures are modeled as hybrid dynamical systems, prevalent across various technological domains, including robotics, power grids, transportation networks, and manufacturing systems. Unlike traditional “smooth” ordinary differential equations or discrete-time recursions, solutions to hybrid dynamical systems are generally discontinuous, lack uniqueness, and have convergence and stability properties that are defined with respect to complex sets. Therefore, effectively designing and controlling such systems, especially under disturbances and uncertainty, is crucial for the development of autonomous and efficient data-driven engineering systems capable of achieving adaptive and self-optimizing behaviors. In this talk, I will delve into recent advancements in the analysis and design of feedback controllers that can achieve such properties in complex scenarios via the synergistic use of adaptive “seeking” dynamics, robust hybrid control, and decision-making algorithms. These controllers can be systematically designed and analyzed using modern tools from hybrid dynamical systems theory, which facilitate the incorporation of "exploration” and “exploitation" behaviors within complex closed-loop systems via multi-time scale tools and perturbation theory. The proposed methodology leads to a family of provably stable and robust algorithms suitable for solving model-free feedback stabilization and decision-making problems in single-agent and multi-agent systems for which smooth feedback solutions fall short.
[2024.06.13]
American Control Conference Practice Talk
Talk 1: On Fixed-Time Stability for a Class of Singularly Perturbed Systems using Composite Lyapunov Functions
Speaker: Michael Tang [Slides]
Abstract: Fixed-time stable dynamical systems are capable of achieving exact convergence to an equilibrium point within a fixed time that is independent of the initial conditions of the system. This property makes them highly appealing for designing control, estimation, and optimization algorithms in applications with stringent performance requirements. However, the set of tools available for analyzing the interconnection of fixed-time stable systems is rather limited compared to their asymptotic counterparts. In this work, we address some of these limitations by exploiting the emergence of multiple time scales in nonlinear singularly perturbed dynamical systems, where the fast dynamics and the slow dynamics are fixed-time stable on their own. By extending the so-called composite Lyapunov method from asymptotic stability to the context of fixed-time stability, we provide a novel class of Lyapunov based sufficient conditions to certify fixed-time stability in a class of singularly perturbed dynamical systems.
Talk 2: Smoothing Mixed Traffic with Robust Data-driven Predictive Control for Connected and Autonomous Vehicles
Speaker: Xu Shang [Slides]
Abstract: Traffic waves, in the form of periodic acceleration and deceleration of individual vehicles, lead to significant reductions of travel efficiency and fuel economy. It has been widely demonstrated that connected and autonomous vehicles (CAVs) have great potential to mitigate this phenomenon. Since the near future will meet with a transition phase of mixed traffic where human-driven vehicles (HDVs) coexist with CAVs, the control of CAVs in mixed traffic has indeed attracted increasing attention. The recently developed DeeP-LCC (Data-EnablEd Predictive Leading Cruise Control) method has shown promising performance for data-driven predictive control of CAVs in mixed traffic. However, its simplistic zero assumption of the future velocity errors for the head vehicle may pose safety concerns and limit its performance of smoothing traffic flow. In this work, we propose a robust DeeP-LCC method to control CAVs in mixed traffic with enhanced safety performance. In particular, we first present a robust formulation that enforces a safety constraint for a range of potential velocity error trajectories, and then estimate all potential velocity errors based on the past data from the head vehicle. We also provide efficient computational approaches to solve the robust optimization for online predictive control. Nonlinear traffic simulations show that our robust DeeP-LCC can provide better traffic efficiency and stronger safety performance while requiring less offline data.
Talk 3: Moving-Horizon Estimators for Hyperbolic and Parabolic PDEs in 1-D
Speaker: Luke Bhan [Slides]
Abstract: Observers for PDEs are themselves PDEs. Therefore, producing real time estimates with such observers is computationally burdensome. For both finite-dimensional and ODE systems, moving-horizon estimators (MHE) are operators whose output is the state estimate, while their inputs are the initial state estimate at the beginning of the horizon as well as the measured output and input signals over the moving time horizon. In this paper we introduce MHEs for PDEs which remove the need for a numerical solution of an observer PDE in real time. We accomplish this using the PDE backstepping method which, for certain classes of both hyperbolic and parabolic PDEs, produces moving-horizon state estimates explicitly. Precisely, to explicitly produce the state estimates, we employ a backstepping transformation of a hard-to-solve observer PDE into a target observer PDE, which is explicitly solvable. The MHEs we propose are not new observer designs but simply the explicit MHE realizations, over a moving horizon of arbitrary length, of the existing backstepping observers. Our PDE MHEs lack the optimality of the MHEs that arose as duals of MPC, but they are given explicitly, even for PDEs. In the paper/talk, we provide explicit formulae for MHEs for both hyperbolic and parabolic PDEs, as well as simulation results that illustrate theoretically guaranteed convergence of the MHEs.
[2024.05.30]
Online Stochasitc Optimization with Decision-dependent Data: Uncertainty Quantification and convergence
Speaker: Abhishek Roy, [Slides]
Abstract: Online optimization algorithms like Stochastic Gradient Descent (SGD) are the main workhorse behind most machine learning tasks with huge datasets. Here we will concentrate on the case of decision-dependent data where the data-distribution is strategically or adversarially modified based on the outputs (decisions) of the algorithms. In this case, an equilibrium decision is of interest which remains optimal even after the data is strategically modified. While most works focus on the prediction accuracy of algorithms in this setting, I will present our new results on online statistical inference of algorithmic estimators for such equilibrium decisions. Specifically, we will focus on three research vignettes. In the first part, I will talk about online covariance estimation for SGD to construct a confidence interval for the equilibrium point under state (decision)-dependent Markovian data. To this end, we establish the convergence rate, which matches with the rate for i.i.d data ignoring logarithmic factors, of an online overlapping batch-means covariance estimator. The second part of my talk focuses on characterizing the asymptotic randomness of an algorithmic estimator of the saddle point in a stochastic min-max optimization with state-dependent Markovian data. These optimization problems arise in multitask strategic classification, relative profit maximization in competitive markets such as electric vehicle charging and ride-sharing platforms, and robust strategic regression. We show that the averaged iterate of the Stochastic Extra-gradient (SEG) Algorithm converges almost surely to the equilibrium saddle point of a globally convex-concave and locally strongly-convex strongly-concave objective, and is asymptotically normal. In the last part, I will talk about nonconvex optimization with decision-dependent data and a brief overview of my future research.
[2024.03.20]
Deceptive Nash Equilibrium Seeking in Noncooperative Games
Speaker: Michael Tang, [Slides]
Abstract: This study introduces a new class of Nash-seeking algorithms that incorporate deception. Deceptive mechanisms are strategically employed to mislead competing entities in a noncooperative environment, leading to the emergence of optimal control policies that account for the inherent adversarial nature of the interactions. We investigate the theoretical foundations of game-theoretic deception in the context of Nash seeking, providing a comprehensive analysis of the impact of our proposed deception scheme on the convergence properties, stability, and performance of the closed loop system. Simulation results and case studies show how the proposed approach can achieve improved objective values for the deceptive players. This work contributes to the broader field of autonomous systems and game theory, offering insights into the application of deception to enhance the robustness and efficiency of multi-agent adaptive systems.
Bio: Michael Tang is a first-year Ph.D. student in the ECE. His research interests include nonlinear control, game theory, and adaptive systems. He received his B.S. degree in Electrical Engineering from UCSD.
[2024.03.06]
A Differentiable PDE Approach for building control
Speaker: Yuexin Bian, [Slides]
Abstract: In this work, we present an innovative partial differential equation (PDE)-based learning and control framework for building HVAC control. The goal is to determine the optimal airflow supply rate and supply air temperature to minimize the energy consumption while maintaining a comfortable and healthy indoor environment. In the proposed framework, the dynamics of airflow, thermal dynamics, and air quality (measured by CO2 concentration) are modeled using PDEs. We formulate both the system learning and optimal HVAC control as PDE-constrained optimization, and we propose a gradient-descent approach based on the adjoint method to effectively learn the unknown PDE model parameters and optimize building control actions. We demonstrate that the proposed approach can accurately learn the building model on both synthetic and real-world datasets. Furthermore, the proposed approach can significantly reduce energy consumption while ensuring occupants’ comfort and safety constraints compared to traditional control methods such as maximum airflow policy, learning-based control with reinforcement learning, and optimization-based control with ODE models.
Bio: Yuexin Bian is a third-year Ph.D. student in the ECE department. Her research interests broadly lie in optimization and control, with applications in power systems. Before starting at UC San Diego, she completed her undergraduate degree in Electrical Engineering from Zhejiang University, China.
[2024.02.07]
Frameworks for High Dimensional Optimization
Speaker: Palma London, [The slides will be updated after the seminar]
Abstract: I present frameworks for solving extremely large, prohibitively massive optimization problems. Today, practical applications require optimization solvers to work at extreme scales, but existing solvers do not often scale as desired. I present black-box acceleration algorithms for speeding up optimization solvers, in both distributed and parallel settings. Given a huge problem, I develop dimension reduction techniques that allow the problem to be solved in a fraction of the original time, and simultaneously makes the computation amenable to distributed computation. Efficient, dependable and secure distributed computing is increasingly fundamental to a wide range of core applications including distributed data centers, decentralized power grid, coordination of autonomous devices, and scheduling and routing problems.
In particular, I consider two optimization settings of interest. First, I consider packing linear programming (LP). LP solvers are fundamental to many problems in supply chain management, routing, learning and inference problems. I present a framework that speeds up linear programming solvers such as Cplex and Gurobi by an order of magnitude, while maintaining provably nearly optimal solutions. Secondly, I present a distributed algorithm that achieves an exponential reduction in message complexity compared to existing distributed methods. I present both empirical demonstrations and theoretical guarantees on the quality of the solution and the speedup provided by my methods.
Bio: Dr. Palma London received her Ph.D. and M.Sc. in Computer Science at Caltech. She received her B.S.E.E. in Electrical Engineering and B.S. in Mathematics at the University of Washington. She is currently a postdoctoral researcher at UCSD. Her research broadly spans convex optimization, machine learning and distributed algorithms.
[2024.01.24]
Spikes in the training loss of SGD, catapults and feature learning
Abstract: We first present an explanation regarding the common occurrence of spikes in the training loss when neural networks are trained with stochastic gradient descent (SGD). We provide evidence that the spikes in the training loss of SGD are "catapults", an optimization phenomenon originally observed in GD with large learning rates in [Lewkowycz et al. 2020]. We empirically show that these catapults occur in a low-dimensional subspace spanned by the top eigenvectors of the tangent kernel, for both GD and SGD. Second, we posit an explanation for how catapults lead to better generalization by demonstrating that catapults promote feature learning by increasing alignment with the Average Gradient Outer Product (AGOP) of the true predictor. Furthermore, we demonstrate that a smaller batch size in SGD induces a larger number of catapults, thereby improving AGOP alignment and test performance.
Bio: Libin Zhu is a PhD candidate in computer science at UCSD, working with Misha Belkin. Prior to UCSD, he received his BS in mathematics from Zhejiang University. His research has been focused on the fundamental understanding of deep learning, e.g., the optimization and generalization of neural networks.
[2024.01.10]
Convex approximations of Data-enabled Predictive Control with Applications to Mixed Traffic
Abstract: Willems' fundamental lemma, which characterizes linear time-invariant (LTI) systems using input and output trajectories, has found many successful applications. Combining this with receding horizon control leads to a popular Data-EnablEd Predictive Control (DeePC) scheme. DeePC is first established for LTI systems and has been extended and applied for practical systems beyond LTI settings. However, the relationship between different DeePC variants, involving regularization and dimension reduction, remains unclear. In this talk, we will first introduce a new bi-level optimization formulation for DeePC which combines a data pre-processing step as an inner problem (system identification) and predictive control as an outer problem (online control). We will next discuss a series of convex approximations by relaxing some hard constraints in the bi-level optimization as suitable regularization terms, accounting for an implicit identification. These include some existing DeePC variants as well as two new variants, for which we establish their equivalence under appropriate settings. In the last part of this talk, we will present some remarkable empirical performances of an adapted method, called DeeP-LCC, in controlling connected and automated vehicles (CAVs) in the mixed traffic system. This talk is based on our recent work: https://arxiv.org/abs/2312.15431, and https://arxiv.org/abs/2310.00509.
Bio: Xu Shang is a first-year Ph.D. student, working with Prof. Yang Zheng in the SOC lab at UCSD. His research interests include learning, optimization, and control, with a particular focus on developing theoretical performance guarantees for utilizing data-driven control in nonlinear, stochastic systems and its integration with various learning algorithms.
Xu received his B.S. in Mechanical Engineering from Shanghai Jiao Tong University and his M.S. from the University of Michigan. While at the University of Michigan, he worked on bipedal robots, which inspired him to develop control theorems and algorithms to address real-world challenges. Away from work, Xu enjoys hiking, reading, and playing soccer.
Supplemental Material:
[2023.11.15]
Information as Control: The Role of Communication in Distributed Systems
Speaker: Bryce Ferguson
Abstract: Distributed decision-making has become an increasingly popular method of reducing physical and computational difficulties in large-scale engineered systems. However, the emergent system behavior induced by these local decisions need not be optimal. As a method to elicit greater coordination, we can design not just how system components act but also how they communicate. I will discuss several ways in which information-communication channels can be exploited as a method to control overall system behavior. Particularly, I will present a game-theoretic model for distributed decision-making and demonstrate the possible benefits and costs of increasing communication in terms of gain/loss to equilibrium efficiency and solution complexity.
Bio: Bryce Ferguson is a PhD candidate in the Department of Electrical and Computer Engineering at the University of California, Santa Barbara. He works under the supervision of Jason R. Marden in the Center for Control, Dynamical-Systems, and Computation (CCDC). Bryce received his B.S. and M.S. degrees in Electrical Engineering from the University of California, Santa Barbara in 2018 and 2020, respectively, and his A.A. in Mathematics from Santa Rosa Junior College in 2016. In 2022, he was named a Rising Star in the NSF Co-sponsored Cyber-Physical Systems Rising Stars Workshop.
[2023.11.01]
Transition to linearity and an optimization theory for wide neural networks
Speaker: Chaoyue Liu
Abstract: In this talk, I will discuss an interesting property of neural networks: transition to linearity. The property shows that neural networks simplify towards linear models, as network width increases to infinity. I will show how this mathematical property is deeply connected with the structures of neural networks. Moreover, I will show that this property regularizes the square loss function of neural networks and provides convergence guarantees for gradient descent and SGD.
Bio: Chaoyue Liu is currently a postdoc at Halıcıoğlu Data Science Institute (HDSI), UC San Diego, working with Dr. Misha Belkin. He obtained my Ph.D. degree in Computer Science from The Ohio State University. After that, he spent one year at Meta Platforms Inc., as a research scientist. He also holds B.S. and M.S. degrees in physics from Tsinghua University.
[2023.10.18]
Structure-preserving Learning of Reduced-order Models for Large-scale Dynamical Systems
Speaker: Harsh Sharma
Abstract: Data-driven reduced-order models of large-scale computational models play a key role in a variety of tasks ranging from control of soft robotics to the design of mechanical structures to climate modeling. However, a vast majority of data-driven reduced-order models are designed to minimize the overall error over the training data, which leads to models that violate the underlying physical principles and provide inaccurate predictions outside the training data regime. This talk will present a structure-preserving learning approach that embeds physics into the data-driven operator inference framework to ensure that the learned models preserve the underlying geometric structure. The first half of this talk will focus on the data-driven symplectic model reduction of Hamiltonian systems. In the second half, we will discuss structure-preserving model reduction of large-scale nonlinear mechanical models. We will also discuss an ML-enhanced operator inference framework for predictive modeling in applications where the underlying physics is not well understood. The advantages of structure preservation in data-driven reduced-order modeling will be illustrated by various examples, which include biomimetic control of a soft robot, PDE parameter estimation from noisy and sparse data, and predictive modeling of a jointed structure from experimental measurements.
Bio: Harsh Sharma is a Postdoctoral Scholar working with Boris Krämer in the Department of Mechanical and Aerospace Engineering at UC San Diego. At UCSD, he taught an undergraduate-level course last year and currently, he is teaching a graduate-level course in the MAE department. Prior to this appointment, he received his PhD in Aerospace Engineering and MS in Mathematics from Virginia Tech. His research focuses on the intersection between structure preservation, reduced-order modeling, and deep learning, with a specific emphasis on nonintrusive model reduction of large-scale dynamical systems.
[2023.10.04]
Koopman-Hopf Reachability Analysis
Speaker: Will Sharpless
Abstract: The Hopf formula for Hamilton-Jacobi reachability (HJR) analysis has been proposed to solve high-dimensional differential games, producing the set of initial states and corresponding controller required to reach (or avoid) a target despite bounded disturbances. As a space-parallelizable optimization problem, the Hopf formula avoids the curse of dimensionality that afflicts standard dynamic-programming HJR, but is restricted to linear time-varying systems and convex games.
To harness the Hopf formula and compute reachable sets for high-dimensional nonlinear systems, we pair the Hopf solution with Koopman theory for global linearization. By first lifting a nonlinear system to a linear space and then solving the Hopf formula, approximate reachable sets can be efficiently computed that are much more accurate than local linearizations.
Furthermore, we construct a Koopman-Hopf disturbance-rejecting controller, and test its ability to drive a 10-dimensional nonlinear glycolysis model. We find that it significantly out-competes expectation-minimizing and game-theoretic model predictive controllers with the same Koopman linearization in the presence of bounded stochastic disturbance. Finally, we conclude with a discussion of future work towards error bounds and guarantees on the linear Hopf solution.
Bio: Will is a Ph.D. student at UCSD working in the SAS lab on applications of HJR algorithms to high-dimensions. His interests revolve around control and optimization in nonlinear, stochastic systems for autonomous devices in robotics, medicine and economics. He is particularly captivated by the graphs underlying differential systems and how their topology influences stability and he sometimes enjoys employing learning methods in these spaces. As an undergraduate, Will studied applied math and biology at UC Berkeley, during which he discovered a fascination for the theory of nonlinear systems and control that arose in the metabolic networks and cellular ecology. When away from work, Will is likely running, reading, or listening to music.
[2023.09.20]
Reset Control for Power Systems
Speaker: Vishal Shenoy, [Slides]
Abstract: This work investigates achieving rapid frequency regulation in Virtual Power Plants using reset control, a type of hybrid control system introduced by J.C. Clegg in 1958. Reset controllers selectively reset certain elements of the control system, such as integrators, to dissipate energy and improve transient performance. Recent research has expanded on these techniques using passivity tools, and they have been studied in power systems to enhance microgrid performance. However, time-triggered resets in hybrid control require optimal tuning of the resetting frequency, which is challenging in systems with unknown and complex dynamics. To overcome this challenge, we explore state-triggered resets for VPPs based on system signals. Our findings are validated using the FlexPower model, which incorporates realistic and high-fidelity representations of wind turbines, photovoltaic generators, and batteries.
[2023.09.06]
Learning Koopman Eigenfunctions and Invariant Subspaces from Data
Speaker: Masih Haseli, [Slides]
Abstract: Koopman operator theory provides an alternative description of dynamical phenomena by encoding the evolution of functions under dynamics within a vector space. The linearity of the Koopman operator, combined with its spectral properties, especially its eigenfunctions and eigenvalues, lead to a regular algebraic structure that can be leveraged for data-driven identification and prediction. Given that the underlying vector space is often infinite-dimensional, it becomes necessary to find finite-dimensional descriptions of the operator’s action. This talk will describe our recent efforts to establish objective measures for assessing the quality of finite-dimensional Koopman-based models. Additionally, we will discuss designing efficient algebraic algorithms to identify or approximate finite-dimensional models with convergence guarantees and tunable accuracy.
This is a joint work with Prof. Jorge Cortes.
Bio: Masih Haseli received the B.Sc. and M.Sc. degrees in electrical engineering from the Amirkabir University of Technology (Tehran Polytechnic), Tehran, Iran, in 2013 and 2015, respectively. He also earned a Ph.D. degree in Engineering Sciences (Mechanical Engineering) from the University of California, San Diego, CA, USA, in 2022. He currently serves as a postdoctoral researcher in the Department of Mechanical and Aerospace Engineering at the University of California, San Diego, CA, USA. His research interests encompass system identification, nonlinear systems, network systems, and data-driven modeling and control.
Dr. Haseli received the Bronze Medal at the 2014 Iran National Mathematics Competition and was awarded the Best Student Paper Award at the 2021 American Control Conference.
[2023.06.07]
Structured Neural-PI Control with End-to-End Stability and Steady-State Optimality Guarantees
Abstract: This talk focuses on the control of networked systems (especially power systems) with the goal of optimizing both transient and steady-state performances. While neural network-based nonlinear controllers have shown superior performance in various applications, their lack of provable guarantees has restricted their adoption in high-stake real-world applications. We will start with the power system frequency control problem to show the construction of nonlinear proportional-integral (PI) controllers that guarantee stability and steady-state economic dispatch. Through a modular abstraction using equilibrium-independent passivity, we further generalize the structured-PI control to a range of networked systems. The key structure is the strict monotonicity on proportional and integral terms, which is parameterized as gradients of strictly convex neural networks (SCNN). In addition, the SCNNs serve as Lyapunov functions, giving us end-to-end performance guarantees. Experiments on power and traffic networks demonstrate the effectiveness of the proposed approach.
Bio: Wenqi Cui is a Ph.D. candidate in the Department of Electrical and Computer Engineering at the University of Washington, advised by Prof. Baosen Zhang. Previously, she received the B.Eng. degree and M.S. degree in electrical engineering from Southeast University and Zhejiang University in 2016 and 2019, respectively. Her research interests lie broadly in machine learning, control, and optimization for cyber-physical energy systems. She was selected to Rising Stars in EECS (2022) and Rising Stars in CPS (2023).
[2023.05.29]
Data-Driven Safety Quantification using Infinite-Dimensional Robust Convex Optimization
Speaker: Jared Miller, [Slides]
Abstract: Safety quantification attaches interpretable numbers to safe trajectories of dynamical systems. Examples of such quantifications include finding the minimum distance of closest approach to an unsafe set, or finding the minimum control effort required to crash into the unsafe set. A safe trajectory with a large distance of closest approach may be acceptable, but an agent that is informed of a small distance of closest approach may want to perform actuation to increase this distance. This work represents the distance and crash-safety problems as infinite-dimensional linear programs (LPs) in smooth auxiliary functions, based on existing work in optimal control, peak estimation, and peak-minimizing control. These LPs can be extended towards modifications in dynamics, such as the analysis of systems with adverse uncertainty processes. The infinite-dimensional LPs are solved using the moment Sum-of-Squares (SOS) hierarchy, which finds a convergent sequence of outer approximations (under compactness and regularity assumptions) to the true safety quantification task.
One particular form of dynamics is highlighted: systems with disturbance-affine dynamics in which the uncertain set is semidefinite-representable (SDR). This setting occurs in data-driven systems analysis, in which the dynamics are described by a linear combination of basis functions consistent with SDR noise descriptions (e.g., L-infinity, L2, energy bounds). Crash-safety can be interpreted in the data-driven framework as finding the minimum noise corruption in the observed data required for the system to contact the unsafe set. The bottleneck in the SOS program is the Lie derivative nonnegativity constraint posed over the time-state-disturbance set. We utilize an infinite-dimensional analogue of robust counterparts (from robust optimization) to eliminate the disturbance variables, forming tractable convex optimization problems for safety quantification without introducing conservatism. This robust counterpart technique can be extended anywhere the disturbance-affine + SDR structure is found, such as barrier functions, reachable set estimation, and optimal control.
Joint work with: Mario Sznaier (Northeastern University)
Bio: Jared Miller is a postdoctoral researcher at the Robust Systems Lab at Northeastern University, advised by Mario Sznaier. He received his B.S. and M.S. degrees in Electrical Engineering from Northeastern University in 2018 and his Ph.D. in Electrical Engineering from Northeastern University in 2023. He is a recipient of the 2020 Chateaubriand Fellowship from the Office for Science Technology of the Embassy of France in the United States. He was given an Outstanding Student Paper award at the IEEE Conference on Decision and Control in 2021 and in 2022. His current research topics include safety verification and data-driven control. His interests include large-scale convex optimization, nonlinear systems, semi-algebraic geometry, and measure theory.
[2023.05.24]
Zeroth-Order Non-Convex Optimization for Cooperative Multi-Agent Systems with Diminishing Step Size and Smoothing Radius
Speaker: Xinran Zheng, [Slides]
Abstract: In this work, we study a class of zeroth-order distributed optimization problems, where each agent can control a partial vector and observe a local cost that depends on the joint vector of all agents, and the agents can communicate with each other with time delay. We propose and study a gradient descent-based algorithm using two-point gradient estimators with diminishing smoothing parameters and diminishing step-size and we establish the convergence rate to a first-order stationary point for general nonconvex problems. A byproduct of our proposed method with diminishing step size and smoothing parameters, as opposed to the fixed-parameter scheme, is that our proposed algorithm does not require any information regarding the local cost functions. This makes the solution appealing in practice as it allows optimizing an unknown (black-box) global function. At the same time, the performance will adaptively match the problem instance parameters.
Bio: Xinran Zheng is a second-year PhD student in the ECE Department at UC San Diego working with Prof. Tara Javidi and Prof. Behrouz Touri. Her research interest lies in zeroth-order distributed optimization and its application in control, federated learning, etc. Prior to studying at UCSD, she received a BSc degree in Mathematics and Physics from Tsinghua University in 2020.
[2023.04.26]
A spectral bundle method for sparse semidefinite programming
Speaker: Hesam Mojtahedi, [Slides]
Abstract: Semidefinite programs (SDPs) have found a wide range of applications in the field of control. When solving SDPs, it is important to take advantage of the inherent sparsity in order to improve scalability. We present a new spectral bundle algorithm that solves sparse SDPs without introducing additional variables. Using chordal decomposition, we replace a large positive semidefinite (PSD) constraint with a set of smaller coupled constraints. We then move the PSD constraints into the cost function using the exact penalty method, leading to an equivalent non-smooth penalized problem. We present a new efficient spectral bundle algorithm, where subgradient information is incorporated to update a lower approximation function at each iteration. We further establish sublinear convergences based on objective value, primal feasibility, dual feasibility, and duality gap. In particular, under Slater’s condition, the algorithm converges with the rate of O(1/ε^3 ) and the rate improves to O (1/ε ) when strict complementarity holds. The theoretical analysis is supported by our numerical experiments.
[2023.04.12]
Neural operators for Provably Accelerated PDE Feedback Control
Abstract: In this presentation, we explore the first provable application of neural operators to feedback control. In particular, we focus on accelerating boundary control of PDEs via backstepping. The PDE backstepping design is a nonlinear infinite-dimensional mapping (operator) transforming the PDE with challenging system model functions (e.g., reaction or advection coefficients) into a known stable target system via the Volterra Operator. This mapping requires the solution of kernel gain functions in the form of a Goursat PDE which is numerically expensive to compute. To overcome this, we prove the existence of an arbitrarily accurate neural operator (DeepONet) approximator for the backstepping gain computation, which is trained offline, "once and for all," using a large enough sample set of PDE model functions. Then, we prove that controllers with the neural operator approximated gains are still stabilizing in a global exponential sense. We also extend this framework to approximate the full feedback law mapping, from plant parameter functions and state measurement functions to the control input, and achieve semiglobal practical stability. The elimination of real-time recomputation of gains is transformative for adaptive control of PDEs and gain scheduling control of nonlinear PDEs.
Bio: Luke is a Ph.D. student in the ECE department at UC San Diego working on the interaction of learning and control. His research interests lie in the areas of neural operators, learning-based control, and control of partial differential equations, all of which he is exploring under the guidance of his advisor, Prof Yuanyuan Shi, with assistance from Prof. Miroslav Krstic. Prior to this, he completed an accelerated Bachelors/Masters program in Physics and Computer Science at Vanderbilt University,
[2023.03.15]
Towards Rigorous Data-driven Control
Speaker: Ya-Chien Chang, [Slides]
Abstract: The progress of artificial intelligence and reinforcement learning is leading to the development of novel control techniques. These technologies utilize data and simulation to solve complex problems that traditional methods cannot handle. However, learning-based control methods' application in real-world scenarios is challenging due to the lack of safety guarantees. Lyapunov functions are a commonly used method for proving the stability of dynamical systems. Nonetheless, finding a Lyapunov function is not always possible as there is no general form for it, and it requires satisfying Lyapunov conditions that depend on the known dynamical systems.
During this presentation, I will discuss my research on the neural Lyapunov method. We introduce a general framework to construct Lyapunov functions and control policies simultaneously, simplifying the control Lyapunov function design process and achieving significantly larger attraction regions than current methods. Furthermore, I will elaborate on extending the neural Lyapunov method to enhance the stability of learned controllers for unknown systems. This method integrates neural Lyapunov techniques as an additional critic value into actor-critic reinforcement learning algorithms. Finally, I will demonstrate how this approach outperforms state-of-the-art policy optimization, highlighting its numerous advantages.
Bio: Ya-Chien is a fourth-year Ph.D. student in CSE at UC San Diego, advised by Prof. Sicun Gao. She received a master's degree in Applied Mathematics from National Tsing Hua University. Her research interests cover safe reinforcement learning, control theory, and optimization. Her current focus is on synthesizing stability certificates for data-driven control systems. Ya-Chien's work has been published in top conferences, and she has been honored with awards for excellence in research from the CSE department.
[2023.03.01]
Direct Policy Search for State-feedback Hinf Robust Control Synthesis
Speaker: Xingang Guo, [Slides]
Abstract: Direct policy search has been widely applied in modern reinforcement learning and continuous control. However, the theoretical properties of direct policy search on nonsmooth robust control synthesis have not been fully understood. The optimal $\mathcal{H}_\infty$ control framework aims at designing a policy to minimize the closed-loop $\mathcal{H}_\infty$ norm, and is arguably the most fundamental robust control paradigm. The primary focus of this talk is on the convergence results for the direct policy search methods on the state-feedback $\mathcal{H}_\infty$ design. We show that direct policy search is guaranteed to find the global solution of the robust $\mathcal{H}_\infty$ state-feedback control design problem despite the resultant policy optimization problem is nonconvex and nonsmooth. In particular, we show that for this nonsmooth optimization problem, all Clarke stationary points are global minimum. Next, we identify the coerciveness of the closed-loop $\mathcal{H}_\infty$ objective function, and prove that all the sublevel sets of the resultant policy search problem are compact. Based on these properties, we show that Goldstein's subgradient method and its implementable variants can be guaranteed to stay in the nonconvex feasible set and eventually find the global optimal solution of the $\mathcal{H}_\infty$ state-feedback synthesis problem. This work builds a new connection between nonconvex nonsmooth optimization theory and robust control, leading to an interesting global convergence result for direct policy search on optimal $\mathcal{H}_\infty$ synthesis.
Bio: Xingang Guo is a Ph.D. student at department Electrical and Computer Engineering (ECE) and Coordinate Science Laboratory (CSL) at the University of Illinois at Urbana-Champaign (UIUC), advised by Prof. Bin Hu. His research interests include control, optimization, machine learning, and their intersections. Previously, he obtained an M.S. degree in Electrical and Computer Engineering (ECE) from King Abdullah University of Science and Technology (KAUST) in 2020, where his research focuses on the process control with applications in membrane based water systems.
[2023.03.01]
Informativity in data-driven control
Speaker: Jaap Eising, [Slides]
Abstract: For a myriad of reasons, data-driven analysis and control have recently attracted a lot of attention. Much of this attention has essentially been building on the well-developed field of system identification, taking the following blueprint to controller design: First, determine the only or "best" model within a given class that is compatible with the collected measurements. Then apply a known, model-based method to obtain the required control performance. Recently, a different approach was introduced, namely one of data informativity, that is, reasoning about the information contained in the measurements directly. A few questions central to this viewpoint are: What conditions on the data are necessary and/or sufficient to guarantee the existence of a stabilizing controller? Can we perform control tasks without measuring or reconstructing the state? What can be deduced from samples of continuous signals? In this talk, we will answer these questions using classical techniques and methods from fields such as system identification and robust control.
Bio: Jaap Eising is a postdoctoral researcher at the Department of Mechanical and Aerospace Engineering at the University of California, San Diego. He attained the Ph.D. degree at the University of Groningen in 2021, after obtaining the master's degree in Mathematics at the same university in 2017. His research interests include constrained linear systems, systems described by difference/differential inclusions, data-driven control and geometric systems theory.