Talk @ EPFL, May 31st, 2024
Some useful resources
Welcome to the seminar! Below you can find information about the seminar, links to github repositories, and the list of publications mentioned during the presentation.
Information
- Where: EPFL, Lausanne, ME C2 405
- When: May 31, 2024, from 11:00 to 12:00
- Speaker: Fabio Bonassi
Repositories
Selected references
2024
- ConferenceStructured state-space models are deep Wiener modelsFabio Bonassi, Carl Andersson, Per Mattsson, and Thomas B SchönIn 20th IFAC Symposium on System Identification (SYSID), 2024
The goal of this paper is to provide a system identification-friendly introduction to the Structured State-space Models (SSMs). These models have become recently popular in the machine learning community since, owing to their parallelizability, they can be efficiently and scalably trained to tackle extremely-long sequence classification and regression problems. Interestingly, SSMs appear as an effective way to learn deep Wiener models, which allows to reframe SSMs as an extension of a model class commonly used in system identification. In order to stimulate a fruitful exchange of ideas between the machine learning and system identification communities, we deem it useful to summarize the recent contributions on the topic in a structured and accessible form. At last, we highlight future research directions for which this community could provide impactful contributions.
@inproceedings{bonassi2024structured, title = {Structured state-space models are deep Wiener models}, year = {2024}, booktitle = {20th IFAC Symposium on System Identification (SYSID)}, author = {Bonassi, Fabio and Andersson, Carl and Mattsson, Per and Sch{\"o}n, Thomas B}, url = {https://www.sciencedirect.com/science/article/pii/S2405896324013168}, doi = {10.1016/j.ifacol.2024.08.536}, }
- JournalNonlinear MPC design for incrementally ISS systems with application to GRU networksFabio Bonassi, Alessio La Bella, Marcello Farina, and Riccardo ScattoliniAutomatica, 2024
This brief addresses the design of a Nonlinear Model Predictive Control (NMPC) strategy for exponentially incremental Input-to-State Stable (ISS) systems. In particular, a novel formulation is devised, which does not necessitate the onerous computation of terminal ingredients, but rather relies on the explicit definition of a minimum prediction horizon ensuring closed-loop stability. The designed methodology is particularly suited for the control of systems learned by Recurrent Neural Networks (RNNs), which are known for their enhanced modeling capabilities and for which the incremental ISS properties can be studied thanks to simple algebraic conditions. The approach is applied to Gated Recurrent Unit (GRU) networks, providing also a method for the design of a tailored state observer with convergence guarantees. The resulting control architecture is tested on a benchmark system, demonstrating its good control performances and efficient applicability.
@article{bonassi2024nonlinear, title = {Nonlinear MPC design for incrementally ISS systems with application to GRU networks}, author = {Bonassi, Fabio and {La Bella}, Alessio and Farina, Marcello and Scattolini, Riccardo}, journal = {Automatica}, publisher = {Elsevier}, volume = {159}, pages = {111381}, year = {2024}, issn = {0005-1098}, doi = {10.1016/j.automatica.2023.111381}, url = {https://www.sciencedirect.com/science/article/pii/S0005109823005484} }
- JournalLearning Control Affine Neural NARX Models for Internal Model Control DesignJing Xie, Fabio Bonassi, and Riccardo ScattoliniIEEE Transactions on Automation Science and Engineering, 2024
This paper explores the use of Control Affine Neural Nonlinear AutoRegressive eXogenous (CA-NNARX) models for nonlinear system identification and model-based control design. The idea behind this architecture is to match the known control-affine structure of the system to achieve improved performance. Coherently with recent literature of neural networks for data-driven control, we first analyze the stability properties of CA-NNARX models, devising sufficient conditions for their incremental Input-to-State Stability (δISS) that can be enforced at the model training stage. The model’s stability property is then leveraged to design a stable Internal Model Control (IMC) architecture. The proposed control scheme is tested on a real Quadruple Tank benchmark system to address the output reference tracking problem. The results achieved show that (i) the modeling accuracy of CA-NNARX is superior to the one of a standard NNARX model for given weight size and training epochs, (ii) the proposed IMC law provides performance comparable to the ones of a standard Model Predictive Controller (MPC) at a significantly lower computational burden, and (iii) the δISS of the model is beneficial to the closed-loop performance. Note to Practitioners —Many engineering systems, such as robotic manipulators and chemical reactors, are described by Control Affine (CA) models, characterized by onlinear dynamics where the control variable enters in a linear way. If only this structural information is available without any additional knowledge, for instance on the order of the system or on the value of its parameters, a black-box identification approach can be followed to estimate the model from data. For these reasons, in this paper we propose a modeling and control design method suited for this class of systems. Specifically, we assume that the system is described by a CA-Neural Nonlinear AutoRegressive eXogenous (CA-NNARX) model. Then, the estimated model is used to design a stable Internal Model Control (IMC) scheme for the solution of output reference tracking problems. The stability, performance, and robustness properties of the proposed approach are studied and tested in the control of a laboratory system. In addition, a simulation analysis shows how IMC represents a valid alternative to the popular Model Predictive Control (MPC) approach, in particular for embedded systems, where the computation power required by MPC can be too high.
@article{xie2022robust, title = {Learning Control Affine Neural NARX Models for Internal Model Control Design}, author = {Xie, Jing and Bonassi, Fabio and Scattolini, Riccardo}, journal = {IEEE Transactions on Automation Science and Engineering}, doi = {10.1109/TASE.2024.3479321}, year = {2024}, }
2023
- ThesisReconciling deep learning and control theory: recurrent neural networks for model-based control designFabio BonassiFeb 2023
Dimitris N. Chorafas Ph.D. Award from the Dimitris N. Chorafas Foundation (Switzerland) to the best Ph.D. theses for their high potential for practical application and the special significance attached to their aftermath
This doctoral thesis aims to establish a theoretically-sound framework for the adoption of Recurrent Neural Network (RNN) models in the context of nonlinear system identification and model-based control design. The idea, long advocated by practitioners, of exploiting the remarkable modeling performances of RNNs to learn black-box models of unknown nonlinear systems, and then using such models to synthesize model-based control laws, has already shown considerable potential in many practical applications. On the other hand, the adoption of these architectures by the control systems community has been so far limited, mainly because the generality of these architectures makes it difficult to attain general properties and to build solid theoretical foundations for their safe and profitable use for control design. To address these gaps, we first provide a control engineer-friendly description of the most common RNN architectures, i.e., Neural NARXs (NNARXs), Gated Recurrent Units (GRUs), and Long Short-Term Memory networks (LSTMs), as well as their training procedure. The stability properties of these architectures are then analyzed, using common nonlinear systems’ stability notions such as the Input-to-State Stability (ISS), the Input-to-State Practical Stability (ISPS), and the Incremental Input-to-State Stability (δISS). In particular, sufficient conditions for these properties are devised for the considered RNN architectures, and it is shown how to enforce these conditions during the training procedure, in order to learn provenly stable RNN models. Model-based control strategies are then synthesized for these models. In particular, nonlinear model predictive control schemes are first designed: in this context, the model’s δISS is shown to enable the attainment of nominal closed-loop stability and, under a suitable design of the control scheme, also robust asymptotic zero-error output regulation. Then, an alternative computationally-lightweight control scheme, based on the internal model control strategy, is proposed, and its closed-loop properties are discussed. The performances of these control schemes are tested on several nonlinear benchmark systems, demonstrating the potentiality of the proposed framework. Finally, some fundamental issues for the practical implementation of RNN-based control strategies are mentioned. In particular, we discuss the need for the safety verification of RNN models and their adaptation in front of changes of the plant’s behavior, the definition of RNN structures that exploit qualitative physical knowledge of the system to boost the performances and interpretability of these models, and the problem of designing control schemes that are robust to the unavoidable plant-model mismatch.
@phdthesis{bonassi2023reconciling, title = {Reconciling deep learning and control theory: recurrent neural networks for model-based control design}, author = {Bonassi, Fabio}, year = {2023}, month = feb, address = {Milan, Italy}, school = {Politecnico di Milano}, type = {PhD thesis}, }
2022
- JournalOn Recurrent Neural Networks for learning-based control: recent results and ideas for future developmentsFabio Bonassi, Marcello Farina, Jing Xie, and Riccardo ScattoliniJournal of Process Control, Feb 2022
This paper aims to discuss and analyze the potentialities of Recurrent Neural Networks (RNN) in control design applications. The main families of RNN are considered, namely Neural Nonlinear AutoRegressive eXogenous, Echo State Networks, Long Short Term Memory, and Gated Recurrent Units. The goal is twofold. Firstly, to survey recent results concerning the training of RNN that enjoy Input-to-State Stability (ISS) and Incremental Input-to-State Stability (𝛿ISS) guarantees. Secondly, to discuss the issues that still hinder the widespread use of RNN for control, namely their robustness, verifiability, and interpretability. The former properties are related to the so-called generalization capabilities of the networks, i.e. their consistency with the underlying real plants, even in presence of unseen or perturbed input trajectories. The latter is instead related to the possibility of providing a clear formal connection between the RNN model and the plant. In this context, we illustrate how ISS and 𝛿ISS represent a significant step towards the robustness and verifiability of the RNN models, while the requirement of interpretability paves the way to the use of physics-based networks. The design of model predictive controllers with RNN as plant’s model is also briefly discussed. Lastly, some of the main topics of the paper are illustrated on a simulated chemical system.
@article{bonassi2022survey, title = {On Recurrent Neural Networks for learning-based control: recent results and ideas for future developments}, author = {Bonassi, Fabio and Farina, Marcello and Xie, Jing and Scattolini, Riccardo}, journal = {Journal of Process Control}, volume = {114}, pages = {92-104}, year = {2022}, issn = {0959-1524}, doi = {10.1016/j.jprocont.2022.04.011}, }
2021
- JournalOn the stability properties of Gated Recurrent Units neural networksFabio Bonassi, Marcello Farina, and Riccardo ScattoliniSystem & Control Letters, Feb 2021
The goal of this paper is to provide sufficient conditions for guaranteeing the Input-to-State Stability (ISS) and the Incremental Input-to-State Stability (δISS) of Gated Recurrent Units (GRUs) neural networks. These conditions, devised for both single-layer and multi-layer architectures, consist of nonlinear inequalities on network’s weights. They can be employed to check the stability of trained networks, or can be enforced as constraints during the training procedure of a GRU. The resulting training procedure is tested on a Quadruple Tank nonlinear benchmark system, showing satisfactory modeling performances.
@article{bonassi2020stability, title = {On the stability properties of Gated Recurrent Units neural networks}, author = {Bonassi, Fabio and Farina, Marcello and Scattolini, Riccardo}, journal = {System \& Control Letters}, publisher = {Elsevier}, volume = {157}, pages = {105049}, year = {2021}, issn = {0167-6911}, doi = {10.1016/j.sysconle.2021.105049}, url = {https://www.sciencedirect.com/science/article/pii/S0167691121001791} }
- ConferenceStability of discrete-time feed-forward neural networks in NARX configurationFabio Bonassi, Marcello Farina, and Riccardo ScattoliniIn 19th IFAC Symposium on System Identification (SYSID), Feb 2021
IFAC Best Student Paper Award at SYSID 2021
The idea of using Feed-Forward Neural Networks (FFNNs) as regression functions for Nonlinear AutoRegressive eXogenous (NARX) models, leading to models herein named Neural NARXs (NNARXs), has been quite popular in the early days of machine learning applied to nonlinear system identification, owing to their simple structure and ease of application to control design. Nonetheless, few theoretical results are available concerning the stability properties of these models. In this paper we address this problem, providing a sufficient condition under which NNARX models are guaranteed to enjoy the Input-to-State Stability (ISS) and the Incremental Input-to-State Stability (δISS) properties. This condition, which is an inequality on the weights of the underlying FFNN, can be enforced during the training procedure to ensure the stability of the model. The proposed model, along with this stability condition, are tested on the pH neutralization process benchmark, showing satisfactory results.
@inproceedings{bonassi2021nnarx, title = {Stability of discrete-time feed-forward neural networks in NARX configuration}, volume = {54}, number = {7}, pages = {547-552}, year = {2021}, booktitle = {19th IFAC Symposium on System Identification (SYSID)}, issn = {2405-8963}, doi = {10.1016/j.ifacol.2021.08.417}, url = {https://www.sciencedirect.com/science/article/pii/S2405896321011915}, author = {Bonassi, Fabio and Farina, Marcello and Scattolini, Riccardo}, keywords = {Neural networks, Nonlinear System Identification, Identification for Control, Input-to-State Stability, Incremental Input-to-State Stability}, }