Safe Rational Interactions in Data-driven Control

In autonomous and multi-agent systems players are normally assumed rational and cooperating or competing in groups to achieve their overall objectives. Useful methods to study the resulting interactions come from game theory whereby notions such as cooperative games, competitive games, local and global optima, power in coalitions, Nash and other forms of equilibria are studied.

While this has resulted in a wide literature in game theory, two aspects are not discussed in the literature. Firstly, the agents are normally assumed to be defined by discrete (memoryful or memoryless) policies, but do not incorporate notions of continuous decision making, typical in control theory, or hybrid policies, including policies synthesised from data. Secondly, policies are synthesised with respect to a single or global utility function, but do not take it into account considerations of safety and trustworthiness. The object of this project is to lay the foundations for this analysis to be carried out.

Objective: Study a system made of multiple autonomous agents that are governed by neural networks interacting with linear and non-linear environments. Examples of these include applications in several emerging AI application areas, such as those in smart traffic networks, smart/active buildings and smart grids.

Control theory is traditionally focused with the design of such policies, when a centralized utility notion is available and systems dynamics are known to obey prescribed models (linear or nonlinear equations). Game play allows to tackle similar questions when multiple objective functions are introduced together with the strategic agents’ interactions that these entail. But differently from the state of the art, where these specifications are achieved against objectives by maximising or minimising some utility notion, in this work these will be tackled in conjunction with safety constraints. Crucially, these will be model based and will be made via logic, and verification, or optimisation standpoint.

In addition to this radical paradigm shift, from a technical standpoint, the work will further include the development of robust and theoretically sound parameterizations of dynamical behaviours resulting from past input-output data. These should allow a more flexible deployment of such technologies in contexts where models are hardly available and may be evolving in time.

Requirements:
Expertise in Control and Optimisation is highly beneficial. Some knowledge of machine learning is advantageous.

  1. Ivanov, K. Jothimurugan, S. Hsu, S. Vaidya, R. Alur, and O. Bastani. Compositional Learning and Verification of Neural Network Controllers.ACM Transactions on  Embedded Computing Systems, 20(5s): 1-26, 2021. Compositional Learning and Verification of Neural Network Controllers | ACM Transactions on Embedded Computing Systems
  2. H.D. Tran, F. Cai, M. L. Diego, P. Musau, T. T. Johnson, and X. Koutsoukos. Safety verification of cyber-physical systems with reinforcement learning control. ACM Transactions on Embedded Computing Systems, 18(5s): 1-22, 2019. Safety Verification of Cyber-Physical Systems with Reinforcement Learning Control | ACM Transactions on Embedded Computing Systems
  3. Z. Wang, H. Chao, W. Yixuan, H. Clara, C. Samarjit, and Z. Qi. Bounding Perception Neural Network Uncertainty for Safe Control of Autonomous Systems. In Design, Automation and Test in Europe Conference and Exhibition (DATE21), 1745-1750, 2021. Grenoble, France. IEEE. Bounding Perception Neural Network Uncertainty for Safe Control of Autonomous Systems | IEEE Conference Publication | IEEE Xplore
  4. B. Karg, and L. Sergio. Stability and feasibility of neural network-based controllers via output range analysis. In 59th IEEE Conference on Decision and Control (CDC20), 4947-4954, 2020. Jeju, Korea (South). IEEE. B. Stability and feasibility of neural network-based controllers via output range analysis | IEEE Conference Publication | IEEE Xplore
  5. Yin, S. Peter, and A. Murat. Stability analysis using quadratic constraints for systems with neural network controllers. IEEE Transactions on Automatic Control, doi: 10.1109/TAC.2021.3069388, 2021. Stability Analysis using Quadratic Constraints for Systems with Neural Network Controllers | IEEE Journals & Magazine | IEEE Xplore

Project ID

STAI-CDT-2021-IC-8