Calinescu, Radu
Safe Reinforcement Learning in Black-Box Environments via Adaptive Shielding
Bethell, Daniel, Gerasimou, Simos, Calinescu, Radu, Imrie, Calum
Empowering safe exploration of reinforcement learning (RL) agents during training is a critical impediment towards deploying RL agents in many real-world scenarios. Training RL agents in unknown, black-box environments poses an even greater safety risk when prior knowledge of the domain/task is unavailable. We introduce ADVICE (Adaptive Shielding with a Contrastive Autoencoder), a novel post-shielding technique that distinguishes safe and unsafe features of state-action pairs during training, thus protecting the RL agent from executing actions that yield potentially hazardous outcomes. Our comprehensive experimental evaluation against state-of-the-art safe RL exploration techniques demonstrates how ADVICE can significantly reduce safety violations during training while maintaining a competitive outcome reward.
Bayesian Learning for the Robust Verification of Autonomous Robots
Zhao, Xingyu, Gerasimou, Simos, Calinescu, Radu, Imrie, Calum, Robu, Valentin, Flynn, David
Autonomous robots used in infrastructure inspection, space exploration and other critical missions operate in highly dynamic environments. As such, they must continually verify their ability to complete the tasks associated with these missions safely and effectively. Here we present a Bayesian learning framework that enables this runtime verification of autonomous robots. The framework uses prior knowledge and observations of the verified robot to learn expected ranges for the occurrence rates of regular and singular (e.g., catastrophic failure) events. Interval continuous-time Markov models defined using these ranges are then analysed to obtain expected intervals of variation for system properties such as mission duration and success probability. We apply the framework to an autonomous robotic mission for underwater infrastructure inspection and repair. The formal proofs and experiments presented in the paper show that our framework produces results that reflect the uncertainty intrinsic to many real-world systems, enabling the robust verification of their quantitative properties under parametric uncertainty.
Robust Uncertainty Quantification using Conformalised Monte Carlo Prediction
Bethell, Daniel, Gerasimou, Simos, Calinescu, Radu
Deploying deep learning models in safety-critical applications remains a very challenging task, mandating the provision of assurances for the dependable operation of these models. Uncertainty quantification (UQ) methods estimate the model's confidence per prediction, informing decision-making by considering the effect of randomness and model misspecification. Despite the advances of state-of-the-art UQ methods, they are computationally expensive or produce conservative prediction sets/intervals. We introduce MC-CP, a novel hybrid UQ method that combines a new adaptive Monte Carlo (MC) dropout method with conformal prediction (CP). MC-CP adaptively modulates the traditional MC dropout at runtime to save memory and computation resources, enabling predictions to be consumed by CP, yielding robust prediction sets/intervals. Throughout comprehensive experiments, we show that MC-CP delivers significant improvements over advanced UQ methods, like MC dropout, RAPS and CQR, both in classification and regression benchmarks. MC-CP can be easily added to existing models, making its deployment simple.
Discrete-Event Controller Synthesis for Autonomous Systems with Deep-Learning Perception Components
Calinescu, Radu, Imrie, Calum, Mangal, Ravi, Rodrigues, Genaína Nunes, Păsăreanu, Corina, Santana, Misael Alpizar, Vázquez, Gricel
We present DeepDECS, a new method for the synthesis of correct-by-construction discrete-event controllers for autonomous systems that use deep neural network (DNN) classifiers for the perception step of their decision-making processes. Despite major advances in deep learning in recent years, providing safety guarantees for these systems remains very challenging. Our controller synthesis method addresses this challenge by integrating DNN verification with the synthesis of verified Markov models. The synthesised models correspond to discrete-event controllers guaranteed to satisfy the safety, dependability and performance requirements of the autonomous system, and to be Pareto optimal with respect to a set of optimisation objectives. We use the method in simulation to synthesise controllers for mobile-robot collision mitigation and for maintaining driver attentiveness in shared-control autonomous driving.
Closed-loop Analysis of Vision-based Autonomous Systems: A Case Study
Pasareanu, Corina S., Mangal, Ravi, Gopinath, Divya, Yaman, Sinem Getir, Imrie, Calum, Calinescu, Radu, Yu, Huafeng
Deep neural networks (DNNs) are increasingly used in safety-critical autonomous systems as perception components processing high-dimensional image data. Formal analysis of these systems is particularly challenging due to the complexity of the perception DNNs, the sensors (cameras), and the environment conditions. We present a case study applying formal probabilistic analysis techniques to an experimental autonomous system that guides airplanes on taxiways using a perception DNN. We address the above challenges by replacing the camera and the network with a compact probabilistic abstraction built from the confusion matrices computed for the DNN on a representative image data set. We also show how to leverage local, DNN-specific analyses as run-time guards to increase the safety of the overall system. Our findings are applicable to other autonomous systems that use complex DNNs for perception.
Verified Synthesis of Optimal Safety Controllers for Human-Robot Collaboration
Gleirscher, Mario, Calinescu, Radu, Douthwaite, James, Lesage, Benjamin, Paterson, Colin, Aitken, Jonathan, Alexander, Rob, Law, James
We present a tool-supported approach for the synthesis, verification and validation of the control software responsible for the safety of the human-robot interaction in manufacturing processes that use collaborative robots. In human-robot collaboration, software-based safety controllers are used to improve operational safety, e.g., by triggering shutdown mechanisms or emergency stops to avoid accidents. Complex robotic tasks and increasingly close human-robot interaction pose new challenges to controller developers and certification authorities. Key among these challenges is the need to assure the correctness of safety controllers under explicit (and preferably weak) assumptions. Our controller synthesis, verification and validation approach is informed by the process, risk analysis, and relevant safety regulations for the target application. Controllers are selected from a design space of feasible controllers according to a set of optimality criteria, are formally verified against correctness criteria, and are translated into executable code and validated in a digital twin. The resulting controller can detect the occurrence of hazards, move the process into a safe state, and, in certain circumstances, return the process to an operational state from which it can resume its original task. We show the effectiveness of our software engineering approach through a case study involving the development of a safety controller for a manufacturing work cell equipped with a collaborative robot.
Maintaining driver attentiveness in shared-control autonomous driving
Calinescu, Radu, Alasmari, Naif, Gleirscher, Mario
We present a work-in-progress approach to improving driver attentiveness in cars provided with automated driving systems. The approach is based on a control loop that monitors the driver's biometrics (eye movement, heart rate, etc.) and the state of the car; analyses the driver's attentiveness level using a deep neural network; plans driver alerts and changes in the speed of the car using a formally verified controller; and executes this plan using actuators ranging from acoustic and visual to haptic devices. The paper presents (i) the self-adaptive system formed by this monitor-analyse-plan-execute (MAPE) control loop, the car and the monitored driver, and (ii) the use of probabilistic model checking to synthesise the controller for the planning step of the MAPE loop.
Guidance on the Assurance of Machine Learning in Autonomous Systems (AMLAS)
Hawkins, Richard, Paterson, Colin, Picardi, Chiara, Jia, Yan, Calinescu, Radu, Habli, Ibrahim
Machine Learning (ML) is now used in a range of systems with results that are reported to exceed, under certain conditions, human performance. Many of these systems, in domains such as healthcare, automotive and manufacturing, exhibit high degrees of autonomy and are safety critical. Establishing justified confidence in ML forms a core part of the safety case for these systems. In this document we introduce a methodology for the Assurance of Machine Learning for use in Autonomous Systems (AMLAS). AMLAS comprises a set of safety case patterns and a process for (1) systematically integrating safety assurance into the development of ML components and (2) for generating the evidence base for explicitly justifying the acceptable safety of these components when integrated into autonomous system applications. The material in this document is provided as guidance only. No responsibility for loss occasioned to any person acting or refraining from action as a result of this material or any comments made can be accepted by the authors or The University of York.
Challenges in the Safety-Security Co-Assurance of Collaborative Industrial Robots
Gleirscher, Mario, Johnson, Nikita, Karachristou, Panayiotis, Calinescu, Radu, Law, James, Clark, John
The coordinated assurance of interrelated critical properties, such as system safety and cyber-security, is one of the toughest challenges in critical systems engineering. In this chapter, we summarise approaches to the coordinated assurance of safety and security. Then, we highlight the state of the art and recent challenges in human-robot collaboration in manufacturing both from a safety and security perspective. We conclude with a list of procedural and technological issues to be tackled in the coordinated assurance of collaborative industrial robots.
Safety Controller Synthesis for Collaborative Robots
Gleirscher, Mario, Calinescu, Radu
Safety Controller Synthesis for Collaborative Robots Mario Gleirscher, Radu Calinescu Assuring Autonomy International Programme, University of Y ork, Y ork, UK Department of Computer Science, University of Y ork, Y ork, UK mario.gleirscher,radu.calinescu@york.ac.uk Abstract --In human-robot collaboration (HRC), software-based automatic safety controllers (ASCs) are used in various forms (e.g. Complex robotic tasks and increasingly close human-robot interaction pose new challenges to ASC developers and certification authorities. Key among these challenges is the need to assure the correctness of ASCs under reasonably weak assumptions. T o address this need, we introduce and evaluate a tool-supported ASC synthesis method for HRC in manufacturing. Our ASC synthesis is: (i) informed by the manufacturing process, risk analysis, and regulations; (ii) formally verified against correctness criteria; and (iii) selected from a design space of feasible controllers according to a set of optimality criteria. The synthesised ASC can detect the occurrence of hazards, move the process into a safe state, and, in certain circumstances, return the process to an operational state from which it can resume its original task. I NTRODUCTION An effective collaboration between industrial robot systems (IRSs) and humans [1], [2] can leverage their complementary skills, but is difficult to achieve because of uncontrolled hazards and unexploited sensing, tracking, and safety measures [3]. Such hazards have been studied since the 1970s, resulting in elaborate risk taxonomies based on workspaces, tasks, and human body regions [2], [4]-[10]. The majority are impact hazards (e.g. Addressing these hazards involves the examination of each mode of operation (e.g. I, a variety of safety measures [3] can prevent or mitigate hazards and accidents by reducing the probability of their occurrence and the severity of their consequences . There are functional measures using electronic equipment (e.g.