Plotting

 Dwyer, Matthew


A DPLL(T) Framework for Verifying Deep Neural Networks

arXiv.org Artificial Intelligence

However, just like traditional software, DNNs can have "bugs", e.g., producing unexpected results on inputs that are different from those in training data, and be attacked, e.g., small perturbations to the inputs by a malicious adversary or even sensorial imperfections result in misclassification [Isac et al. 2022; Ren et al. 2020; Yang et al. 2022; Zhang et al. 2019; Zรผgner et al. 2018]. These issues, which have been observed in many DNNs [Goodfellow et al. 2014; Szegedy et al. 2014] and demonstrated in the real world [Eykholt et al. 2018], naturally raise the question of how DNNs should be tested, validated, and ultimately verified to meet the requirements of relevant robustness or safety standards [Huang et al. 2020; Katz et al. 2017b]. To address this question, researchers have developed a variety of techniques and tools to verify DNNs (e.g., [Huang et al. 2017; Katz et al. 2022, 2019; Liu et al. 2021; Mรผller et al. 2021; Urban and Minรฉ 2021; Wang et al. 2021]). Constraint-based approaches [Ehlers 2017; Huang et al. 2017; Katz et al. 2017a] aim to both correctly prove and disprove properties, but do not scale to large networks. In contrast, abstraction-based approaches [Mรผller et al. 2021; Singh et al. 2018a, 2019b; Wang et al. 2018b, 2021] scale much better, but while modern abstraction verification tools can often refine their abstractions to avoid returning spurious counterexamples they are incomplete. The problem of verifying non-trivial properties of DNNs with piecewise linear activation functions, such as "ReLU", has been shown to be reducible [Katz et al. 2017a] from the classical satisfiability (SAT) problem [Cook 1971]. Despite this complexity, the ability of satisfiability modulo theories (SMT) solvers to scale to large formulae encoding real-world verification problems [Kroening and Strichman 2016] suggests that a similar approaches might be effective for DNN verification. However, the constraint-based DNN verifiers developed to date [Ehlers 2017; Katz et al. 2017a, 2022, 2019] are not among the state-of-the-art as determined by DNN verification competitions [Bak et al. 2021; Mรผller et al. 2022]. Techniques like Planet [Ehlers 2017] and Reluplex [Katz et al. 2017a, 2022] demonstrated how the semantics of a trained DNN could be encoded as a constraint in Linear Real Arithmetic (LRA) and


Deep Learning & Software Engineering: State of Research and Future Directions

arXiv.org Artificial Intelligence

The advent of deep learning (DL) has fundamentally changed the landscape of modern software. Generally, a DL system is comprised of several interconnected computational units that form "layers" which perform mathematical transformations, according to sets of learnable parameters, on data passing through them. These architectures can be "trained" for specific tasks by updating the parameters according to a model's performance on a labeled set of training data. DL represents a fundamental shift in the manner by which machines learn patterns from data by automatically extracting salient features for a given computational task, as opposed to relying upon human intuition. These DL systems can be viewed as an inflection point for software development, as they enable new capabilities that cannot be realized cost-effectively through "traditional" software wherein the behavior of a program must be specified analytically.


Formal Language Constraints for Markov Decision Processes

arXiv.org Machine Learning

In order to satisfy safety conditions, a reinforcement learned (RL) agent maybe constrained from acting freely, e.g., to prevent trajectories that might cause unwanted behavior or physical damage in a robot. We propose a general framework for augmenting a Markov decision process (MDP) with constraints that are described in formal languages over sequences of MDP states and agent actions. Constraint enforcement is implemented by filtering the allowed action set or by applying potential-based reward shaping to implement hard and soft constraint enforcement, respectively. We instantiate this framework using deterministic finite automata to encode constraints and propose methods of augmenting MDP observations with the state of the constraint automaton for learning. We empirically evaluate these methods with a variety of constraints by training Deep Q-Networks in Atari games as well as Proximal Policy Optimization in MuJoCo environments. We experimentally find that our approaches are effective in significantly reducing or eliminating constraint violations with either minimal negative or, depending on the constraint, a clear positive impact on final performance.