Laddaga, Robert
Reports of the Workshops of the Thirty-First AAAI Conference on Artificial Intelligence
Anderson, Monica (University of Alabama) | Barták, Roman (Charles University) | Brownstein, John S. (Boston Children's Hospital, Harvard University) | Buckeridge, David L. (McGill University) | Eldardiry, Hoda (Palo Alto Research Center) | Geib, Christopher (Drexel University) | Gini, Maria (University of Minnesota) | Isaksen, Aaron (New York University) | Keren, Sarah (Technion University) | Laddaga, Robert (Vanderbilt University) | Lisy, Viliam (Czech Technical University) | Martin, Rodney (NASA Ames Research Center) | Martinez, David R. (MIT Lincoln Laboratory) | Michalowski, Martin (University of Ottawa) | Michael, Loizos (Open University of Cyprus) | Mirsky, Reuth (Ben-Gurion University) | Nguyen, Thanh (University of Michigan) | Paul, Michael J. (University of Colorado Boulder) | Pontelli, Enrico (New Mexico State University) | Sanner, Scott (University of Toronto) | Shaban-Nejad, Arash (University of Tennessee) | Sinha, Arunesh (University of Michigan) | Sohrabi, Shirin (IBM T. J. Watson Research Center) | Sricharan, Kumar (Palo Alto Research Center) | Srivastava, Biplav (IBM T. J. Watson Research Center) | Stefik, Mark (Palo Alto Research Center) | Streilein, William W. (MIT Lincoln Laboratory) | Sturtevant, Nathan (University of Denver) | Talamadupula, Kartik (IBM T. J. Watson Research Center) | Thielscher, Michael (University of New South Wales) | Togelius, Julian (New York University) | Tran, So Cao (New Mexico State University) | Tran-Thanh, Long (University of Southampton) | Wagner, Neal (MIT Lincoln Laboratory) | Wallace, Byron C. (Northeastern University) | Wilk, Szymon (Poznan University of Technology) | Zhu, Jichen (Drexel University)
Reports of the Workshops of the Thirty-First AAAI Conference on Artificial Intelligence
Anderson, Monica (University of Alabama) | Barták, Roman (Charles University) | Brownstein, John S. (Boston Children's Hospital, Harvard University) | Buckeridge, David L. (McGill University) | Eldardiry, Hoda (Palo Alto Research Center) | Geib, Christopher (Drexel University) | Gini, Maria (University of Minnesota) | Isaksen, Aaron (New York University) | Keren, Sarah (Technion University) | Laddaga, Robert (Vanderbilt University) | Lisy, Viliam (Czech Technical University) | Martin, Rodney (NASA Ames Research Center) | Martinez, David R. (MIT Lincoln Laboratory) | Michalowski, Martin (University of Ottawa) | Michael, Loizos (Open University of Cyprus) | Mirsky, Reuth (Ben-Gurion University) | Nguyen, Thanh (University of Michigan) | Paul, Michael J. (University of Colorado Boulder) | Pontelli, Enrico (New Mexico State University) | Sanner, Scott (University of Toronto) | Shaban-Nejad, Arash (University of Tennessee) | Sinha, Arunesh (University of Michigan) | Sohrabi, Shirin (IBM T. J. Watson Research Center) | Sricharan, Kumar (Palo Alto Research Center) | Srivastava, Biplav (IBM T. J. Watson Research Center) | Stefik, Mark (Palo Alto Research Center) | Streilein, William W. (MIT Lincoln Laboratory) | Sturtevant, Nathan (University of Denver) | Talamadupula, Kartik (IBM T. J. Watson Research Center) | Thielscher, Michael (University of New South Wales) | Togelius, Julian (New York University) | Tran, So Cao (New Mexico State University) | Tran-Thanh, Long (University of Southampton) | Wagner, Neal (MIT Lincoln Laboratory) | Wallace, Byron C. (Northeastern University) | Wilk, Szymon (Poznan University of Technology) | Zhu, Jichen (Drexel University)
The AAAI-17 workshop program included 17 workshops covering a wide range of topics in AI. Workshops were held Sunday and Monday, February 4-5, 2017 at the Hilton San Francisco Union Square in San Francisco, California, USA. This report contains summaries of 12 of the workshops, and brief abstracts of the remaining 5
Qualitative Reasoning about Cyber Intrusions
Robertson, Paul (DOLL Inc.) | Laddaga, Robert (Vanderbilt University) | Goldman, Robert (SIFT) | Burstein, Mark (SIFT) | Cerys, Daniel (DOLL Inc.)
In this paper we discuss work performed in an ambitious DARPA funded cyber security effort. The broad approach taken by the project was for the network to be self-aware and to self-adapt in order to dodge attacks. In critical systems, it is not always the best or practical thing, to shut down the network under attack. The paper describes the qualitative trust modeling and diagnosis system that maintains a model of trust for networked resources using a combination of two basic ideas: Conditional trust (based on conditional preference (CP-Nets) and the principle of maximum entropy (PME)). We describe Monte-Carlo simulations of using adaptive security based on our trust model. The results of the simulations show the trade-off, under ideal conditions, between additional resource provisioning and attack mitigation.
AWDRAT: A Cognitive Middleware System for Information Survivability
Shrobe, Howard, Laddaga, Robert, Balzer, Bob, Goldman, Neil, Wile, Dave, Tallis, Marcelo, Hollebeek, Tim, Egyed, Alexander
The infrastructure of modern society is controlled by software systems that are vulnerable to attacks. Many such attacks, launched by "recreational hackers" have already led to severe disruptions and significant cost. It, therefore, is critical that we find ways to protect such systems and to enable them to continue functioning even after a successful attack. This article describes AWDRAT, a prototype middleware system for providing survivability to both new and legacy applications. AWDRAT stands for architectural differencing, wrappers, diagnosis, recovery, adaptive software, and trust modeling. AWDRAT uses these techniques to gain visibility into the execution of an application system and to compare the application's actual behavior to that which is expected. In the case of a deviation, AWDRAT conducts a diagnosis that determines which computational resources are likely to have been compromised and then adds these assessments to its trust model. The trust model in turn guides the recovery process, particularly by guiding the system in its choice among functionally equivalent methods and resources.AWDRAT has been applied to and evaluated on an example application system, a graphical editor for constructing mission plans. We describe a series of experiments that were performed to test the effectiveness of AWDRAT in recognizing and recovering from simulated attacks, and we present data showing the effectiveness of AWDRAT in detecting a variety of compromises to the application system (approximately 90 percent of all simulated attacks are detected, diagnosed, and corrected). We also summarize some lessons learned from the AWDRAT experiments and suggest approaches for comprehensive application protection methods and techniques.