ml-based system
Capturing Semantic Flow of ML-based Systems
Yoo, Shin, Feldt, Robert, Kim, Somin, Kim, Naryeong
ML-based systems are software systems that incorporates machine learning components such as Deep Neural Networks (DNNs) or Large Language Models (LLMs). While such systems enable advanced features such as high performance computer vision, natural language processing, and code generation, their internal behaviour remain largely opaque to traditional dynamic analysis such as testing: existing analysis typically concern only what is observable from the outside, such as input similarity or class label changes. We propose semantic flow, a concept designed to capture the internal behaviour of ML-based system and to provide a platform for traditional dynamic analysis techniques to be adapted to. Semantic flow combines the idea of control flow with internal states taken from executions of ML-based systems, such as activation values of a specific layer in a DNN, or embeddings of LLM responses at a specific inference step of LLM agents. The resulting representation, summarised as semantic flow graphs, can capture internal decisions that are not explicitly represented in the traditional control flow of ML-based systems. We propose the idea of semantic flow, introduce two examples using a DNN and an LLM agent, and finally sketch its properties and how it can be used to adapt existing dynamic analysis techniques for use in ML-based software systems.
- North America > United States (0.28)
- Europe > Norway (0.15)
- North America > Canada > Ontario > Toronto (0.14)
- (2 more...)
Outline of an Independent Systematic Blackbox Test for ML-based Systems
Wiesbrock, Hans-Werner, Großmann, Jürgen
ML-based systems are used today in a wide range of areas, and increasingly also in safety-critical domains. Their range of application is growing exponentially. At the same time, more and more experts are warning of the uncertainties and risks associated with the uncontrolled and overly rapid development of AI systems Bengio et al. [22.03.2023]. In general, there is a growing need to provide methods and procedures for testing functioning and quality characteristics of these systems. Various methods currently exist to test and verify ML-based systems, be it formal verification, simulation approaches or classical testing Albarghouthi, Jackson et al., Vasu Singh et al., or new analysis methods in the context of XAI Hoyer et al., Guidotti et al.. The methods aim for providing evidence on the robustness and trustworthiness of the ML models or ML-based system (ML - Machine Learning). Similar to the traditional development of complex software systems, testing has also proven to be the most effective method for proving quality and gaining trust in ML.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
- (13 more...)
Real-world Machine Learning Systems: A survey from a Data-Oriented Architecture Perspective
Cabrera, Christian, Paleyes, Andrei, Thodoroff, Pierre, Lawrence, Neil D.
Machine Learning models are being deployed as parts of real-world systems with the upsurge of interest in artificial intelligence. The design, implementation, and maintenance of such systems are challenged by real-world environments that produce larger amounts of heterogeneous data and users requiring increasingly faster responses with efficient resource consumption. These requirements push prevalent software architectures to the limit when deploying ML-based systems. Data-oriented Architecture (DOA) is an emerging concept that equips systems better for integrating ML models. DOA extends current architectures to create data-driven, loosely coupled, decentralised, open systems. Even though papers on deployed ML-based systems do not mention DOA, their authors made design decisions that implicitly follow DOA. The reasons why, how, and the extent to which DOA is adopted in these systems are unclear. Implicit design decisions limit the practitioners' knowledge of DOA to design ML-based systems in the real world. This paper answers these questions by surveying real-world deployments of ML-based systems. The survey shows the design decisions of the systems and the requirements these satisfy. Based on the survey findings, we also formulate practical advice to facilitate the deployment of ML-based systems. Finally, we outline open challenges to deploying DOA-based systems that integrate ML models.
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- Asia > Pakistan (0.04)
- (19 more...)
- Overview (1.00)
- Research Report > New Finding (0.67)
- Information Technology > Services (1.00)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine > Therapeutic Area (1.00)
- (7 more...)
- Information Technology > Data Science > Data Mining > Big Data (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Information Technology > Artificial Intelligence > Robots > Autonomous Vehicles (0.93)
Adversarial Attacks Against Uncertainty Quantification
Ledda, Emanuele, Angioni, Daniele, Piras, Giorgio, Fumera, Giorgio, Biggio, Battista, Roli, Fabio
Machine-learning models can be fooled by adversarial examples, i.e., carefully-crafted input perturbations that force models to output wrong predictions. While uncertainty quantification has been recently proposed to detect adversarial inputs, under the assumption that such attacks exhibit a higher prediction uncertainty than pristine data, it has been shown that adaptive attacks specifically aimed at reducing also the uncertainty estimate can easily bypass this defense mechanism. In this work, we focus on a different adversarial scenario in which the attacker is still interested in manipulating the uncertainty estimate, but regardless of the correctness of the prediction; in particular, the goal is to undermine the use of machine-learning models when their outputs are consumed by a downstream module or by a human operator. Following such direction, we: \textit{(i)} design a threat model for attacks targeting uncertainty quantification; \textit{(ii)} devise different attack strategies on conceptually different UQ techniques spanning for both classification and semantic segmentation problems; \textit{(iii)} conduct a first complete and extensive analysis to compare the differences between some of the most employed UQ approaches under attack. Our extensive experimental analysis shows that our attacks are more effective in manipulating uncertainty quantification measures than attacks aimed to also induce misclassifications.
- North America > United States > California > Los Angeles County > Long Beach (0.04)
- Europe > Italy > Sardinia > Cagliari (0.04)
- North America > United States > New York > Richmond County > New York City (0.04)
- (8 more...)
- Information Technology > Security & Privacy (1.00)
- Government > Military (0.85)
To Classify is to Interpret: Building Taxonomies from Heterogeneous Data through Human-AI Collaboration
Meier, Sebastian, Glinka, Katrin
Taxonomies serve this purpose as structured classification schemes that adhere to domain-specific standards. The importance of organizing, segmenting, and classifying data is especially obvious in light of the ever growing amount of information that is being created, aggregated, and made available through specialized data repositories or on the Internet. In light of the amount and heterogeneity of the available data, classification can hardly be addressed by means of manual-cognitive processing alone. Systems that integrate machine learning (ML) are able to process large amounts of data and, thus, can help with the task of classification and organization. However, delegating this task to ML-based systems in their entirety would mean that we sideline human interpretation and rely on the output of black-boxed systems that reproduce language ideologies and representational harms (see, e.g., [5]). As an attempt to highlight the interpretative character of classification and taxonomy building, we propose to design ML-based systems that enable human-AI collaboration. Such systems are designed with the goal to effectively combine human competencies and computational capabilities (see, e.g.,[27, 29]). Our approach enables domain experts to iteratively interact with the suggestions of the system while retaining interpretative authority. We report on the concept and implementation of this approach that we realized for two real-world use cases.
- North America > United States > New York > New York County > New York City (0.05)
- Europe > Belgium > Flanders (0.05)
- North America > United States > Texas > Travis County > Austin (0.04)
- (9 more...)
Bug Characterization in Machine Learning-based Systems
Morovati, Mohammad Mehdi, Nikanjam, Amin, Tambon, Florian, Khomh, Foutse, Ming, Zhen, Jiang, null
Rapid growth of applying Machine Learning (ML) in different domains, especially in safety-critical areas, increases the need for reliable ML components, i.e., a software component operating based on ML. Understanding the bugs characteristics and maintenance challenges in ML-based systems can help developers of these systems to identify where to focus maintenance and testing efforts, by giving insights into the most error-prone components, most common bugs, etc. In this paper, we investigate the characteristics of bugs in ML-based software systems and the difference between ML and non-ML bugs from the maintenance viewpoint. We extracted 447,948 GitHub repositories that used one of the three most popular ML frameworks, i.e., TensorFlow, Keras, and PyTorch. After multiple filtering steps, we select the top 300 repositories with the highest number of closed issues. We manually investigate the extracted repositories to exclude non-ML-based systems. Our investigation involved a manual inspection of 386 sampled reported issues in the identified ML-based systems to indicate whether they affect ML components or not. Our analysis shows that nearly half of the real issues reported in ML-based systems are ML bugs, indicating that ML components are more error-prone than non-ML components. Next, we thoroughly examined 109 identified ML bugs to identify their root causes, symptoms, and calculate their required fixing time. The results also revealed that ML bugs have significantly different characteristics compared to non-ML bugs, in terms of the complexity of bug-fixing (number of commits, changed files, and changed lines of code). Based on our results, fixing ML bugs are more costly and ML components are more error-prone, compared to non-ML bugs and non-ML components respectively. Hence, paying a significant attention to the reliability of the ML components is crucial in ML-based systems.
- North America > United States > New York > New York County > New York City (0.06)
- North America > Canada > Quebec > Montreal (0.04)
- North America > United States > Pennsylvania (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study > Negative Result (0.93)
Towards a safe MLOps Process for the Continuous Development and Safety Assurance of ML-based Systems in the Railway Domain
Zeller, Marc, Waschulzik, Thomas, Schmid, Reiner, Bahlmann, Claus
Traditional automation technologies alone are not sufficient to enable driverless operation of trains (called Grade of Automation (GoA) 4) on non-restricted infrastructure. The required perception tasks are nowadays realized using Machine Learning (ML) and thus need to be developed and deployed reliably and efficiently. One important aspect to achieve this is to use an MLOps process for tackling improved reproducibility, traceability, collaboration, and continuous adaptation of a driverless operation to changing conditions. MLOps mixes ML application development and operation (Ops) and enables high frequency software releases and continuous innovation based on the feedback from operations. In this paper, we outline a safe MLOps process for the continuous development and safety assurance of ML-based systems in the railway domain. It integrates system engineering, safety assurance, and the ML life-cycle in a comprehensive workflow. We present the individual stages of the process and their interactions. Moreover, we describe relevant challenges to automate the different stages of the safe MLOps process.
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Europe > Germany > Berlin (0.04)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
A Survey on XAI for Beyond 5G Security: Technical Aspects, Use Cases, Challenges and Research Directions
Senevirathna, Thulitha, La, Vinh Hoa, Marchal, Samuel, Siniarski, Bartlomiej, Liyanage, Madhusanka, Wang, Shen
With the advent of 5G commercialization, the need for more reliable, faster, and intelligent telecommunication systems are envisaged for the next generation beyond 5G (B5G) radio access technologies. Artificial Intelligence (AI) and Machine Learning (ML) are not just immensely popular in the service layer applications but also have been proposed as essential enablers in many aspects of B5G networks, from IoT devices and edge computing to cloud-based infrastructures. However, existing B5G ML-security surveys tend to place more emphasis on AI/ML model performance and accuracy than on the models' accountability and trustworthiness. In contrast, this paper explores the potential of Explainable AI (XAI) methods, which would allow B5G stakeholders to inspect intelligent black-box systems used to secure B5G networks. The goal of using XAI in the security domain of B5G is to allow the decision-making processes of the ML-based security systems to be transparent and comprehensible to B5G stakeholders making the systems accountable for automated actions. In every facet of the forthcoming B5G era, including B5G technologies such as RAN, zero-touch network management, E2E slicing, this survey emphasizes the role of XAI in them and the use cases that the general users would ultimately enjoy. Furthermore, we presented the lessons learned from recent efforts and future research directions on top of the currently conducted projects involving XAI.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.13)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Europe > Finland > Northern Ostrobothnia > Oulu (0.04)
- (15 more...)
- Overview (1.00)
- Research Report > New Finding (0.45)
- Research Report > Promising Solution (0.45)
Threats, Vulnerabilities, and Controls of Machine Learning Based Systems: A Survey and Taxonomy
Kawamoto, Yusuke, Miyake, Kazumasa, Konishi, Koichi, Oiwa, Yutaka
In this article, we propose the Artificial Intelligence Security Taxonomy to systematize the knowledge of threats, vulnerabilities, and security controls of machine-learning-based (ML-based) systems. We first classify the damage caused by attacks against ML-based systems, define ML-specific security, and discuss its characteristics. Next, we enumerate all relevant assets and stakeholders and provide a general taxonomy for ML-specific threats. Then, we collect a wide range of security controls against ML-specific threats through an extensive review of recent literature. Finally, we classify the vulnerabilities and controls of an ML-based system in terms of each vulnerable asset in the system's entire lifecycle.
- North America > United States > California > San Francisco County > San Francisco (0.28)
- North America > United States > California > Los Angeles County > Long Beach (0.14)
- Europe > Austria > Vienna (0.14)
- (42 more...)
- Research Report (1.00)
- Overview (1.00)
Toward Certification of Machine-Learning Systems for Low Criticality Airborne Applications
Dmitriev, K., Schumann, J., Holzapfel, F.
The exceptional progress in the field of machine learning (ML) in recent years has attracted a lot of interest in using this technology in aviation. Possible airborne applications of ML include safety-critical functions, which must be developed in compliance with rigorous certification standards of the aviation industry. Current certification standards for the aviation industry were developed prior to the ML renaissance without taking specifics of ML technology into account. There are some fundamental incompatibilities between traditional design assurance approaches and certain aspects of ML-based systems. In this paper, we analyze the current airborne certification standards and show that all objectives of the standards can be achieved for a low-criticality ML-based system if certain assumptions about ML development workflow are applied.
- North America > United States (0.68)
- Europe > Germany > Bavaria > Upper Bavaria > Munich (0.04)
- Research Report (0.64)
- Workflow (0.51)
- Transportation > Air (1.00)
- Aerospace & Defense > Aircraft (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)