Goto

Collaborating Authors

 internal feature


SAFE: Multitask Failure Detection for Vision-Language-Action Models

Gu, Qiao, Ju, Yuanliang, Sun, Shengxiang, Gilitschenski, Igor, Nishimura, Haruki, Itkina, Masha, Shkurti, Florian

arXiv.org Artificial Intelligence

While vision-language-action models (VLAs) have shown promising robotic behaviors across a diverse set of manipulation tasks, they achieve limited success rates when deployed on novel tasks out of the box. To allow these policies to safely interact with their environments, we need a failure detector that gives a timely alert such that the robot can stop, backtrack, or ask for help. However, existing failure detectors are trained and tested only on one or a few specific tasks, while generalist VLAs require the detector to generalize and detect failures also in unseen tasks and novel environments. In this paper, we introduce the multitask failure detection problem and propose SAFE, a failure detector for generalist robot policies such as VLAs. We analyze the VLA feature space and find that VLAs have sufficient high-level knowledge about task success and failure, which is generic across different tasks. Based on this insight, we design SAFE to learn from VLA internal features and predict a single scalar indicating the likelihood of task failure. SAFE is trained on both successful and failed rollouts and is evaluated on unseen tasks. SAFE is compatible with different policy architectures. We test it on OpenVLA, $π_0$, and $π_0$-FAST in both simulated and real-world environments extensively. We compare SAFE with diverse baselines and show that SAFE achieves state-of-the-art failure detection performance and the best trade-off between accuracy and detection time using conformal prediction. More qualitative results and code can be found at the project webpage: https://vla-safe.github.io/


Conformal Prediction for Multimodal Regression

Bose, Alexis, Ethier, Jonathan, Guinand, Paul

arXiv.org Artificial Intelligence

This paper introduces multimodal conformal regression. Traditionally confined to scenarios with solely numerical input features, conformal prediction is now extended to multimodal contexts through our methodology, which harnesses internal features from complex neural network architectures processing images and unstructured text. Our findings highlight the potential for internal neural network features, extracted from convergence points where multimodal information is combined, to be used by conformal prediction to construct prediction intervals (PIs). This capability paves new paths for deploying conformal prediction in domains abundant with multimodal data, enabling a broader range of problems to benefit from guaranteed distribution-free uncertainty quantification.


Self-Normalizing Neural Network, Enabling One Shot Transfer Learning for Modeling EDFA Wavelength Dependent Gain

Raj, Agastya, Wang, Zehao, Slyne, Frank, Chen, Tingjun, Kilper, Dan, Ruffini, Marco

arXiv.org Artificial Intelligence

We present a novel ML framework for modeling the wavelength-dependent gain of multiple EDFAs, based on semi-supervised, self-normalizing neural networks, enabling one-shot transfer learning. Our experiments on 22 EDFAs in Open Ireland and COSMOS testbeds show high-accuracy transfer-learning even when operated across different amplifier types.


Test-Time Adaptation for Point Cloud Upsampling Using Meta-Learning

Hatem, Ahmed, Qian, Yiming, Wang, Yang

arXiv.org Artificial Intelligence

Affordable 3D scanners often produce sparse and non-uniform point clouds that negatively impact downstream applications in robotic systems. While existing point cloud upsampling architectures have demonstrated promising results on standard benchmarks, they tend to experience significant performance drops when the test data have different distributions from the training data. To address this issue, this paper proposes a test-time adaption approach to enhance model generality of point cloud upsampling. The proposed approach leverages meta-learning to explicitly learn network parameters for test-time adaption. Our method does not require any prior information about the test data. During meta-training, the model parameters are learned from a collection of instance-level tasks, each of which consists of a sparse-dense pair of point clouds from the training data. During meta-testing, the trained model is fine-tuned with a few gradient updates to produce a unique set of network parameters for each test instance. The updated model is then used for the final prediction. Our framework is generic and can be applied in a plug-and-play manner with existing backbone networks in point cloud upsampling. Extensive experiments demonstrate that our approach improves the performance of state-of-the-art models.


Security of Deep Learning Methodologies: Challenges and Opportunities

Rezaei, Shahbaz, Liu, Xin

arXiv.org Artificial Intelligence

University of California, Davis Abstract--Despite the plethora of studies about security vulnerabilities and defenses of deep learning models, security aspects of deep learning methodologies, such as transfer learning, have been rarely studied. In this article, we highlight the security challenges and research opportunities of these methodologies, focusing on vulnerabilities and attacks unique to them. W ith the widespread adaptation of deep neural networks (DNN), their security challenges have received significant attention from both academia and industry, especially for mission critical applications, such as road sign detection for autonomous vehicles, face recognition in authentication systems, and fraud detection in financial systems. There are three major types of attacks on deep learning models, namely adversarial attacks, data poisoning, and exploratory attacks. Particularly, adversarial attacks, which aim to carefully craft inputs that cause the model to misclassify, has been extensively studied and many defence mechanisms have been proposed to alleviate them. These attacks are of paramount importance because they are effective, moderately simple to launch, and often transferable from one model to another. In literature, there are several survey and review papers on deep learning security and defence mechanisms. In this article, we focus on security of a much less explored area of machine learning - machine learning methodologies. Machine learning methodologies have been widely used to mitigate the restrictions and assumptions of a typical machine learning process. A typical DNN training process assumes large labeled dataset(s), access to high computational resources, non-private and centralized data, standard training and hyper-parameter tuning, and fixed task distribution over time. However, these assumptions are often difficult to realize in practice. Notwithstanding the proliferation of these machine learning methodologies, their security aspects have not been comprehensively analyzed, if ever studied. In this article, we focus on potential attacks, security vulnerabilities, and future directions specific to each learning methodology.


An empirical learning-based validation procedure for simulation workflow

Liu, Zhuqing, Lai, Liyuanjun, Zhang, Lin

arXiv.org Machine Learning

Simulation workflow is a top-level model for the design and control of simulation process. It connects multiple simulation components with time and interaction restrictions to form a complete simulation system. Before the construction and evaluation of the component models, the validation of upper-layer simulation workflow is of the most importance in a simulation system. However, the methods especially for validating simulation workflow is very limit. Many of the existing validation techniques are domain-dependent with cumbersome questionnaire design and expert scoring. Therefore, this paper present an empirical learning-based validation procedure to implement a semi-automated evaluation for simulation workflow. First, representative features of general simulation workflow and their relations with validation indices are proposed. The calculation process of workflow credibility based on Analytic Hierarchy Process (AHP) is then introduced. In order to make full use of the historical data and implement more efficient validation, four learning algorithms, including back propagation neural network (BPNN), extreme learning machine (ELM), evolving new-neuron (eNFN) and fast incremental gaussian mixture model (FIGMN), are introduced for constructing the empirical relation between the workflow credibility and its features. A case study on a landing-process simulation workflow is established to test the feasibility of the proposed procedure. The experimental results also provide some useful overview of the state-of-the-art learning algorithms on the credibility evaluation of simulation models.