Goto

Collaborating Authors

 Military


Stability and Generalization of Adversarial Training for Shallow Neural Networks with Smooth Activation

Neural Information Processing Systems

Adversarial training has emerged as a popular approach for training models that are robust to inference-time adversarial attacks. However, our theoretical understanding of why and when it works remains limited. Prior work has offered generalization analysis of adversarial training, but they are either restricted to the Neural Tangent Kernel (NTK) regime or they make restrictive assumptions about data such as (noisy) linear separability or robust realizability. In this work, we study the stability and generalization of adversarial training for two-layer networks without any data distribution assumptions and beyond the NTK regime. Our findings suggest that for networks with any given initialization and sufficiently large width, the generalization bound can be effectively controlled via early stopping. We further improve the generalization bound by leveraging smoothing using Moreau's envelope.



Regret Minimization in Stackelberg Games with Side Information

Neural Information Processing Systems

Algorithms for playing in Stackelberg games have been deployed in real-world domains including airport security, anti-poaching efforts, and cyber-crime prevention. However, these algorithms often fail to take into consideration the additional information available to each player (e.g.


MMDU: A Multi-Turn Multi-Image Dialog Understanding Benchmark and Instruction-Tuning Dataset for LVLMs

Neural Information Processing Systems

Generating natural and meaningful responses to communicate with multi-modal human inputs is a fundamental capability of Large Vision-Language Models (LVLMs). While current open-source LVLMs demonstrate promising performance in simplified scenarios such as single-turn single-image input, they fall short in real-world conversation scenarios such as following instructions in a long context history with multi-turn and multi-images. Existing LVLM benchmarks primarily focus on single-choice questions or short-form responses, which do not adequately assess the capabilities of LVLMs in real-world human-AI interaction applications. Therefore, we introduce MMDU, a comprehensive benchmark, and MMDU-45k, a large-scale instruction tuning dataset, designed to evaluate and improve LVLMs' abilities in multi-turn and multi-image conversations. We employ the clustering algorithm to find the relevant images and textual descriptions from the open-source Wikipedia and construct the question-answer pairs by human annotators with the assistance of the GPT-4o model.


Japan backs AI chip startup EdgeCortix in boost to defense tech

The Japan Times

EdgeCortix, a Tokyo-based artificial intelligence (AI) chip startup, is riding a wave of interest to foster Japanese semiconductors with defense applications. EdgeCortix, which has won a contract tied to the U.S. Department of Defense, on Wednesday secured government subsidies of 3 billion ( 21 million) to develop energy-efficient AI chiplets for commercialization in 2027. The contract may help revenue more than double this year, founder Sakyasingha Dasgupta said. The products, designed to help robots make real-time decisions and fill the country's labor shortages, target mass production at Taiwan Semiconductor Manufacturing Co.'s plant in Japan. The subsidies are on top of 4 billion in support the semiconductor designer won in November to make chips for next-generation communication systems.


Efficient Adversarial Training in LLMs with Continuous Attacks

Neural Information Processing Systems

Large language models (LLMs) are vulnerable to adversarial attacks that can bypass their safety guardrails. In many domains, adversarial training has proven to be one of the most promising methods to reliably improve robustness against such attacks. Yet, in the context of LLMs, current methods for adversarial training are hindered by the high computational costs required to perform discrete adversarial attacks at each training iteration. We address this problem by instead calculating adversarial attacks in the continuous embedding space of the LLM, which is orders of magnitudes more efficient. We propose a fast adversarial training algorithm (CAT) composed of two losses: the first makes the model robust on continuous embedding attacks computed on an adversarial behaviour dataset; the second ensures the usefulness of the final model by fine-tuning on utility data. Moreover, we introduce CAPO, an adversarial variant of IPO that does not require utility data for adversarially robust alignment. Our empirical evaluation on five models from different families (Gemma, Phi3, Mistral, Zephyr, Llama2) and at different scales (2B, 3.8B, 7B) shows that both algorithms substantially enhance LLM robustness against discrete attacks (GCG, AutoDAN, PAIR), while maintaining utility. Our results demonstrate that robustness to continuous perturbations can extrapolate to discrete threat models. Thereby, we present a path toward scalable adversarial training algorithms for robustly aligning LLMs.


LLaNA: Large Language and NeRF Assistant

Neural Information Processing Systems

Multimodal Large Language Models (MLLMs) have demonstrated an excellent understanding of images and 3D data. However, both modalities have shortcomings in holistically capturing the appearance and geometry of objects. Meanwhile, Neural Radiance Fields (NeRFs), which encode information within the weights of a simple Multi-Layer Perceptron (MLP), have emerged as an increasingly widespread modality that simultaneously encodes the geometry and photorealistic appearance of objects. This paper investigates the feasibility and effectiveness of ingesting NeRF into MLLM. We create LLaNA, the first general-purpose NeRFlanguage assistant capable of performing new tasks such as NeRF captioning and Q&A. Notably, our method directly processes the weights of the NeRF's MLP to extract information about the represented objects without the need to render images or materialize 3D data structures. Moreover, we build a dataset of NeRFs with text annotations for various NeRF-language tasks with no human intervention. Based on this dataset, we develop a benchmark to evaluate the NeRF understanding capability of our method. Results show that processing NeRF weights performs favourably against extracting 2D or 3D representations from NeRFs.



Lockheed Martin CEO shares path to making Trump's 'Golden Dome' missile shield a reality

FOX News

Lockheed Martin CEO Jim Taiclet weighs in on the Trump administration's Golden Dome defense system announcement on'Special Report.' Lockheed Martin CEO Jim Taiclet said President Donald Trump's proposed "Golden Dome" missile shield for the United States is a "fantastic vision" for the country as defense contracting companies work to implement the commander-in-chief's bold idea by the end of his term. "We'll be able to use the Golden Dome concept to make sure the country is increasingly protected against hypersonic threats," Taiclet said in an exclusive interview Tuesday on "Special Report." Trump unveiled his ambitious missile defense plan at the White House last week, which he says will be operational by the time he leaves office. The announcement comes as the United States faces growing threats from adversaries around the world who are making significant inroads in artificial intelligence and drone technology.


Parallel Streaming Wasserstein Barycenters

Neural Information Processing Systems

Efficiently aggregating data from different sources is a challenging problem, particularly when samples from each source are distributed differently. These differences can be inherent to the inference task or present for other reasons: sensors in a sensor network may be placed far apart, affecting their individual measurements. Conversely, it is computationally advantageous to split Bayesian inference tasks across subsets of data, but data need not be identically distributed across subsets. One principled way to fuse probability distributions is via the lens of optimal transport: the Wasserstein barycenter is a single distribution that summarizes a collection of input measures while respecting their geometry. However, computing the barycenter scales poorly and requires discretization of all input distributions and the barycenter itself.