Goto

Collaborating Authors

 llava-1



DeepStack: DeeplyStackingVisualTokens isSurprisinglySimpleandEffectiveforLMMs

Neural Information Processing Systems

This inevitably introduces a tremendous memory andcompute overheadintotheLLMs, whichisparticularly significant when it comes to high-resolution images and multi-frame videos. Several previous works attempt to mitigate this issue by proposing various token compression strategies. A straightforward way is to reduce the number of tokens with spatial grouping [70, 47]. Instead of pooling vision tokens, a few work instead to concatenate local tokens along the feature dimension to preserve visual information [11, 48]. Moreover, other works seek more sophisticated token resampling, such as Q-Former [43], Perceiver [4]and Abstractor [8],etc.


VisualAnchorsAreStrongInformationAggregators ForMultimodalLargeLanguageModel

Neural Information Processing Systems

IntherealmofMultimodal LargeLanguage Models(MLLMs), vision-language connector plays acrucial role to link the pre-trained vision encoders with Large Language Models (LLMs). Despite itsimportance, thevision-language connector has been relatively less explored. In this study, we aim to propose a strong vision-language connector that enables MLLMs toachievehigh accuracywhile maintainlowcomputationcost.


CVQA: Culturally-diverseMultilingual VisualQuestionAnsweringBenchmark

Neural Information Processing Systems

Visual Question Answering (VQA) is an important task in multimodal AI, and it is often used to test the ability of vision-language models to understand and reason on knowledge present in both visual and textual data.



Bench LanguageBenchmark

Neural Information Processing Systems

Wefurther evaluated state-of-the-art models on this benchmark forthree vision-language tasks: image captioning, visual grounding, and visual question answering. Our work aims to significantly contribute to the development ofadvanced vision-language models inthefieldofremote sensing.


DetectionUsingCommonSenseReasoning

Neural Information Processing Systems

Explainability in artificial intelligence is crucial for restoring trust, particularly in areas like face forgery detection, where viewers often struggle to distinguish between real and fabricated content.


DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effective for LMMs

Neural Information Processing Systems

Most large multimodal models (LMMs) are implemented by feeding visual tokens as a sequence into the first layer of a large language model (LLM). The resulting architecture is simple but significantly increases computation and memory costs, as it has to handle a large number of additional tokens in its input layer. This paper presents a new architecture *DeepStack* for LMMs. Considering $N$ layers in the language and vision transformer of LMMs, we stack the visual tokens into $N$ groups and feed each group to its aligned transformer layer from bottom to top. Surprisingly, this simple method greatly enhances the power of LMMs to model interactions among visual tokens across layers but with minimal additional cost. We apply *DeepStack* to both language and vision transformer in LMMs, and validate the effectiveness of *DeepStack* LMMs with extensive empirical results. Using the same context length, our DeepStack 7B and 13B parameters surpass their counterparts by 2.7 and 2.9 on average across 9 benchmarks, respectively.


Adversarial Confusion Attack: Disrupting Multimodal Large Language Models

Hoscilowicz, Jakub, Janicki, Artur

arXiv.org Artificial Intelligence

We introduce the Adversarial Confusion Attack, a new class of threats against multimodal large language models (MLLMs). Unlike jailbreaks or targeted misclassification, the goal is to induce systematic disruption that makes the model generate incoherent or confidently incorrect outputs. Practical applications include embedding such adversarial images into websites to prevent MLLM-powered AI Agents from operating reliably. The proposed attack maximizes next-token entropy using a small ensemble of open-source MLLMs. In the white-box setting, we show that a single adversarial image can disrupt all models in the ensemble, both in the full-image and Adversarial CAPTCHA settings. Despite relying on a basic adversarial technique (PGD), the attack generates perturbations that transfer to both unseen open-source (e.g., Qwen3-VL) and proprietary (e.g., GPT-5.1)


Minimal neuron ablation triggers catastrophic collapse in the language core of Large Vision-Language Models

Lu, Cen, Tang, Yung-Chen, Cavallaro, Andrea

arXiv.org Artificial Intelligence

Large Vision-Language Models (LVLMs) have shown impressive multimodal understanding capabilities, yet their robustness is poorly understood. In this paper, we investigate the structural vulnerabilities of LVLMs to identify any critical neurons whose removal triggers catastrophic collapse. In this context, we propose CAN, a method to detect Consistently Activated Neurons and to locate critical neurons by progressive masking. Experiments on LLaVA-1.5-7b-hf and InstructBLIP-Vicuna-7b reveal that masking only a tiny portion of the language model's feed-forward networks (just as few as four neurons in extreme cases) suffices to trigger catastrophic collapse. Notably, critical neurons are predominantly localized in the language model rather than in the vision components, and the down-projection layer is a particularly vulnerable structure. We also observe a consistent two-stage collapse pattern: initial expressive degradation followed by sudden, complete collapse. Our findings provide important insights for safety research in LVLMs.