Goto

Collaborating Authors

 dude



DUDE: Diffusion-Based Unsupervised Cross-Domain Image Retrieval

Yang, Ruohong, Hu, Peng, Li, Yunfan, Peng, Xi

arXiv.org Artificial Intelligence

Unsupervised cross-domain image retrieval (UCIR) aims to retrieve images of the same category across diverse domains without relying on annotations. Existing UCIR methods, which align cross-domain features for the entire image, often struggle with the domain gap, as the object features critical for retrieval are frequently entangled with domain-specific styles. To address this challenge, we propose DUDE, a novel UCIR method building upon feature disentanglement. In brief, DUDE leverages a text-to-image generative model to disentangle object features from domain-specific styles, thus facilitating semantical image retrieval. To further achieve reliable alignment of the disentangled object features, DUDE aligns mutual neighbors from within domains to across domains in a progressive manner. Extensive experiments demonstrate that DUDE achieves state-of-the-art performance across three benchmark datasets over 13 domains. The code will be released.


Dual Decomposition of Weights and Singular Value Low Rank Adaptation

Han, Jialong, Zhang, Si, Zhang, Ke

arXiv.org Artificial Intelligence

Parameter-Efficient Fine-Tuning (PEFT) has emerged as a critical paradigm for adapting Large Language Models (LLMs) to downstream tasks, among which Low-rank Adaptation (LoRA) represents one of the most widely adopted methodologies. However, existing LoRA-based approaches exhibit two fundamental limitations: unstable training dynamics and inefficient knowledge transfer from pre-trained models, both stemming from random initialization of adapter parameters. To overcome these challenges, we propose DuDe, a novel approach that decomposes weight matrices into magnitude and direction components, employing Singular Value Decomposition (SVD) for principled initialization. Our comprehensive evaluation demonstrates DuDe's superior performance and robustness, achieving up to 48.35\% accuracy on MMLU and 62.53\% ($\pm$ 1.59) accuracy on GSM8K. Our theoretical analysis and empirical validation collectively demonstrate that DuDe's decomposition strategy enhances optimization stability and better preserves pre-trained representations, particularly for domain-specific tasks requiring specialized knowledge. The combination of robust empirical performance and rigorous theoretical foundations establishes DuDe as a significant contribution to PEFT methodologies for LLMs.


Lexicalization Is All You Need: Examining the Impact of Lexical Knowledge in a Compositional QALD System

Schmidt, David Maria, Elahi, Mohammad Fazleh, Cimiano, Philipp

arXiv.org Artificial Intelligence

In this paper, we examine the impact of lexicalization on Question Answering over Linked Data (QALD). It is well known that one of the key challenges in interpreting natural language questions with respect to SPARQL lies in bridging the lexical gap, that is mapping the words in the query to the correct vocabulary elements. We argue in this paper that lexicalization, that is explicit knowledge about the potential interpretations of a word with respect to the given vocabulary, significantly eases the task and increases the performance of QA systems. Towards this goal, we present a compositional QA system that can leverage explicit lexical knowledge in a compositional manner to infer the meaning of a question in terms of a SPARQL query. We show that such a system, given lexical knowledge, has a performance well beyond current QA systems, achieving up to a $35.8\%$ increase in the micro $F_1$ score compared to the best QA system on QALD-9. This shows the importance and potential of including explicit lexical knowledge. In contrast, we show that LLMs have limited abilities to exploit lexical knowledge, with only marginal improvements compared to a version without lexical knowledge. This shows that LLMs have no ability to compositionally interpret a question on the basis of the meaning of its parts, a key feature of compositional approaches. Taken together, our work shows new avenues for QALD research, emphasizing the importance of lexicalization and compositionality.


Neural Universal Discrete Denoiser

Neural Information Processing Systems

We present a new framework of applying deep neural networks (DNN) to devise a universal discrete denoiser. Unlike other approaches that utilize supervised learning for denoising, we do not require any additional training data. In such setting, while the ground-truth label, i.e., the clean data, is not available, we devise "pseudolabels" and a novel objective function such that DNN can be trained in a same way as supervised learning to become a discrete denoiser. We experimentally show that our resulting algorithm, dubbed as Neural DUDE, significantly outperforms the previous state-of-the-art in several applications with a systematic rule of choosing the hyperparameter, which is an attractive feature in practice.


Document Understanding Dataset and Evaluation (DUDE)

Van Landeghem, Jordy, Tito, Rubén, Borchmann, Łukasz, Pietruszka, Michał, Józiak, Paweł, Powalski, Rafał, Jurkiewicz, Dawid, Coustaty, Mickaël, Ackaert, Bertrand, Valveny, Ernest, Blaschko, Matthew, Moens, Sien, Stanisławek, Tomasz

arXiv.org Artificial Intelligence

We call on the Document AI (DocAI) community to reevaluate current methodologies and embrace the challenge of creating more practically-oriented benchmarks. Document Understanding Dataset and Evaluation (DUDE) seeks to remediate the halted research progress in understanding visually-rich documents (VRDs). We present a new dataset with novelties related to types of questions, answers, and document layouts based on multi-industry, multi-domain, and multi-page VRDs of various origins, and dates. Moreover, we are pushing the boundaries of current methods by creating multi-task and multi-domain evaluation setups that more accurately simulate real-world situations where powerful generalization and adaptation under low-resource settings are desired. DUDE aims to set a new standard as a more practical, long-standing benchmark for the community, and we hope that it will lead to future extensions and contributions that address real-world challenges. Finally, our work illustrates the importance of finding more efficient ways to model language, images, and layout in DocAI.


Snoop Dogg addresses risks of artificial intelligence: 'Sh-- what the f---'

FOX News

American rapper Snoop Dogg expressed confusion about recent developments in artificial intelligence, comparing the technology to movies he saw as a child. At the Milken Institute Global Conference in Beverly Hills this week, Snoop, whose given name is Calvin Broadus, turned his focus to artificial intelligence while discussing a strike of the Writers Guild of America. The writers strike is, in part, about the potential for artificial intelligence to take writing jobs. "I got a motherf---ing AI right now that they did made for me," Snoop said. "This n----- could talk to me. I'm like, man, this thing can hold a real conversation? Like it's blowing my mind because I watched movies on this as a kid years ago."


DUDES: Deep Uncertainty Distillation using Ensembles for Semantic Segmentation

Landgraf, Steven, Wursthorn, Kira, Hillemann, Markus, Ulrich, Markus

arXiv.org Artificial Intelligence

Deep neural networks lack interpretability and tend to be overconfident, which poses a serious problem in safety-critical applications like autonomous driving, medical imaging, or machine vision tasks with high demands on reliability. Quantifying the predictive uncertainty is a promising endeavour to open up the use of deep neural networks for such applications. Unfortunately, current available methods are computationally expensive. In this work, we present a novel approach for efficient and reliable uncertainty estimation which we call Deep Uncertainty Distillation using Ensembles for Segmentation (DUDES). DUDES applies student-teacher distillation with a Deep Ensemble to accurately approximate predictive uncertainties with a single forward pass while maintaining simplicity and adaptability. Experimentally, DUDES accurately captures predictive uncertainties without sacrificing performance on the segmentation task and indicates impressive capabilities of identifying wrongly classified pixels and out-of-domain samples on the Cityscapes dataset. With DUDES, we manage to simultaneously simplify and outperform previous work on Deep Ensemble-based Uncertainty Distillation.



How Apple Uses Big Data And AI To Build The Future - HData Systems

#artificialintelligence

Imagine for a moment, if you will, that you were a "dude" from the 1970s. And when I mean "The Dude" I really mean the "Dude". I mean the dude who is so much like the dude that it will make the protagonist of Big Lebowski shiver. And now think, if you had a friend called "Walter". And he says, "Dude, there is like this big company in like… 10 years. It'll sell computers and stuff and will revolutionize the tech industry."