Goto

Collaborating Authors

 Jain, Devansh


PolygloToxicityPrompts: Multilingual Evaluation of Neural Toxic Degeneration in Large Language Models

arXiv.org Artificial Intelligence

Recent advances in large language models (LLMs) have led to their extensive global deployment, and ensuring their safety calls for comprehensive and multilingual toxicity evaluations. However, existing toxicity benchmarks are overwhelmingly focused on English, posing serious risks to deploying LLMs in other languages. We address this by introducing PolygloToxicityPrompts (PTP), the first large-scale multilingual toxicity evaluation benchmark of 425K naturally occurring prompts spanning 17 languages. We overcome the scarcity of naturally occurring toxicity in web-text and ensure coverage across languages with varying resources by automatically scraping over 100M web-text documents. Using PTP, we investigate research questions to study the impact of model size, prompt language, and instruction and preference-tuning methods on toxicity by benchmarking over 60 LLMs. Notably, we find that toxicity increases as language resources decrease or model size increases. Although instruction- and preference-tuning reduce toxicity, the choice of preference-tuning method does not have any significant impact. Our findings shed light on crucial shortcomings of LLM safeguarding and highlight areas for future research.


GlobalFlowNet: Video Stabilization using Deep Distilled Global Motion Estimates

arXiv.org Artificial Intelligence

Videos shot by laymen using hand-held cameras contain undesirable shaky motion. Estimating the global motion between successive frames, in a manner not influenced by moving objects, is central to many video stabilization techniques, but poses significant challenges. A large body of work uses 2D affine transformations or homography for the global motion. However, in this work, we introduce a more general representation scheme, which adapts any existing optical flow network to ignore the moving objects and obtain a spatially smooth approximation of the global motion between video frames. We achieve this by a knowledge distillation approach, where we first introduce a low pass filter module into the optical flow network to constrain the predicted optical flow to be spatially smooth. This becomes our student network, named as \textsc{GlobalFlowNet}. Then, using the original optical flow network as the teacher network, we train the student network using a robust loss function. Given a trained \textsc{GlobalFlowNet}, we stabilize videos using a two stage process. In the first stage, we correct the instability in affine parameters using a quadratic programming approach constrained by a user-specified cropping limit to control loss of field of view. In the second stage, we stabilize the video further by smoothing global motion parameters, expressed using a small number of discrete cosine transform coefficients. In extensive experiments on a variety of different videos, our technique outperforms state of the art techniques in terms of subjective quality and different quantitative measures of video stability. The source code is publicly available at \href{https://github.com/GlobalFlowNet/GlobalFlowNet}{https://github.com/GlobalFlowNet/GlobalFlowNet}


Distilling Hypernymy Relations from Language Models: On the Effectiveness of Zero-Shot Taxonomy Induction

arXiv.org Artificial Intelligence

Ushio et al., 2021; Paranjape et al., 2021), multiword expression (MWE) identification (Espinosa-In this paper, we analyze zero-shot taxonomy In this paper, we evaluate LMs on TL benchmarks Taxonomy learning (TL) is the task of arranging using prompt-based and sentence-scoring domain terminologies into hierarchical structures techniques, and find not only that they are competitive where terms are nodes and edges denote is-a (hypernymic) with common approaches proposed in the relationships (Hwang et al., 2012). Domainspecific literature (which are typically supervised and/or concept generalization is at the core of human reliant on external resources), but that they achieve cognition (Yu et al., 2015), and a key enabler SoTa results in certain domains.