Gupta, Mansi
Do GFlowNets Transfer? Case Study on the Game of 24/42
Gupta, Adesh, Kumar, Abhinav, Gupta, Mansi, Chopra, Paras
Generating diverse solutions is key to human-like reasoning, yet autoregres-sive language models focus on single accurate responses, limiting creativity. Our case study shows their limited zero-shot transferability by fine-tuning small and medium-sized large language models on the Game of 24 and testing them on the Game of 42 datasets. Results revealed that GFlowNets struggle to maintain solution diversity and accuracy, highlighting key limitations in their cross-task generalization and the need for future research in improved transfer learning capabilities. Recent advances have introduced approaches showing significant improvement in LLM reasoning capabilities (Touvron et al., 2023a), including supervised fine-tuning with synthetic datasets (Y u et al.; Y ue et al.), modified decoding mechanisms (Holtzman et al.; Nguyen et al., 2024), and enhanced pretraining data quality (Akter et al., 2024; Trinh et al., 2024). While these approaches demonstrate improved accuracy, they rarely account for the diversity of correct solutions, an essential aspect of human-like reasoning and creativity (Y u et al., 2024a; Hu et al.).
Pruning as a Defense: Reducing Memorization in Large Language Models
Gupta, Mansi, Waghela, Nikhar, Gupta, Sarthak, Goel, Shourya, Shanmugavelu, Sanjif
Large language models have been shown to memorize significan t portions of their training data, which they can reproduce when appropriately prompted. This work investigates the impact of simple pruning techniques on thi s behavior. Our findings reveal that pruning effectively reduces the extent of m emorization in LLMs, demonstrating its potential as a foundational approach for mitigating membership inference attacks. Large language models are known to memorize portions of thei r training data, which poses significant privacy and security risks. Although various studies h ave explored the extent of memorization in LLMs, most of these efforts are qualitative (Carlini et al .
Are VLMs Really Blind
Singh, Ayush, Gupta, Mansi, Garg, Shivank
Vision Language Models excel in handling a wide range of complex tasks, including Optical Character Recognition (OCR), Visual Question Answering (VQA), and advanced geometric reasoning. However, these models fail to perform well on low-level basic visual tasks which are especially easy for humans. Our goal in this work was to determine if these models are truly "blind" to geometric reasoning or if there are ways to enhance their capabilities in this area. Our work presents a novel automatic pipeline designed to extract key information from images in response to specific questions. Instead of just relying on direct VQA, we use question-derived keywords to create a caption that highlights important details in the image related to the question. This caption is then used by a language model to provide a precise answer to the question without requiring external fine-tuning.
Beyond Captioning: Task-Specific Prompting for Improved VLM Performance in Mathematical Reasoning
Singh, Ayush, Gupta, Mansi, Garg, Shivank, Kumar, Abhinav, Agrawal, Vansh
Vision-Language Models (VLMs) have transformed tasks requiring visual and reasoning abilities, such as image retrieval and Visual Question Answering (VQA). Despite their success, VLMs face significant challenges with tasks involving geometric reasoning, algebraic problem-solving, and counting. These limitations stem from difficulties effectively integrating multiple modalities and accurately interpreting geometry-related tasks. Various works claim that introducing a captioning pipeline before VQA tasks enhances performance. We incorporated this pipeline for tasks involving geometry, algebra, and counting. We found that captioning results are not generalizable, specifically with larger VLMs primarily trained on downstream QnA tasks showing random performance on math-related challenges. However, we present a promising alternative: task-based prompting, enriching the prompt with task-specific guidance. This approach shows promise and proves more effective than direct captioning methods for math-heavy problems.
Confidence Is All You Need for MI Attacks
Sinha, Abhishek, Tibrewal, Himanshi, Gupta, Mansi, Waghela, Nikhar, Garg, Shivank
In this evolving era of machine learning security, membership inference attacks have emerged as a potent threat to the confidentiality of sensitive data. In this attack, adversaries aim to determine whether a particular point was used during the training of a target model. This paper proposes a new method to gauge a data point's membership in a model's training set. Instead of correlating loss with membership, as is traditionally done, we have leveraged the fact that training examples generally exhibit higher confidence values when classified into their actual class. During training, the model is essentially being 'fit' to the training data and might face particular difficulties in generalization to unseen data. This asymmetry leads to the model achieving higher confidence on the training data as it exploits the specific patterns and noise present in the training data. Our proposed approach leverages the confidence values generated by the machine learning model. These confidence values provide a probabilistic measure of the model's certainty in its predictions and can further be used to infer the membership of a given data point. Additionally, we also introduce another variant of our method that allows us to carry out this attack without knowing the ground truth(true class) of a given data point, thus offering an edge over existing label-dependent attack methods.
A Data-Centric Framework for Composable NLP Workflows
Liu, Zhengzhong, Ding, Guanxiong, Bukkittu, Avinash, Gupta, Mansi, Gao, Pengzhi, Ahmed, Atif, Zhang, Shikun, Gao, Xin, Singhavi, Swapnil, Li, Linwei, Wei, Wei, Hu, Zecong, Shi, Haoran, Liang, Xiaodan, Mitamura, Teruko, Xing, Eric P., Hu, Zhiting
Empirical natural language processing (NLP) systems in application domains (e.g., healthcare, finance, education) involve interoperation among multiple components, ranging from data ingestion, human annotation, to text retrieval, analysis, generation, and visualization. We establish a unified open-source framework to support fast development of such sophisticated NLP workflows in a composable manner. The framework introduces a uniform data representation to encode heterogeneous results by a wide range of NLP tasks. It offers a large repository of processors for NLP tasks, visualization, and annotation, which can be easily assembled with full interoperability under the unified representation. The highly extensible framework allows plugging in custom processors from external off-the-shelf NLP and deep learning libraries. The whole framework is delivered through two modularized yet integratable open-source projects, namely Forte1 (for workflow infrastructure and NLP function processors) and Stave2 (for user interaction, visualization, and annotation).