Goto

Collaborating Authors

 sharma


Top safety researcher issues shock resignation from major tech firm, warning 'world is in peril'

Daily Mail - Science & tech

Doctor at Jeffrey Epstein's post-mortem says the paedophile was strangled and NOT hanged Married California coffee growing pioneers die in'tragic accident' leaving their three children orphaned Lindsey Vonn's'primary goal is to keep her leg': Knee specialist warns Winter Olympics legend could face amputation or'lifelong consequences' after'motorcycle-style' skiing crash Disturbing Kurt Cobain autopsy details revealed for first time: As new probe claims Nirvana singer's death was a HOMICIDE, here's the evidence that convinced forensic investigators HGTV star was canned after vile video leak... now insiders make bombshell claims about what else she was doing behind the scenes: 'Unhinged' Gwyneth Paltrow's'nepo baby' Apple Martin, 21, reveals what cosmetic procedures she has had done Nashville's hottest couple engulfed by cheating storm as insiders declare: 'It's over' Jill Zarin's replacement REVEALED after racist Bad Bunny rant got her sacked from her TV comeback Scientists are ...


AI researcher says 'world is in peril' and quits to study poetry

BBC News

AI researcher says'world is in peril' and quits to study poetry An AI safety researcher has quit US firm Anthropic with a cryptic warning that the world is in peril. In his resignation letter shared on X, Mrinank Sharma told the firm he was leaving amid concerns about AI, bioweapons and the state of the wider world. He said he would instead look to pursue writing and studying poetry, and move back to the UK to become invisible. It comes in the same week that a OpenAI researcher said she had resigned, sharing concerns about the ChatGPT maker's decision to deploy adverts in its chatbot . Anthropic, best known for its Claude chatbot, had released a series of commercials aimed at OpenAI, criticising the company's move to include adverts for some users.


Chain of Unit-Physics: A Primitive-Centric Approach to Scientific Code Synthesis

Sharma, Vansh, Raman, Venkat

arXiv.org Artificial Intelligence

Agentic large language models are proposed as autonomous code generators for scientific computing, yet their reliability in high-stakes problems remains unclear. Developing computational scientific software from natural-language queries remains challenging broadly due to (a) sparse representation of domain codes during training and (b) the limited feasibility of RLHF with a small expert community. To address these limitations, this work conceptualizes an inverse approach to code design, embodied in the Chain of Unit-Physics framework: a first-principles (or primitives)-centric, multi-agent system in which human expert knowledge is encoded as unit-physics tests that explicitly constrain code generation. The framework is evaluated on a nontrivial combustion task, used here as a representative benchmark for scientific problem with realistic physical constraints. Closed-weight systems and code-focused agentic variants fail to produce correct end-to-end solvers, despite tool and web access, exhibiting four recurrent error classes: interface (syntax/API) hallucinations, overconfident assumptions, numerical/physical incoherence, and configuration fragility. Open-weight models with chain-of-thought (CoT) decoding reduce interface errors but still yield incorrect solutions. On the benchmark task, the proposed framework converges within 5-6 iterations, matches the human-expert implementation (mean error of $3.1\times10^{-3}$ %), with a $\sim$33.4 % faster runtime and a $\sim$30 % efficient memory usage at a cost comparable to mid-sized commercial APIs, yielding a practical template for physics-grounded scientific code generation. As datasets and models evolve, zero-shot code accuracy will improve; however, the Chain of Unit-Physics framework goes further by embedding first-principles analysis that is foundational to scientific codes.


The Indian woman who stood up to moral policing - and won a pageant

BBC News

Muskan Sharma stood up to men who tried to bully her over her clothes - and went on to win hearts and a beauty pageant. The 23-year-old, who was crowned Miss Rishikesh 2025 last week in the northern Indian state of Uttarakhand, told the BBC that even though it was a small local pageant, it made me feel like Miss Universe. Sharma's win has made headlines in India as it came after a viral video that showed her spiritedly arguing with a man who barged into their rehearsals just a day before the 4 October contest. Sharma, who wanted to be a model and participate in a pageant since I was in school, said the intruders came in just as they broke for lunch. We were sitting around, chilling, having a laugh when they walked in, she said.


When and Where do Events Switch in Multi-Event Video Generation?

Liao, Ruotong, Huang, Guowen, Cheng, Qing, Seidl, Thomas, Cremers, Daniel, Tresp, Volker

arXiv.org Artificial Intelligence

Text-to-video (T2V) generation has surged in response to challenging questions, especially when a long video must depict multiple sequential events with temporal coherence and controllable content. Existing methods that extend to multi-event generation omit an inspection of the intrinsic factor in event shifting. The paper aims to answer the central question: When and where multi-event prompts control event transition during T2V generation. This work introduces MEve, a self-curated prompt suite for evaluating multi-event text-to-video (T2V) generation, and conducts a systematic study of two representative model families, i.e., OpenSora and CogVideoX. Extensive experiments demonstrate the importance of early intervention in denoising steps and block-wise model layers, revealing the essential factor for multi-event video generation and highlighting the possibilities for multi-event conditioning in future models.


Decoding Memes: Benchmarking Narrative Role Classification across Multilingual and Multimodal Models

Sharma, Shivam, Chakraborty, Tanmoy

arXiv.org Artificial Intelligence

Abstract--This work investigates the challenging task of identifying narrative roles - Hero, Villain, Victim, and Other - in Internet memes, across three diverse test sets spanning English and code-mixed (English-Hindi) languages. Building on an annotated dataset originally skewed toward the'Other' class, we explore a more balanced and linguistically diverse extension, originally introduced as part of the CLEF 2024 shared task. Comprehensive lexical and structural analyses highlight the nuanced, culture-specific, and context-rich language used in real memes, in contrast to synthetically curated hateful content, which exhibits explicit and repetitive lexical markers. T o benchmark the role detection task, we evaluate a wide spectrum of models, including fine-tuned multilingual transformers, sentiment and abuse-aware classifiers, instruction-tuned LLMs, and multimodal vision-language models. Performance is assessed under zero-shot settings using precision, recall, and F1 metrics. W e also explore prompt design strategies to guide multi-modal models and find that hybrid prompts incorporating structured instructions and role definitions offer marginal yet consistent improvements. Our findings underscore the importance of cultural grounding, prompt engineering, and multimodal reasoning in modelling subtle narrative framings in visual-textual content. W arning: This paper contains potentially harmful and offensive content. I. Introduction Social media platforms have become pivotal arenas for rapid information dissemination. However, this openness has also catalysed the proliferation of harmful content - including hate speech, propaganda, and misinformation, often embedded within memes [1], [2]. Memes, with their multimodal structure and cultural resonance, are particularly potent in shaping public opinion and propagating ideologies.


Towards Physics-Guided Foundation Models

Farhadloo, Majid, Sharma, Arun, Yang, Mingzhou, Jayaprakash, Bharat, Northrop, William, Shekhar, Shashi

arXiv.org Artificial Intelligence

Traditional foundation models are pre-trained on broad datasets to reduce the training resources (e.g., time, energy, labeled samples) needed for fine-tuning a wide range of downstream tasks. However, traditional foundation models struggle with out-of-distribution prediction and can produce outputs that are unrealistic and physically infeasible. We propose the notation of physics-guided foundation models (PGFM), that is, foundation models integrated with broad or general domain (e.g., scientific) physical knowledge applicable to a wide range of downstream tasks.


PromptPex: Automatic Test Generation for Language Model Prompts

Sharma, Reshabh K, De Halleux, Jonathan, Barke, Shraddha, Zorn, Benjamin

arXiv.org Artificial Intelligence

Large language models (LLMs) are being used in many applications and prompts for these models are integrated into software applications as code-like artifacts. These prompts behave much like traditional software in that they take inputs, generate outputs, and perform some specific function. However, prompts differ from traditional code in many ways and require new approaches to ensure that they are robust. For example, unlike traditional software the output of a prompt depends on the AI model that interprets it. Also, while natural language prompts are easy to modify, the impact of updates is harder to predict. New approaches to testing, debugging, and modifying prompts with respect to the model running them are required. To address some of these issues, we developed PromptPex, an LLM-based tool to automatically generate and evaluate unit tests for a given prompt. PromptPex extracts input and output specifications from a prompt and uses them to generate diverse, targeted, and valid unit tests. These tests are instrumental in identifying regressions when a prompt is changed and also serve as a tool to understand how prompts are interpreted by different models. We use PromptPex to generate tests for eight benchmark prompts and evaluate the quality of the generated tests by seeing if they can cause each of four diverse models to produce invalid output. PromptPex consistently creates tests that result in more invalid model outputs than a carefully constructed baseline LLM-based test generator. Furthermore, by extracting concrete specifications from the input prompt, PromptPex allows prompt writers to clearly understand and test specific aspects of their prompts. The source code of PromptPex is available at https://github.com/microsoft/promptpex.


AI abortion training has arrived: New tech tools navigate blurry line between healthcare and politics

FOX News

Artificial intelligence (AI) tools are now available for future medical professionals at one Texas university to navigate the complexities of pregnancy and abortion--a development that further blurs the line between technology, politics and healthcare. A group of medical students at the University of Texas Medical Branch in Galveston recently created a simulation of a pregnant patient, powered by AI, that the next generation of health experts can use to interpret various maternal health situations, including abortion. The new tech allows users to engage in all-options pregnancy counseling in Texas while also avoiding the potential consequences of the state's abortion restrictions. Anu Sharma, the CEO and founder of a tech-enabled maternity care company called Millie, told Fox News Digital that while this kind of tech is not without controversy or political discourse, it could provide much needed innovation and help to a healthcare system with significant gaps. Texas medical students have developed new AI tools to assist women with different pregnancy options, including abortion.


GAME-ON: Graph Attention Network based Multimodal Fusion for Fake News Detection

Dhawan, Mudit, Sharma, Shakshi, Kadam, Aditya, Sharma, Rajesh, Kumaraguru, Ponnurangam

arXiv.org Artificial Intelligence

Social media in present times has a significant and growing influence. Fake news being spread on these platforms have a disruptive and damaging impact on our lives. Furthermore, as multimedia content improves the visibility of posts more than text data, it has been observed that often multimedia is being used for creating fake content. A plethora of previous multimodal-based work has tried to address the problem of modeling heterogeneous modalities in identifying fake content. However, these works have the following limitations: (1) inefficient encoding of inter-modal relations by utilizing a simple concatenation operator on the modalities at a later stage in a model, which might result in information loss; (2) training very deep neural networks with a disproportionate number of parameters on small but complex real-life multimodal datasets result in higher chances of overfitting. To address these limitations, we propose GAME-ON, a Graph Neural Network based end-to-end trainable framework that allows granular interactions within and across different modalities to learn more robust data representations for multimodal fake news detection. We use two publicly available fake news datasets, Twitter and Weibo, for evaluations. Our model outperforms on Twitter by an average of 11% and keeps competitive performance on Weibo, within a 2.6% margin, while using 65% fewer parameters than the best comparable state-of-the-art baseline.