misconduct
- Asia > Singapore (0.29)
- Europe > Austria > Vienna (0.14)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- (7 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.92)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.92)
- (4 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
- Asia > Singapore (0.29)
- Europe > Austria > Vienna (0.14)
- North America > United States > Illinois > Champaign County > Urbana (0.04)
- (7 more...)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.92)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.92)
- (4 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Agents (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
LionGuard 2: Building Lightweight, Data-Efficient & Localised Multilingual Content Moderators
Tan, Leanne, Chua, Gabriel, Ge, Ziyu, Lee, Roy Ka-Wei
Modern moderation systems increasingly support multiple languages, but often fail to address localisation and low-resource variants - creating safety gaps in real-world deployments. Small models offer a potential alternative to large LLMs, yet still demand considerable data and compute. We present LionGuard 2, a lightweight, multilingual moderation classifier tailored to the Singapore context, supporting English, Chinese, Malay, and partial Tamil. Built on pre-trained OpenAI embeddings and a multi-head ordinal classifier, LionGuard 2 outperforms several commercial and open-source systems across 17 benchmarks, including both Singapore-specific and public English datasets. The system is actively deployed within the Singapore Government, demonstrating practical efficacy at scale. Our findings show that high-quality local data and robust multilingual embeddings can achieve strong moderation performance, without fine-tuning large models. We release our model weights and part of our training data to support future work on LLM safety.
- Asia > Singapore (1.00)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- (2 more...)
BMDetect: A Multimodal Deep Learning Framework for Comprehensive Biomedical Misconduct Detection
Zhou, Yize, Zhang, Jie, Wang, Meijie, Yu, Lun
Academic misconduct detection in biomedical research remains challenging due to algorithmic narrowness in existing methods and fragmented analytical pipelines. We present BMDetect, a multimodal deep learning framework that integrates journal metadata (SJR, institutional data), semantic embeddings (PubMedBERT), and GPT-4o-mined textual attributes (methodological statistics, data anomalies) for holistic manuscript evaluation. Key innovations include: (1) multimodal fusion of domain-specific features to reduce detection bias; (2) quantitative evaluation of feature importance, identifying journal authority metrics (e.g., SJR-index) and textual anomalies (e.g., statistical outliers) as dominant predictors; and (3) the BioMCD dataset, a large-scale benchmark with 13,160 retracted articles and 53,411 controls. BMDetect achieves 74.33% AUC, outperforming single-modality baselines by 8.6%, and demonstrates transferability across biomedical subfields. This work advances scalable, interpretable tools for safeguarding research integrity.
- North America > United States > California > San Francisco County > San Francisco (0.14)
- Asia > Japan (0.04)
- Asia > China > Hubei Province > Wuhan (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.68)
- Health & Medicine > Pharmaceuticals & Biotechnology (1.00)
- Education (1.00)
- Health & Medicine > Therapeutic Area (0.68)
RabakBench: Scaling Human Annotations to Construct Localized Multilingual Safety Benchmarks for Low-Resource Languages
Chua, Gabriel, Tan, Leanne, Ge, Ziyu, Lee, Roy Ka-Wei
Large language models (LLMs) and their safety classifiers often perform poorly on low-resource languages due to limited training data and evaluation benchmarks. This paper introduces RabakBench, a new multilingual safety benchmark localized to Singapore's unique linguistic context, covering Singlish, Chinese, Malay, and Tamil. RabakBench is constructed through a scalable three-stage pipeline: (i) Generate - adversarial example generation by augmenting real Singlish web content with LLM-driven red teaming; (ii) Label - semi-automated multi-label safety annotation using majority-voted LLM labelers aligned with human judgments; and (iii) Translate - high-fidelity translation preserving linguistic nuance and toxicity across languages. The final dataset comprises over 5,000 safety-labeled examples across four languages and six fine-grained safety categories with severity levels. Evaluations of 11 popular open-source and closed-source guardrail classifiers reveal significant performance degradation. RabakBench not only enables robust safety evaluation in Southeast Asian multilingual settings but also offers a reproducible framework for building localized safety datasets in low-resource environments. The benchmark dataset, including the human-verified translations, and evaluation code are publicly available.
- Asia > Singapore (0.49)
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.14)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- (9 more...)
- Law (1.00)
- Information Technology (0.93)
- Health & Medicine (0.70)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.47)
Revealed: Thousands of UK university students caught cheating using AI
Thousands of university students in the UK have been caught misusing ChatGPT and other artificial intelligence tools in recent years, while traditional forms of plagiarism show a marked decline, a Guardian investigation can reveal. A survey of academic integrity violations found almost 7,000 proven cases of cheating using AI tools in 2023-24, equivalent to 5.1 for every 1,000 students. That was up from 1.6 cases per 1,000 in 2022-23. Figures up to May suggest that number will increase again this year to about 7.5 proven cases per 1,000 students – but recorded cases represent only the tip of the iceberg, according to experts. The data highlights a rapidly evolving challenge for universities: trying to adapt assessment methods to the advent of technologies such as ChatGPT and other AI-powered writing tools.
- Europe > United Kingdom (0.25)
- North America > United States (0.05)
- North America > Canada (0.05)
- Information Technology > Artificial Intelligence > Applied AI (0.90)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.75)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (0.58)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.58)
AI can be a powerful tool for scientists. But it can also fuel research misconduct
In February this year, Google announced it was launching "a new AI system for scientists". It said this system was a collaborative tool designed to help scientists "in creating novel hypotheses and research plans". It's too early to tell just how useful this particular tool will be to scientists. But what is clear is that artificial intelligence (AI) more generally is already transforming science. Last year for example, computer scientists won the Nobel Prize for Chemistry for developing an AI model to predict the shape of every protein known to mankind.
- Personal > Honors (0.91)
- Research Report (0.75)
'Ghost stops': Lieutenant claims LAPD officials were warned about troubled gang unit
A Los Angeles police lieutenant has filed a legal claim against the city, alleging his superiors ignored his warnings about misconduct in an anti-gang unit until it became a public scandal, leading to him facing termination. The claim, which typically serves as the precursor to a lawsuit, was brought this month by Lt. Mark Garza. It's the first litigation being pursued by a former member the Mission Division gang unit, whose officers came under investigation last year over allegations they illegally stopped and searched vehicles and stole from people they pulled over. Garza, who was in charge of the unit, said he reported his suspicion in June 2023 that some of his officers were conducting "ghost stops," which meant their actions could go unnoticed because they didn't document the encounters or turn on their body-worn or dashboard cameras and never informed police dispatch of where they were. At that time, Garza said, the department's body camera policy required supervisors to review only footage related to "complaints, use of force and pursuits."
- North America > United States > California > Los Angeles County > Los Angeles (0.55)
- Europe > United Kingdom (0.05)
- Law > Litigation (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (1.00)
Former ByteDance Intern Accused of Sabotage Among Winners of Prestigious AI Award
A former ByteDance intern who was allegedly dismissed for professional misconduct, including sabotaging colleagues' work, was announced as a winner of one of the most prestigious annual awards for AI research this week. Keyu Tian, whose LinkedIn and Google Scholar pages list him as a master's student in computer science at Peking University, is the first author of one of two papers chosen Tuesday for the main "Best Paper Award" at the Neural Information Processing Systems (NeurIPS) conference, the largest gathering of machine learning researchers in the world. The paper, titled "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction," presents a new method for creating AI-generated images that Tian and four coauthors--all affiliated with either ByteDance or Peking University--claim is faster and more efficient than its predecessors. "The overall quality of the paper presentation, experimental validation and insights (scaling laws) give compelling reasons to experiment with this model," the NeurIPS Best Paper Award committee wrote in a statement. The committee's decision to grant the honor to Tian, whom ByteDance reportedly sued for over 1 million in damages last month, claiming deliberate sabotage of other company research projects, quickly became the focus of wider discussions online about how NeurIPS is run and the way top AI researchers evaluate the work of their colleagues.
DocETL: Agentic Query Rewriting and Evaluation for Complex Document Processing
Shankar, Shreya, Chambers, Tristan, Shah, Tarak, Parameswaran, Aditya G., Wu, Eugene
Analyzing unstructured data has been a persistent challenge in data processing. Large Language Models (LLMs) have shown promise in this regard, leading to recent proposals for declarative frameworks for LLM-powered processing of unstructured data. However, these frameworks focus on reducing cost when executing user-specified operations using LLMs, rather than improving accuracy, executing most operations as-is (in a single LLM call). This is problematic for complex tasks and data, where LLM outputs for user-defined operations are often inaccurate, even with optimized prompts. For example, an LLM may struggle to identify {\em all} instances of specific clauses, like force majeure or indemnification, in lengthy legal documents, requiring decomposition of the data, the task, or both. We present DocETL, a system that optimizes complex document processing pipelines, while accounting for LLM shortcomings. DocETL offers a declarative interface for users to define such pipelines and uses an agent-based approach to automatically optimize them, leveraging novel agent-based rewrites (that we call rewrite directives), as well as an optimization and evaluation framework. We introduce (i) logical rewriting of pipelines, tailored for LLM-based tasks, (ii) an agent-guided plan evaluation mechanism that synthesizes and orchestrates task-specific validation prompts, and (iii) an optimization algorithm that efficiently finds promising plans, considering the latencies of agent-based plan generation and evaluation. Our evaluation on four different unstructured document analysis tasks demonstrates that DocETL finds plans with outputs that are 25 to 80% more accurate than well-engineered baselines, addressing a critical gap in unstructured data analysis. DocETL is open-source at docetl.org, and as of November 2024, has amassed over 1.3k GitHub Stars, with users spanning a variety of domains.
- North America > United States > California (0.04)
- North America > United States > New York (0.04)
- Asia > Singapore (0.04)
- Asia > Indonesia > Bali (0.04)
- Law (1.00)
- Law Enforcement & Public Safety > Crime Prevention & Enforcement (0.93)
- Health & Medicine (0.92)