credibility
Reddit's human content wins amid the AI flood
Reddit's human content wins amid the AI flood For Ines Tan there's one particular site she turns to again and again for advice - and that's Reddit. Tan, who works in communications, regularly jumps on the site for skincare advice, to view reactions to shows she watches, such as The Traitors, and for help planning her upcoming wedding in May. It's a very empathetic place, she says of Reddit. For my wedding, I've found help emotionally, logistically and inspiration-wise. Tan believes people are consulting the online discussion platform more as they're craving human interaction in the world of increasing AI slop.
- North America > United States (0.15)
- North America > Central America (0.15)
- Oceania > Australia (0.05)
- (11 more...)
- Leisure & Entertainment (1.00)
- Media > News (0.93)
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (0.48)
- North America > United States > Florida > Broward County (0.04)
- Europe > Italy > Trentino-Alto Adige/Südtirol > Trentino Province > Trento (0.04)
- Asia > Middle East > Israel (0.04)
- Banking & Finance (1.00)
- Health & Medicine > Therapeutic Area > Endocrinology > Diabetes (0.68)
Source Coverage and Citation Bias in LLM-based vs. Traditional Search Engines
Zhang, Peixian, Ye, Qiming, Peng, Zifan, Garimella, Kiran, Tyson, Gareth
LLM-based Search Engines (LLM-SEs) introduces a new paradigm for information seeking. Unlike Traditional Search Engines (TSEs) (e.g., Google), these systems summarize results, often providing limited citation transparency. The implications of this shift remain largely unexplored, yet raises key questions regarding trust and transparency. In this paper, we present a large-scale empirical study of LLM-SEs, analyzing 55,936 queries and the corresponding search results across six LLM-SEs and two TSEs. We confirm that LLM-SEs cites domain resources with greater diversity than TSEs. Indeed, 37% of domains are unique to LLM-SEs. However, certain risks still persist: LLM-SEs do not outperform TSEs in credibility, political neutrality and safety metrics. Finally, to understand the selection criteria of LLM-SEs, we perform a feature-based analysis to identify key factors influencing source choice. Our findings provide actionable insights for end users, website owners, and developers.
- Asia > China > Hong Kong (0.40)
- Asia > China > Guangdong Province > Guangzhou (0.05)
- North America > United States > New York > New York County > New York City (0.04)
- (4 more...)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.93)
- Information Technology > Security & Privacy (1.00)
- Media > News (0.67)
- Leisure & Entertainment (0.67)
- (2 more...)
VP-AutoTest: A Virtual-Physical Fusion Autonomous Driving Testing Platform
Cui, Yiming, Fang, Shiyu, Zhang, Jiarui, Huang, Yan, Xu, Chengkai, Zhu, Bing, Zhang, Hao, Hang, Peng, Sun, Jian
The rapid development of autonomous vehicles has led to a surge in testing demand. Traditional testing methods, such as virtual simulation, closed-course, and public road testing, face several challenges, including unrealistic vehicle states, limited testing capabilities, and high costs. These issues have prompted increasing interest in virtual-physical fusion testing. However, despite its potential, virtual-physical fusion testing still faces challenges, such as limited element types, narrow testing scope, and fixed evaluation metrics. To address these challenges, we propose the Virtual-Physical Testing Platform for Autonomous Vehicles (VP-AutoTest), which integrates over ten types of virtual and physical elements, including vehicles, pedestrians, and roadside infrastructure, to replicate the diversity of real-world traffic participants. The platform also supports both single-vehicle interaction and multi-vehicle cooperation testing, employing adversarial testing and parallel deduction to accelerate fault detection and explore algorithmic limits, while OBU and Redis communication enable seamless vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I) cooperation across all levels of cooperative automation. Furthermore, VP-AutoTest incorporates a multidimensional evaluation framework and AI-driven expert systems to conduct comprehensive performance assessment and defect diagnosis. Finally, by comparing virtual-physical fusion test results with real-world experiments, the platform performs credibility self-evaluation to ensure both the fidelity and efficiency of autonomous driving testing. Please refer to the website for the full testing functionalities on the autonomous driving public service platform OnSite:https://www.onsite.com.cn.
- Asia > China > Shanghai > Shanghai (0.04)
- North America > United States > Hawaii > Honolulu County > Honolulu (0.04)
- Transportation > Ground > Road (1.00)
- Information Technology (1.00)
Misalignment of LLM-Generated Personas with Human Perceptions in Low-Resource Settings
Prama, Tabia Tanzin, Danforth, Christopher M., Dodds, Peter Sheridan
Recent advances enable Large Language Models (LLMs) to generate AI personas, yet their lack of deep contextual, cultural, and emotional understanding poses a significant limitation. This study quantitatively compared human responses with those of eight LLM-generated social personas (e.g., Male, Female, Muslim, Political Supporter) within a low-resource environment like Bangladesh, using culturally specific questions. Results show human responses significantly outperform all LLMs in answering questions, and across all matrices of persona perception, with particularly large gaps in empathy and credibility. Furthermore, LLM-generated content exhibited a systematic bias along the lines of the ``Pollyanna Principle'', scoring measurably higher in positive sentiment ($Φ_{avg} = 5.99$ for LLMs vs. $5.60$ for Humans). These findings suggest that LLM personas do not accurately reflect the authentic experience of real people in resource-scarce environments. It is essential to validate LLM personas against real-world human data to ensure their alignment and reliability before deploying them in social science research.
- Asia > Bangladesh (0.29)
- North America > United States > Vermont > Chittenden County > Burlington (0.04)
- North America > United States > New Mexico > Santa Fe County > Santa Fe (0.04)
- Asia > China (0.04)
Building Resilient Information Ecosystems: Large LLM-Generated Dataset of Persuasion Attacks
Kao, Hsien-Te, Panasyuk, Aleksey, Bautista, Peter, Dupree, William, Ganberg, Gabriel, Beaubien, Jeffrey M., Cassani, Laura, Volkova, Svitlana
Organization's communication is essential for public trust, but the rise of generative AI models has introduced significant challenges by generating persuasive content that can form competing narratives with official messages from government and commercial organizations at speed and scale. This has left agencies in a reactive position, often unaware of how these models construct their persuasive strategies, making it more difficult to sustain communication effectiveness. In this paper, we introduce a large LLM-generated persuasion attack dataset, which includes 134,136 attacks generated by GPT-4, Gemma 2, and Llama 3.1 on agency news. These attacks span 23 persuasive techniques from SemEval 2023 Task 3, directed toward 972 press releases from ten agencies. The generated attacks come in two mediums, press release statements and social media posts, covering both long-form and short-form communication strategies. We analyzed the moral resonance of these persuasion attacks to understand their attack vectors. GPT-4's attacks mainly focus on Care, with Authority and Loyalty also playing a role. Gemma 2 emphasizes Care and Authority, while Llama 3.1 centers on Loyalty and Care. Analyzing LLM-generated persuasive attacks across models will enable proactive defense, allow to create the reputation armor for organizations, and propel the development of both effective and resilient communications in the information ecosystem.
- Press Release (0.55)
- Research Report (0.50)
- Government > Regional Government > North America Government > United States Government (1.00)
- Government > Military (1.00)
A Cross-Cultural Assessment of Human Ability to Detect LLM-Generated Fake News about South Africa
Schlippe, Tim, Wölfel, Matthias, Mabokela, Koena Ronny
This study investigates how cultural proximity affects the ability to detect AI-generated fake news by comparing South African participants with those from other nationalities. As large language models increasingly enable the creation of sophisticated fake news, understanding human detection capabilities becomes crucial, particularly across different cultural contexts. We conducted a survey where 89 participants (56 South Africans, 33 from other nationalities) evaluated 10 true South African news articles and 10 AI-generated fake versions. Results reveal an asymmetric pattern: South Africans demonstrated superior performance in detecting true news about their country (40% deviation from ideal rating) compared to other participants (52%), but performed worse at identifying fake news (62% vs. 55%). This difference may reflect South Africans' higher overall trust in news sources. Our analysis further shows that South Africans relied more on content knowledge and contextual understanding when judging credibility, while participants from other countries emphasised formal linguistic features such as grammar and structure. Overall, the deviation from ideal rating was similar between groups (51% vs. 53%), suggesting that cultural familiarity appears to aid verification of authentic information but may also introduce bias when evaluating fabricated content. These insights contribute to understanding cross-cultural dimensions of misinformation detection and inform strategies for combating AI-generated fake news in increasingly globalised information ecosystems where content crosses cultural and geographical boundaries.
Trust in foundation models and GenAI: A geographic perspective
McKenzie, Grant, Janowicz, Krzysztof, Kessler, Carsten
Large-scale pre-trained machine learning models have reshaped our understanding of artificial intelligence across numerous domains, including our own field of geography. As with any new technology, trust has taken on an important role in this discussion. In this chapter, we examine the multifaceted concept of trust in foundation models, particularly within a geographic context. As reliance on these models increases and they become relied upon for critical decision-making, trust, while essential, has become a fractured concept. Here we categorize trust into three types: epistemic trust in the training data, operational trust in the model's functionality, and interpersonal trust in the model developers. Each type of trust brings with it unique implications for geographic applications. Topics such as cultural context, data heterogeneity, and spatial relationships are fundamental to the spatial sciences and play an important role in developing trust. The chapter continues with a discussion of the challenges posed by different forms of biases, the importance of transparency and explainability, and ethical responsibilities in model development. Finally, the novel perspective of geographic information scientists is emphasized with a call for further transparency, bias mitigation, and regionally-informed policies. Simply put, this chapter aims to provide a conceptual starting point for researchers, practitioners, and policy-makers to better understand trust in (generative) GeoAI.
- Europe > Austria > Vienna (0.14)
- North America > Canada > Quebec > Montreal (0.14)
- Europe > Ukraine (0.04)
- (4 more...)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- Government (0.87)
- (4 more...)
Melania Trump Used as 'Window-Dressing' in Elaborate Memecoin Fraud, Legal Filing Claims
Melania Trump Used as'Window-Dressing' in Elaborate Memecoin Fraud, Legal Filing Claims The first lady of the United States became a pawn in an intricate memecoin scam that resulted in millions of dollars in losses, crypto investors have alleged. A cryptocurrency promoted in January by US first lady Melania Trump was part of a sophisticated fraud that "leveraged celebrity association and'borrowed fame' to sell legitimacy to unsuspecting investors," a new legal filing has alleged. In April, crypto investors brought a federal class action lawsuit against Benjamin Chow, cofounder of crypto exchange Meteora, and Hayden Davis, cofounder of crypto venture capital firm Kelsier Labs, among other defendants, accusing them of a multimillion-dollar fraud involving a single memecoin, $M3M3. Later, the plaintiffs filed an amended complaint, expanding the allegations to include racketeering activity. They claimed the pair had colluded to rig the market for $LIBRA, a coin promoted by Javier Milei, president of Argentina, which collapsed in value shortly after launch.
- South America > Argentina (0.55)
- North America > United States > Utah (0.05)
- North America > United States > New York (0.05)
- (3 more...)
- Law > Litigation (1.00)
- Government > Regional Government > North America Government > United States Government (1.00)
- Banking & Finance > Trading (1.00)
Assessing Web Search Credibility and Response Groundedness in Chat Assistants
Vykopal, Ivan, Pikuliak, Matúš, Ostermann, Simon, Šimko, Marián
Chat assistants increasingly integrate web search functionality, enabling them to retrieve and cite external sources. While this promises more reliable answers, it also raises the risk of amplifying misinformation from low-credibility sources. In this paper, we introduce a novel methodology for evaluating assistants' web search behavior, focusing on source credibility and the groundedness of responses with respect to cited sources. Using 100 claims across five misinformation-prone topics, we assess GPT-4o, GPT-5, Perplexity, and Qwen Chat. Our findings reveal differences between the assistants, with Perplexity achieving the highest source credibility, whereas GPT-4o exhibits elevated citation of non-credibility sources on sensitive topics. This work provides the first systematic comparison of commonly used chat assistants for fact-checking behavior, offering a foundation for evaluating AI systems in high-stakes information environments.
- Media > News (1.00)
- Law (1.00)
- Health & Medicine > Therapeutic Area > Infections and Infectious Diseases (1.00)
- (7 more...)
- Information Technology > Information Management > Search (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)