objectivity
Objectivity seen as key to screening AI weapons
The Defense Ministry has compiled guidelines on ensuring appropriate human involvement in the research and development of defense equipment using artificial intelligence. The guidelines are expected to cover R&D activities for equipment such as unmanned combat-support drones and ships, but ensuring objectivity and reliability remains a key challenge, since ministry officials are responsible for screening these activities. The effectiveness of the guidelines also depends on how much AI data related to intellectual property private-sector companies disclose during R&D programs.
Examining Linguistic Shifts in Academic Writing Before and After the Launch of ChatGPT: A Study on Preprint Papers
Bao, Tong, Zhao, Yi, Mao, Jin, Zhang, Chengzhi
Large Language Models (LLMs), such as ChatGPT, have prompted academic concerns about their impact on academic writing. Existing studies have primarily examined LLM usage in academic writing through quantitative approaches, such as word frequency statistics and probability-based analyses. However, few have systematically examined the potential impact of LLMs on the linguistic characteristics of academic writing. To address this gap, we conducted a large-scale analysis across 823,798 abstracts published in last decade from arXiv dataset. Through the linguistic analysis of features such as the frequency of LLM-preferred words, lexical complexity, syntactic complexity, cohesion, readability and sentiment, the results indicate a significant increase in the proportion of LLM-preferred words in abstracts, revealing the widespread influence of LLMs on academic writing. Additionally, we observed an increase in lexical complexity and sentiment in the abstracts, but a decrease in syntactic complexity, suggesting that LLMs introduce more new vocabulary and simplify sentence structure. However, the significant decrease in cohesion and readability indicates that abstracts have fewer connecting words and are becoming more difficult to read. Moreover, our analysis reveals that scholars with weaker English proficiency were more likely to use the LLMs for academic writing, and focused on improving the overall logic and fluency of the abstracts. Finally, at discipline level, we found that scholars in Computer Science showed more pronounced changes in writing style, while the changes in Mathematics were minimal.
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (0.70)
Laypeople's Attitudes Towards Fair, Affirmative, and Discriminatory Decision-Making Algorithms
Lima, Gabriel, Grgić-Hlača, Nina, Langer, Markus, Zou, Yixin
Affirmative algorithms have emerged as a potential answer to algorithmic discrimination, seeking to redress past harms and rectify the source of historical injustices. We present the results of two experiments ($N$$=$$1193$) capturing laypeople's perceptions of affirmative algorithms -- those which explicitly prioritize the historically marginalized -- in hiring and criminal justice. We contrast these opinions about affirmative algorithms with folk attitudes towards algorithms that prioritize the privileged (i.e., discriminatory) and systems that make decisions independently of demographic groups (i.e., fair). We find that people -- regardless of their political leaning and identity -- view fair algorithms favorably and denounce discriminatory systems. In contrast, we identify disagreements concerning affirmative algorithms: liberals and racial minorities rate affirmative systems as positively as their fair counterparts, whereas conservatives and those from the dominant racial group evaluate affirmative algorithms as negatively as discriminatory systems. We identify a source of these divisions: people have varying beliefs about who (if anyone) is marginalized, shaping their views of affirmative algorithms. We discuss the possibility of bridging these disagreements to bring people together towards affirmative algorithms.
- North America > Canada > Ontario > Toronto (0.14)
- Europe > Germany > Baden-Württemberg > Freiburg (0.04)
- North America > United States > Alaska (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Questionnaire & Opinion Survey (1.00)
A Proposed S.C.O.R.E. Evaluation Framework for Large Language Models : Safety, Consensus, Objectivity, Reproducibility and Explainability
Tan, Ting Fang, Elangovan, Kabilan, Ong, Jasmine, Shah, Nigam, Sung, Joseph, Wong, Tien Yin, Xue, Lan, Liu, Nan, Wang, Haibo, Kuo, Chang Fu, Chesterman, Simon, Yeong, Zee Kin, Ting, Daniel SW
A comprehensive qualitative evaluation framework for large language models (LLM) in healthcare that expands beyond traditional accuracy and quantitative metrics needed. We propose 5 key aspects for evaluation of LLMs: Safety, Consensus, Objectivity, Reproducibility and Explainability (S.C.O.R.E.). We suggest that S.C.O.R.E. may form the basis for an evaluation framework for future LLM-based models that are safe, reliable, trustworthy, and ethical for healthcare and clinical applications.
- North America > United States > California > San Francisco County > San Francisco (0.28)
- Asia > Singapore > Central Region > Singapore (0.05)
- North America > United States > Pennsylvania > Philadelphia County > Philadelphia (0.04)
- (3 more...)
HYBRINFOX at CheckThat! 2024 -- Task 2: Enriching BERT Models with the Expert System VAGO for Subjectivity Detection
Casanova, Morgane, Chanson, Julien, Icard, Benjamin, Faye, Géraud, Gadek, Guillaume, Gravier, Guillaume, Égré, Paul
This paper presents the HYBRINFOX method used to solve Task 2 of Subjectivity detection of the CLEF 2024 CheckThat! competition. The specificity of the method is to use a hybrid system, combining a RoBERTa model, fine-tuned for subjectivity detection, a frozen sentence-BERT (sBERT) model to capture semantics, and several scores calculated by the English version of the expert system VAGO, developed independently of this task to measure vagueness and subjectivity in texts based on the lexicon. In English, the HYBRINFOX method ranked 1st with a macro F1 score of 0.7442 on the evaluation data. For the other languages, the method used a translation step into English, producing more mixed results (ranking 1st in Multilingual and 2nd in Italian over the baseline, but under the baseline in Bulgarian, German, and Arabic). We explain the principles of our hybrid approach, and outline ways in which the method could be improved for other languages besides English.
- Europe > France > Auvergne-Rhône-Alpes > Isère > Grenoble (0.04)
- Asia (0.04)
- Europe > Switzerland (0.04)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Expert Systems (0.62)
Reinforcement Learning from Human Feedback: Whose Culture, Whose Values, Whose Perspectives?
Barman, Kristian González, Lohse, Simon, de Regt, Henk
This approach is partic ularly useful when designing AI systems for tasks where it is difficult to specify a precise reward function or when it is important to align the model's behaviour with certain human expectations and values. For instance, RLHF has notably improved language models for context - aware text generation (Ziegler et al. 2020) and taught robots to navigate cluttered environments (Henry et al. 2010) . RLHF is commonly employed in the later stages of fine - tuning models, particularly in the development of prominent Large Language Models (LLMs) like GPT - 3.5 or GPT - 4. Initially, these models undergo training using vast text corpora to grasp a broad range of language patterns and contexts. This foundational training is supplemented by task - specific fine - tuning, where the models are adjusted to excel in particular applications, such as understanding and generating dialogues. The refinement process is then furt her enhanced through RLHF.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- (9 more...)
- Law > Civil Rights & Constitutional Law (0.67)
- Information Technology > Security & Privacy (0.46)
- Leisure & Entertainment > Games (0.46)
- Health & Medicine > Epidemiology (0.46)
Drawing the Line: Deep Segmentation for Extracting Art from Ancient Etruscan Mirrors
Sterzinger, Rafael, Brenner, Simon, Sablatnig, Robert
Etruscan mirrors constitute a significant category within Etruscan art and, therefore, undergo systematic examinations to obtain insights into ancient times. A crucial aspect of their analysis involves the labor-intensive task of manually tracing engravings from the backside. Additionally, this task is inherently challenging due to the damage these mirrors have sustained, introducing subjectivity into the process. We address these challenges by automating the process through photometric-stereo scanning in conjunction with deep segmentation networks which, however, requires effective usage of the limited data at hand. We accomplish this by incorporating predictions on a per-patch level, and various data augmentations, as well as exploring self-supervised learning. Compared to our baseline, we improve predictive performance w.r.t. the pseudo-F-Measure by around 16%. When assessing performance on complete mirrors against a human baseline, our approach yields quantitative similar performance to a human annotator and significantly outperforms existing binarization methods. With our proposed methodology, we streamline the annotation process, enhance its objectivity, and reduce overall workload, offering a valuable contribution to the examination of these historical artifacts and other non-traditional documents.
Maintaining Journalistic Integrity in the Digital Age: A Comprehensive NLP Framework for Evaluating Online News Content
Bojic, Ljubisa, Prodanovic, Nikola, Samala, Agariadne Dwinggo
The rapid growth of online news platforms has led to an increased need for reliable methods to evaluate the quality and credibility of news articles. This paper proposes a comprehensive framework to analyze online news texts using natural language processing (NLP) techniques, particularly a language model specifically trained for this purpose, alongside other well-established NLP methods. The framework incorporates ten journalism standards-objectivity, balance and fairness, readability and clarity, sensationalism and clickbait, ethical considerations, public interest and value, source credibility, relevance and timeliness, factual accuracy, and attribution and transparency-to assess the quality of news articles. By establishing these standards, researchers, media organizations, and readers can better evaluate and understand the content they consume and produce. The proposed method has some limitations, such as potential difficulty in detecting subtle biases and the need for continuous updating of the language model to keep pace with evolving language patterns.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- Europe > Serbia > Vojvodina > South Bačka District > Novi Sad (0.04)
- North America > United States > New York (0.04)
- (8 more...)
How China's New AI Rules Could Affect U.S. Companies
Soon after China's artificial intelligence rules came into effect last month, a series of new AI chatbots began trickling onto the market, with government approval. The rules have already been watered down from what was initially proposed, and so far, China hasn't enforced them as strictly as it could, experts say. China's regulatory approach will likely have huge implications for the technological competition between the country and its AI superpower rival the U.S. The Cyberspace Administration of China's (CAC) Generative AI Measures, which came into effect on Aug. 15, are some of the strictest in the world. They state that the generative AI services should not generate content "inciting subversion of national sovereignty or the overturn of the socialist system," or "advocating terrorism or extremism, promoting ethnic hatred and ethnic discrimination, violence and obscenity, as well as fake and harmful information." Preventing AI chatbots from spewing out unwanted or even toxic content has been a challenge for AI developers around the world.
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.05)
- Asia > Middle East > Jordan (0.05)
- Asia > China > Beijing > Beijing (0.05)
- Law > Statutes (1.00)
- Government (1.00)
What Is AI Going To Do To Art? The History Of Photography Offers Clues.
Lois Rosson is a historian of science and technology based in Los Angeles. She is currently writing a book about images of outer space and their legibility. In 1835, William Henry Fox Talbot finally succeeded in producing a crude photograph of his country estate. He triumphantly declared that his was the first house ever known to have drawn its own picture. Fox Talbot described the calotype, his contribution to the photomechanical process, as an eradication of human intervention.
- North America > United States > California > Los Angeles County > Los Angeles (0.24)
- Europe > France (0.05)
- Law (1.00)
- Media > Photography (0.74)