Goto

Collaborating Authors

 springer nature


Disclosing Generative AI Use in Digital Humanities Research

Ma, Rongqian, Zhang, Xuhan, Wisnicki, Adrian

arXiv.org Artificial Intelligence

This survey study investigates how digital humanists perceive and approach generative AI disclosure in research. The results indicate that while digital humanities scholars acknowledge the importance of disclosing GenAI use, the actual rate of disclosure in research practice remains low. Respondents differ in their views on which activities most require disclosure and on the most appropriate methods for doing so. Most also believe that safeguards for AI disclosure should be established through institutional policies rather than left to individual decisions. The study's findings will offer empirical guidance to scholars, institutional leaders, funders, and other stakeholders responsible for shaping effective disclosure policies.


Inductive Biases for Zero-shot Systematic Generalization in Language-informed Reinforcement Learning

Dijujin, Negin Hashemi, Rohani, Seyed Roozbeh Razavi, Samiei, Mohammad Mahdi, Baghshah, Mahdieh Soleymani

arXiv.org Artificial Intelligence

Sample efficiency and systematic generalization are two long-standing challenges in reinforcement learning. Previous studies have shown that involving natural language along with other observation modalities can improve generalization and sample efficiency due to its compositional and open-ended nature. However, to transfer these properties of language to the decision-making process, it is necessary to establish a proper language grounding mechanism. One approach to this problem is applying inductive biases to extract fine-grained and informative representations from the observations, which makes them more connectable to the language units. We provide architecture-level inductive biases for modularity and sparsity mainly based on Neural Production Systems (NPS). Alongside NPS, we assign a central role to memory in our architecture. It can be seen as a high-level information aggregator which feeds policy/value heads with comprehensive information and simultaneously guides selective attention in NPS through attentional feedback. Our results in the BabyAI environment suggest that the proposed model's systematic generalization and sample efficiency are improved significantly compared to previous models. An extensive ablation study on variants of the proposed method is conducted, and the effectiveness of each employed technique on generalization, sample efficiency, and training stability is specified.


NeoPhysIx: An Ultra Fast 3D Physical Simulator as Development Tool for AI Algorithms

Fischer, Jörn, Ihme, Thomas

arXiv.org Artificial Intelligence

Traditional AI algorithms, such as Genetic Programming and Reinforcement Learning, often require extensive computational resources to simulate real-world physical scenarios effectively. While advancements in multi-core processing have been made, the inherent limitations of parallelizing rigid body dynamics lead to significant communication overheads, hindering substantial performance gains for simple simulations. This paper introduces NeoPhysIx, a novel 3D physical simulator designed to overcome these challenges. By adopting innovative simulation paradigms and focusing on essential algorithmic elements, NeoPhysIx achieves unprecedented speedups exceeding 1000x compared to real-time. This acceleration is realized through strategic simplifications, including point cloud collision detection, joint angle determination, and friction force estimation. The efficacy of NeoPhysIx is demonstrated through its application in training a legged robot with 18 degrees of freedom and six sensors, controlled by an evolved genetic program. Remarkably, simulating half a year of robot lifetime within a mere 9 hours on a single core of a standard mid-range CPU highlights the significant efficiency gains offered by NeoPhysIx. This breakthrough paves the way for accelerated AI development and training in physically-grounded domains.


Exploring ChatGPT and its Impact on Society

Haque, Md. Asraful, Li, Shuai

arXiv.org Artificial Intelligence

Artificial intelligence has been around for a while, but suddenly it has received more attention than ever before. Thanks to innovations from companies like Google, Microsoft, Meta, and other major brands in technology. OpenAI, though, has triggered the button with its ground-breaking invention ChatGPT. ChatGPT is a Large Language Model (LLM) based on Transformer architecture that has the ability to generate human-like responses in a conversational context. It uses deep learning algorithms to generate natural language responses to input text. Its large number of parameters, contextual generation, and open-domain training make it a versatile and effective tool for a wide range of applications, from chatbots to customer service to language translation. It has the potential to revolutionize various industries and transform the way we interact with technology. However, the use of ChatGPT has also raised several concerns, including ethical, social, and employment challenges, which must be carefully considered to ensure the responsible use of this technology. The article provides an overview of ChatGPT, delving into its architecture and training process. It highlights the potential impacts of ChatGPT on the society. In this paper, we suggest some approaches involving technology, regulation, education, and ethics in an effort to maximize ChatGPT's benefits while minimizing its negative impacts. This study is expected to contribute to a greater understanding of ChatGPT and aid in predicting the potential changes it may bring about.


Affective social anthropomorphic intelligent system

Mamun, Md. Adyelullahil, Abdullah, Hasnat Md., Alam, Md. Golam Rabiul, Hassan, Muhammad Mehedi, Uddin, Md. Zia

arXiv.org Artificial Intelligence

Human conversational styles are measured by the sense of humor, personality, and tone of voice. These characteristics have become essential for conversational intelligent virtual assistants. However, most of the state-of-the-art intelligent virtual assistants (IVAs) are failed to interpret the affective semantics of human voices. This research proposes an anthropomorphic intelligent system that can hold a proper human-like conversation with emotion and personality. A voice style transfer method is also proposed to map the attributes of a specific emotion. Initially, the frequency domain data (Mel-Spectrogram) is created by converting the temporal audio wave data, which comprises discrete patterns for audio features such as notes, pitch, rhythm, and melody. A collateral CNN-Transformer-Encoder is used to predict seven different affective states from voice. The voice is also fed parallelly to the deep-speech, an RNN model that generates the text transcription from the spectrogram. Then the transcripted text is transferred to the multi-domain conversation agent using blended skill talk, transformer-based retrieve-and-generate generation strategy, and beam-search decoding, and an appropriate textual response is generated. The system learns an invertible mapping of data to a latent space that can be manipulated and generates a Mel-spectrogram frame based on previous Mel-spectrogram frames to voice synthesize and style transfer. Finally, the waveform is generated using WaveGlow from the spectrogram. The outcomes of the studies we conducted on individual models were auspicious. Furthermore, users who interacted with the system provided positive feedback, demonstrating the system's effectiveness.


Efficient learning of large sets of locally optimal classification rules

Huynh, Van Quoc Phuong, Fürnkranz, Johannes, Beck, Florian

arXiv.org Artificial Intelligence

Conventional rule learning algorithms aim at finding a set of simple rules, where each rule covers as many examples as possible. In this paper, we argue that the rules found in this way may not be the optimal explanations for each of the examples they cover. Instead, we propose an efficient algorithm that aims at finding the best rule covering each training example in a greedy optimization consisting of one specialization and one generalization loop. These locally optimal rules are collected and then filtered for a final rule set, which is much larger than the sets learned by conventional rule learning algorithms. A new example is classified by selecting the best among the rules that cover this example. In our experiments on small to very large datasets, the approach's average classification accuracy is higher than that of state-of-the-art rule learning algorithms. Moreover, the algorithm is highly efficient and can inherently be processed in parallel without affecting the learned rule set and so the classification accuracy. We thus believe that it closes an important gap for large-scale classification rule induction.


Unknown area exploration for robots with energy constraints using a modified Butterfly Optimization Algorithm

Bendahmane, Amine, Tlemsani, Redouane

arXiv.org Artificial Intelligence

Butterfly Optimization Algorithm (BOA) is a recent metaheuristic that has been used in several optimization problems. In this paper, we propose a new version of the algorithm (xBOA) based on the crossover operator and compare its results to the original BOA and 3 other variants recently introduced in the literature. We also proposed a framework for solving the unknown area exploration problem with energy constraints using metaheuristics in both single- and multi-robot scenarios. This framework allowed us to benchmark the performances of different metaheuristics for the robotics exploration problem. We conducted several experiments to validate this framework and used it to compare the effectiveness of xBOA with wellknown metaheuristics used in the literature through 5 evaluation criteria. Although BOA and xBOA are not optimal in all these criteria, we found that BOA can be a good alternative to many metaheuristics in terms of the exploration time, while xBOA is more robust to local optima; has better fitness convergence; and achieves better exploration rates than the original BOA and its other variants.


This is how AI will change the way we treat spinal damage.

#artificialintelligence

This is how AI will change the way we treat spinal damage.. A quick overview of Machine learning strategies for the structure-property relationship of copolymers..


"Augmented Reality" Science-Research, January 2022, Week 3 -- summary from Springer Nature, Europe…

#artificialintelligence

In the last few years, there has been a progressive increase in clinical production on the application of Augmented Reality in students with Autism Spectrum Disorders. The results show that this is an area with numerous defined thematic lines: to start with, the incorporation of students with ASD into institution. As future lines of research, the opportunity of including new bibliometric software that makes it feasible to get even more bibliometric signs on the records is considered. History The aim of this research was to objectively compare clinical augmented reality glasses and standard monitors in video-assisted surgical treatment and to methodically evaluate its ergonomic benefits. When ARG was made use of contrasted to those with traditional screen, outcomes NASA-TLX ratings of 3 surgeons were reduced.


"Autonomous vehicle" Science-Research, January 2022, Week 1 -- summary from Springer Nature

#artificialintelligence

Recent advancement in the assumption of autonomous vehicles is generally stemmed by deep learning. There is constantly fantastic difficulty to processing deep learning formula on an ingrained system. When actuator mistakes happen, this post presents a unique control strategy based on dual-loop adaptive dynamic programs to enhance the tracking performance and make certain the safety and security of autonomous vehicles. Stereo-vision is one of the most widely utilized techniques in the advancement of ecological understanding systems for smart transportation. The primary need for the application of stereo vision on a vehicle is the processing time, which has to be extremely quick for autonomous driving in actual time, whereas the computation of the document of the pictures in the formula of stereo-vision calls for more computing power.