krauss
Author-Specific Linguistic Patterns Unveiled: A Deep Learning Study on Word Class Distributions
Krauss, Patrick, Schilling, Achim
Deep learning methods have been increasingly applied to computational linguistics to uncover patterns in text data. This study investigates author-specific word class distributions using part-of-speech (POS) tagging and bigram analysis. By leveraging deep neural networks, we classify literary authors based on POS tag vectors and bigram frequency matrices derived from their works. We employ fully connected and convolutional neural network architectures to explore the efficacy of unigram and bigram-based representations. Our results demonstrate that while unigram features achieve moderate classification accuracy, bigram-based models significantly improve performance, suggesting that sequential word class patterns are more distinctive of authorial style. Multi-dimensional scaling (MDS) visualizations reveal meaningful clustering of authors' works, supporting the hypothesis that stylistic nuances can be captured through computational methods. These findings highlight the potential of deep learning and linguistic feature analysis for author profiling and literary studies.
Exploring Narrative Clustering in Large Language Models: A Layerwise Analysis of BERT
Banerjee, Awritrojit, Schilling, Achim, Krauss, Patrick
This study investigates the internal mechanisms of BERT, a transformer-based large language model, with a focus on its ability to cluster narrative content and authorial style across its layers. Using a dataset of narratives developed via GPT-4, featuring diverse semantic content and stylistic variations, we analyze BERT's layerwise activations to uncover patterns of localized neural processing. Through dimensionality reduction techniques such as Principal Component Analysis (PCA) and Multidimensional Scaling (MDS), we reveal that BERT exhibits strong clustering based on narrative content in its later layers, with progressively compact and distinct clusters. While strong stylistic clustering might occur when narratives are rephrased into different text types (e.g., fables, sci-fi, kids' stories), minimal clustering is observed for authorial style specific to individual writers. These findings highlight BERT's prioritization of semantic content over stylistic features, offering insights into its representational capabilities and processing hierarchy. This study contributes to understanding how transformer models like BERT encode linguistic information, paving the way for future interdisciplinary research in artificial intelligence and cognitive neuroscience.
Analysis of Argument Structure Constructions in the Large Language Model BERT
Ramezani, Pegah, Schilling, Achim, Krauss, Patrick
This study investigates how BERT processes and represents Argument Structure Constructions (ASCs), extending previous LSTM analyses. Using a dataset of 2000 sentences across four ASC types (transitive, ditransitive, caused-motion, resultative), we analyzed BERT's token embeddings across 12 layers. Visualizations with MDS and t-SNE and clustering quantified by Generalized Discrimination Value (GDV) were used. Feedforward classifiers (probes) predicted construction categories from embeddings. CLS token embeddings clustered best in layers 2-4, decreased in intermediate layers, and slightly increased in final layers. DET and SUBJ embeddings showed consistent clustering in intermediate layers, VERB embeddings increased in clustering from layer 1 to 12, and OBJ embeddings peaked in layer 10. Probe accuracies indicated low construction information in layer 1, with over 90 percent accuracy from layer 2 onward, revealing latent construction information beyond GDV clustering. Fisher Discriminant Ratio (FDR) analysis of attention weights showed OBJ tokens were crucial for differentiating ASCs, followed by VERB and DET tokens. SUBJ, CLS, and SEP tokens had insignificant FDR scores. This study highlights BERT's layered processing of linguistic constructions and its differences from LSTMs. Future research will compare these findings with neuroimaging data to understand the neural correlates of ASC processing. This research underscores neural language models' potential to mirror linguistic processing in the human brain, offering insights into the computational and neural mechanisms underlying language understanding.
- Europe > Germany > Bavaria > Middle Franconia > Nuremberg (0.04)
- North America > United States > New York (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- (4 more...)
Analysis of Argument Structure Constructions in a Deep Recurrent Language Model
Ramezani, Pegah, Schilling, Achim, Krauss, Patrick
Understanding how language and linguistic constructions are processed in the brain is a fundamental question in cognitive computational neuroscience. In this study, we explore the representation and processing of Argument Structure Constructions (ASCs) in a recurrent neural language model. We trained a Long Short-Term Memory (LSTM) network on a custom-made dataset consisting of 2000 sentences, generated using GPT-4, representing four distinct ASCs: transitive, ditransitive, caused-motion, and resultative constructions. We analyzed the internal activations of the LSTM model's hidden layers using Multidimensional Scaling (MDS) and t-Distributed Stochastic Neighbor Embedding (t-SNE) to visualize the sentence representations. The Generalized Discrimination Value (GDV) was calculated to quantify the degree of clustering within these representations. Our results show that sentence representations form distinct clusters corresponding to the four ASCs across all hidden layers, with the most pronounced clustering observed in the last hidden layer before the output layer. This indicates that even a relatively simple, brain-constrained recurrent neural network can effectively differentiate between various construction types. These findings are consistent with previous studies demonstrating the emergence of word class and syntax rule representations in recurrent language models trained on next word prediction tasks. In future work, we aim to validate these results using larger language models and compare them with neuroimaging data obtained during continuous speech perception. This study highlights the potential of recurrent neural language models to mirror linguistic processing in the human brain, providing valuable insights into the computational and neural mechanisms underlying language understanding.
- Europe > Germany > Bavaria > Middle Franconia > Nuremberg (0.04)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- (2 more...)
Analyzing Narrative Processing in Large Language Models (LLMs): Using GPT4 to test BERT
Krauss, Patrick, Hösch, Jannik, Metzner, Claus, Maier, Andreas, Uhrig, Peter, Schilling, Achim
The ability to transmit and receive complex information via language is unique to humans and is the basis of traditions, culture and versatile social interactions. Through the disruptive introduction of transformer based large language models (LLMs) humans are not the only entity to "understand" and produce language any more. In the present study, we have performed the first steps to use LLMs as a model to understand fundamental mechanisms of language processing in neural networks, in order to make predictions and generate hypotheses on how the human brain does language processing. Thus, we have used ChatGPT to generate seven different stylistic variations of ten different narratives (Aesop's fables). We used these stories as input for the open source LLM BERT and have analyzed the activation patterns of the hidden units of BERT using multi-dimensional scaling and cluster analysis. We found that the activation vectors of the hidden units cluster according to stylistic variations in earlier layers of BERT (1) than narrative content (4-5). Despite the fact that BERT consists of 12 identical building blocks that are stacked and trained on large text corpora, the different layers perform different tasks. This is a very useful model of the human brain, where self-similar structures, i.e. different areas of the cerebral cortex, can have different functions and are therefore well suited to processing language in a very efficient way. The proposed approach has the potential to open the black box of LLMs on the one hand, and might be a further step to unravel the neural processes underlying human language processing and cognition in general.
Use of a low-cost forward-looking sonar for collision avoidance in small AUVs, analysis and experimental results
Morency, Christopher, Stilwell, Daniel J., Krauss, Stephen T.
In this paper, we seek to evaluate the effectiveness of a novel forward-looking sonar system with a limited number of beams for collision avoidance for small autonomous underwater vehicles (AUVs). We present a collision avoidance strategy specifically designed for a novel forward-looking sonar system based on posterior expected loss, explicitly coupling the obstacle detection, collision avoidance, and planning. We demonstrate the strategy with field trials using the 690 AUV, built by the Center for Marine Autonomy and Robotics at Virginia Tech, and verify the forward-looking sonar system using a prototype sonar with nine beams. Post-processed simulations are performed while changing parameters in the sensitivity of the system to demonstrate the trade-off between the detection and false alarm rates.
- North America > United States > Virginia (0.25)
- North America > United States > New York (0.04)
- Europe > Spain (0.04)
- (2 more...)
Conceptual Cognitive Maps Formation with Neural Successor Networks and Word Embeddings
Stoewer, Paul, Schilling, Achim, Maier, Andreas, Krauss, Patrick
The human brain possesses the extraordinary capability to contextualize the information it receives from our environment. The entorhinal-hippocampal plays a critical role in this function, as it is deeply engaged in memory processing and constructing cognitive maps using place and grid cells. Comprehending and leveraging this ability could significantly augment the field of artificial intelligence. The multi-scale successor representation serves as a good model for the functionality of place and grid cells and has already shown promise in this role. Here, we introduce a model that employs successor representations and neural networks, along with word embedding vectors, to construct a cognitive map of three separate concepts. The network adeptly learns two different scaled maps and situates new information in proximity to related pre-existing representations. The dispersion of information across the cognitive map varies according to its scale - either being heavily concentrated, resulting in the formation of the three concepts, or spread evenly throughout the map. We suggest that our model could potentially improve current AI models by providing multi-modal context information to any input, based on a similarity metric for the input and pre-existing knowledge representations.
- Europe > Germany > Bavaria > Middle Franconia > Nuremberg (0.14)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
Insights into black box of artificial intelligence
At many banks, insurance companies and online retailers, self-learning computer algorithms are used to make decisions that have major consequences for customers. However, just how algorithms in artificial intelligence (AI) represent and process their input data internally is largely unknown. They have published their results in the journal Neural Networks. 'What we call artificial intelligence today is based on deep artificial neural networks that roughly mimic human brain functions,' explains Dr. Patrick Krauss from the Cognitive Computational Neuroscience Group at FAU. As is the case in children learning their native language without being aware of the rules of grammar, AI algorithms can learn to make the right choice by independently comparing a large amount of input data.
- Europe > France > Provence-Alpes-Côte d'Azur > Bouches-du-Rhône > Marseille (0.07)
- North America > Canada > Ontario > Toronto (0.05)
- Europe > Germany > Bavaria > Middle Franconia > Nuremberg (0.05)
- Banking & Finance (1.00)
- Health & Medicine > Therapeutic Area > Neurology (0.99)
The godless, leftist nature of artificial intelligence
The fast-moving field of artificial intelligence development is a lucrative one -- a head-spinning one -- an oft-surprising and exciting one. But peer past the frenzy of media headlines announcing the latest discoveries and newest breakthroughs and it's sad but true, the world of science, including technology, is a field dominated by godless leftists, too. Look at this headline, from a New Yorker piece in September of 2015: "All Scientists Should be Militant Atheists." That was by cosmologist Lawrence Krauss, who also served as the director of the Origins Project at Arizona State University, an endeavor aimed at exploring the universe, humanity and technology. Krauss saw his career suffer a bit of a hit in 2018, on the heels of accusations by women of sexual harassment -- accusations which he has strenuously denied, by the way.
- North America > United States > New York (0.26)
- North America > United States > Arizona (0.25)
- North America > United States > Texas (0.05)
- (2 more...)