Goto

Collaborating Authors

 gru layer


DynBERG: Dynamic BERT-based Graph neural network for financial fraud detection

Kulkarni, Omkar, Chandra, Rohitash

arXiv.org Artificial Intelligence

Financial fraud detection is critical for maintaining the integrity of financial systems, particularly in decentralised environments such as cryptocurrency networks. Although Graph Convolutional Networks (GCNs) are widely used for financial fraud detection, graph Transformer models such as Graph-BERT are gaining prominence due to their Transformer-based architecture, which mitigates issues such as over-smoothing. Graph-BERT is designed for static graphs and primarily evaluated on citation networks with undirected edges. However, financial transaction networks are inherently dynamic, with evolving structures and directed edges representing the flow of money. To address these challenges, we introduce DynBERG, a novel architecture that integrates Graph-BERT with a Gated Recurrent Unit (GRU) layer to capture temporal evolution over multiple time steps. Additionally, we modify the underlying algorithm to support directed edges, making DynBERG well-suited for dynamic financial transaction analysis. We evaluate our model on the Elliptic dataset, which includes Bitcoin transactions, including all transactions during a major cryptocurrency market event, the Dark Market Shutdown. By assessing DynBERG's resilience before and after this event, we analyse its ability to adapt to significant market shifts that impact transaction behaviours. Our model is benchmarked against state-of-the-art dynamic graph classification approaches, such as EvolveGCN and GCN, demonstrating superior performance, outperforming EvolveGCN before the market shutdown and surpassing GCN after the event. Additionally, an ablation study highlights the critical role of incorporating a time-series deep learning component, showcasing the effectiveness of GRU in modelling the temporal dynamics of financial transactions.


Intelligent DoS and DDoS Detection: A Hybrid GRU-NTM Approach to Network Security

Panggabean, Caroline, Venkatachalam, Chandrasekar, Shah, Priyanka, John, Sincy, P, Renuka Devi, Venkatachalam, Shanmugavalli

arXiv.org Artificial Intelligence

Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any cur rent or future media. Caroline Panggabean Departement of CSE (AI) JAIN (Deemed - to - be University) Bangalore, Karnataka carolinepgabean@gmail.com Sincy John Departement of CSE (AIM) JAIN (Deemed - to - be University) Bangalore, Karnataka sincyjohn@jainuniversity.ac.in Chandrasekar Venkatachalam Departement of CSE (AI) JAIN (Deemed - to - be University) Bangalore, Karnataka chandrasekar.v@jainuniversity.ac.in Renuka Devi P Departement of CSE (AIML) JAIN (Deemed - to - be University) Bangalore, Karnataka renukadevi.p@jainuniversity.ac.in Priyanka Shah Departement of CSE (AI) JAIN (Deemed - to - be University) Bangalore, Karnataka priyankashah8324@gmail.com Shanmugavalli Venkatachalam Department of CSE KSR College of Engineering Namakkal, Tamil N adu drvshanmugavalli@gmail.com Abstract -- Detecting Denial of Service (DoS) and Distributed Denial of Service (DDoS) attacks remains a critical challenge in cybersecurity. This research introduces a hybrid deep learning model combining Gated Recurrent Units (GRUs) and a Neural Turing Machine (NTM) for enhanced intrusion detection. Trained on UNSW - NB15 and BoT - IoT datasets, the model employs GRU layers for sequential data processing and an NTM for long - term pattern recognition.


Leveraging Recurrent Neural Networks for Predicting Motor Movements from Primate Motor Cortex Neural Recordings

Wang, Yuanxi, Wang, Zuowen, Liu, Shih-Chii

arXiv.org Artificial Intelligence

This paper presents an efficient deep learning solution for decoding motor movements from neural recordings in non-human primates. An Autoencoder Gated Recurrent Unit (AEGRU) model was adopted as the model architecture for this task. The autoencoder is only used during the training stage to achieve better generalization. Together with the preprocessing techniques, our model achieved 0.71 $R^2$ score, surpassing the baseline models in Neurobench and is ranked first for $R^2$ in the IEEE BioCAS 2024 Grand Challenge on Neural Decoding. Model pruning is also applied leading to a reduction of 41.4% of the multiply-accumulate (MAC) operations with little change in the $R^2$ score compared to the unpruned model.


Personalized Speech Enhancement Without a Separate Speaker Embedding Model

Pärnamaa, Tanel, Saabas, Ando

arXiv.org Artificial Intelligence

Personalized speech enhancement (PSE) models can improve the audio quality of teleconferencing systems by adapting to the characteristics of a speaker's voice. However, most existing methods require a separate speaker embedding model to extract a vector representation of the speaker from enrollment audio, which adds complexity to the training and deployment process. We propose to use the internal representation of the PSE model itself as the speaker embedding, thereby avoiding the need for a separate model. We show that our approach performs equally well or better than the standard method of using a pre-trained speaker embedding model on noise suppression and echo cancellation tasks. Moreover, our approach surpasses the ICASSP 2023 Deep Noise Suppression Challenge winner by 0.15 in Mean Opinion Score.


Classifying Objects in 3D Point Clouds Using Recurrent Neural Network: A GRU LSTM Hybrid Approach

Mousa, Ramin, Khezli, Mitra, Azadi, Mohamadreza, Nikoofard, Vahid, Hesaraki, Saba

arXiv.org Artificial Intelligence

Accurate classification of objects in 3D point clouds is a significant problem in several applications, such as autonomous navigation and augmented/virtual reality scenarios, which has become a research hot spot. In this paper, we presented a deep learning strategy for 3D object classification in augmented reality. The proposed approach is a combination of the GRU and LSTM. LSTM networks learn longer dependencies well, but due to the number of gates, it takes longer to train; on the other hand, GRU networks have a weaker performance than LSTM, but their training speed is much higher than GRU, which is The speed is due to its fewer gates. The proposed approach used the combination of speed and accuracy of these two networks. The proposed approach achieved an accuracy of 0.99 in the 4,499,0641 points dataset, which includes eight classes (unlabeled, man-made terrain, natural terrain, high vegetation, low vegetation, buildings, hardscape, scanning artifacts, cars). Meanwhile, the traditional machine learning approaches could achieve a maximum accuracy of 0.9489 in the best case. Keywords: Point Cloud Classification, Virtual Reality, Hybrid Model, GRULSTM, GRU, LSTM


Recurrent Neural Networks (RNN) with Keras

#artificialintelligence

Recurrent neural networks (RNN) are a class of neural networks that is powerful for modeling sequence data such as time series or natural language. Schematically, a RNN layer uses a for loop to iterate over the timesteps of a sequence, while maintaining an internal state that encodes information about the timesteps it has seen so far. Ease of use: the built-in keras.layers.RNN, keras.layers.LSTM, keras.layers.GRU layers enable you to quickly build recurrent models without having to make difficult configuration choices. Ease of customization: You can also define your own RNN cell layer (the inner part of the for loop) with custom behavior, and use it with the generic keras.layers.RNN layer (the for loop itself). This allows you to quickly prototype different research ideas in a flexible way with minimal code.


Improving EEG based Continuous Speech Recognition

Krishna, Gautam, Tran, Co, Carnahan, Mason, Han, Yan, Tewfik, Ahmed H

arXiv.org Machine Learning

Improving EEG based Continuous Speech Recognition Gautam Krishna Brain Machine Interface Lab The University of T exas at Austin Austin, Texas Co Tran Brain Machine Interface Lab The University of T exas at Austin Austin, Texas Mason Carnahan Brain Machine Interface Lab The University of T exas at Austin Austin, Texas Y an Han Brain Machine Interface Lab The University of T exas at Austin Austin, Texas Ahmed H Tewfik Brain Machine Interface Lab The University of T exas at Austin Austin, Texas Abstract --In this paper we introduce various techniques to improve the performance of electroencephalography (EEG) features based continuous speech recognition (CSR) systems. A connectionist temporal classification (CTC) based automatic speech recognition (ASR) system was implemented for performing recognition. We introduce techniques to initialize the weights of the recurrent layers in the encoder of the CTC model with more meaningful weights rather than with random weights and we make use of an external language model to improve the beam search during decoding time. We finally study the problem of predicting articulatory features from EEG features in this paper . ASR systems forms front end or back end in many state of the art voice assistant systems like Bixby, Alexa,Siri,Cortana etc.


Falls Prediction in eldery people using Gated Recurrent Units

Radzio, Marcin, Wielgosz, Maciej, Mertik, Matej

arXiv.org Machine Learning

Falls prevention, especially in older people, becomes an increasingly important topic in the times of aging societies. In this work, we present Gated Recurrent Unit-based neural networks models designed for predicting falls (syncope). The cardiovascular systems signals used in the study come from Gravitational Physiology, Aging and Medicine Research Unit, Institute of Physiology, Medical University of Graz. We used two of the collected signals, heart rate, and mean blood pressure. By using bidirectional GRU model, it was possible to predict the syncope occurrence approximately ten minutes before the manual marker.


Amobee at SemEval-2018 Task 1: GRU Neural Network with a CNN Attention Mechanism for Sentiment Classification

Rozental, Alon, Fleischer, Daniel

arXiv.org Machine Learning

This paper describes the participation of Amobee in the shared sentiment analysis task at SemEval 2018. We participated in all the English sub-tasks and the Spanish valence tasks. Our system consists of three parts: training task-specific word embeddings, training a model consisting of gated-recurrent-units (GRU) with a convolution neural network (CNN) attention mechanism and training stacking-based ensembles for each of the sub-tasks. Our algorithm reached 3rd and 1st places in the valence ordinal classification sub-tasks in English and Spanish, respectively.