Goto

Collaborating Authors

 dmn


The immense interconnectivity of the brain: Best ideas of the century

New Scientist

You have probably heard the parable of the blind men and the elephant. One feels the trunk and says it's a snake, another feels a leg and claims it's a tree. It warns of how focusing on single parts can obscure the whole. Neuroscience made the same mistake for decades, viewing the brain as a collection of specialised regions, each working on a distinct function. Our understanding of what each region did often stemmed from incredible accidents, like the case of Phineas Gage, a 19th-century railway worker who survived having an iron rod blown through his brain.


Functional Brain Network Identification in Opioid Use Disorder Using Machine Learning Analysis of Resting-State fMRI BOLD Signals

Temtam, Ahmed, Witherow, Megan A., Ma, Liangsuo, Sadique, M. Shibly, Moeller, F. Gerard, Iftekharuddin, Khan M.

arXiv.org Artificial Intelligence

Understanding the neurobiology of opioid use disorder (OUD) using resting-state functional magnetic resonance imaging (rs-fMRI) may help inform treatment strategies to improve patient outcomes. Recent literature suggests temporal characteristics of rs-fMRI blood oxygenation level-dependent (BOLD) signals may offer complementary information to functional connectivity analysis. However, existing studies of OUD analyze BOLD signals using measures computed across all time points. This study, for the first time in the literature, employs data-driven machine learning (ML) modeling of rs-fMRI BOLD features representing multiple time points to identify region(s) of interest that differentiate OUD subjects from healthy controls (HC). Following the triple network model, we obtain rs-fMRI BOLD features from the default mode network (DMN), salience network (SN), and executive control network (ECN) for 31 OUD and 45 HC subjects. Then, we use the Boruta ML algorithm to identify statistically significant BOLD features that differentiate OUD from HC, identifying the DMN as the most salient functional network for OUD. Furthermore, we conduct brain activity mapping, showing heightened neural activity within the DMN for OUD. We perform 5-fold cross-validation classification (OUD vs. HC) experiments to study the discriminative power of functional network features with and without fusing demographic features. The DMN shows the most discriminative power, achieving mean AUC and F1 scores of 80.91% and 73.97%, respectively, when fusing BOLD and demographic features. Follow-up Boruta analysis using BOLD features extracted from the medial prefrontal cortex, posterior cingulate cortex, and left and right temporoparietal junctions reveals significant features for all four functional hubs within the DMN.


Diffusion-based Negative Sampling on Graphs for Link Prediction

Nguyen, Trung-Kien, Fang, Yuan

arXiv.org Artificial Intelligence

Link prediction is a fundamental task for graph analysis with important applications on the Web, such as social network analysis and recommendation systems, etc. Modern graph link prediction methods often employ a contrastive approach to learn robust node representations, where negative sampling is pivotal. Typical negative sampling methods aim to retrieve hard examples based on either predefined heuristics or automatic adversarial approaches, which might be inflexible or difficult to control. Furthermore, in the context of link prediction, most previous methods sample negative nodes from existing substructures of the graph, missing out on potentially more optimal samples in the latent space. To address these issues, we investigate a novel strategy of multi-level negative sampling that enables negative node generation with flexible and controllable ``hardness'' levels from the latent space. Our method, called Conditional Diffusion-based Multi-level Negative Sampling (DMNS), leverages the Markov chain property of diffusion models to generate negative nodes in multiple levels of variable hardness and reconcile them for effective graph link prediction. We further demonstrate that DMNS follows the sub-linear positivity principle for robust negative sampling. Extensive experiments on several benchmark datasets demonstrate the effectiveness of DMNS.


An epistemic logic for modeling decisions in the context of incomplete knowledge

Marković, Đorđe, Vandevelde, Simon, Vanbesien, Linde, Vennekens, Joost, Denecker, Marc

arXiv.org Artificial Intelligence

Substantial efforts have been made in developing various Decision Modeling formalisms, both from industry and academia. A challenging problem is that of expressing decision knowledge in the context of incomplete knowledge. In such contexts, decisions depend on what is known or not known. We argue that none of the existing formalisms for modeling decisions are capable of correctly capturing the epistemic nature of such decisions, inevitably causing issues in situations of uncertainty. This paper presents a new language for modeling decisions with incomplete knowledge. It combines three principles: stratification, autoepistemic logic, and definitions. A knowledge base in this language is a hierarchy of epistemic theories, where each component theory may epistemically reason on the knowledge in lower theories, and decisions are made using definitions with epistemic conditions.


Slow Momentum with Fast Reversion: A Trading Strategy Using Deep Learning and Changepoint Detection

Wood, Kieran, Roberts, Stephen, Zohren, Stefan

arXiv.org Machine Learning

Momentum strategies are an important part of alternative investments and are at the heart of commodity trading advisors (CTAs). These strategies have however been found to have difficulties adjusting to rapid changes in market conditions, such as during the 2020 market crash. In particular, immediately after momentum turning points, where a trend reverses from an uptrend (downtrend) to a downtrend (uptrend), time-series momentum (TSMOM) strategies are prone to making bad bets. To improve the response to regime change, we introduce a novel approach, where we insert an online change-point detection (CPD) module into a Deep Momentum Network (DMN) [1904.04912] pipeline, which uses an LSTM deep-learning architecture to simultaneously learn both trend estimation and position sizing. Furthermore, our model is able to optimise the way in which it balances 1) a slow momentum strategy which exploits persisting trends, but does not overreact to localised price moves, and 2) a fast mean-reversion strategy regime by quickly flipping its position, then swapping it back again to exploit localised price moves. Our CPD module outputs a changepoint location and severity score, allowing our model to learn to respond to varying degrees of disequilibrium, or smaller and more localised changepoints, in a data driven manner. Using a portfolio of 50, liquid, continuous futures contracts over the period 1990-2020, the addition of the CPD module leads to an improvement in Sharpe ratio of $33\%$. Even more notably, this module is especially beneficial in periods of significant nonstationarity, and in particular, over the most recent years tested (2015-2020) the performance boost is approximately $400\%$. This is especially interesting as traditional momentum strategies have been underperforming in this period.


Dreaming Is Like Taking LSD - Issue 95: Escape

Nautilus

Without a doubt, the biggest questions about dreaming are all variants on this question: Why do we dream? We began studying dreaming in the early 1990s and, between the two of us, have published over 200 scientific papers on sleep and dreams. Pulling together a variety of compelling neuroscientific ideas and state-of-the-art findings in the fields of sleep and dream research, we propose a new and innovative model of why we dream. We call this model NEXTUP. It proposes that our dreams allow us to explore the brain's neural network connections in order to understand possibilities.



Causality based Feature Fusion for Brain Neuro-Developmental Analysis

Kassani, Peyman Hosseinzadeh, Xiao, Li, Zhang, Gemeng, Stephen, Julia M., Wilson, Tony W., Calhoun, Vince D., Wang, Yu Ping

arXiv.org Artificial Intelligence

REPLACE THIS LINE WITH YOUR PAPER IDENTIFICATION NUMBER (DOUBLE - CLICK HERE TO EDIT) 1 Abstract -- Human brain development is a complex and dynamic process that is affected by several factors such as genetic s, sex hormones, and environmental changes . A number of recent studies on brain development have examined functional connectivity (FC) defi ned by the temporal correlation between time series of different brain regions. We propose to add the directional flow of information during brain maturation . To do so, w e extract effective connectivity (EC) through Granger causality (GC) for two different groups of subjects, i.e., children and young adults. The motivation is that the inclusion of causal interaction may further discriminate brain connections between two age groups and help to discover new conn ections between brain regions. The contributions of this study are three fold. First, t here has been a lack of attention to EC - based feature extraction in the context of brain development . T o this end, we propose a new kernel - based GC ( K GC) method to learn nonlinearity of complex brain network, where a reduced Sine hyperbolic polynomial ( RSP) neural network wa s used as our proposed learner . S econd, we use d causality values as the weight for the directional connectivity between brain regions . Our f indings indicate d that the strength of connections was significantly higher in young adult s relative to children. In addition, our new EC - based feature outperform ed FC - based analysis from Philadelphia neurocohort (PNC) study wi th better discrimination of the different age groups . Moreover, the fusion of these two sets of features (FC EC) improve d brain age prediction accuracy by more than 4 %, indicating that they should be used together for brain development stud ies . I NTRODUCTION uman brain development is a prolonged process that is initiated from the third gestational week (GW) to late adolescence, and presumably to the entire lifespan [ 1 ].


bpmNEXT 2018 Demonstrates Next Gen Processes

#artificialintelligence

There were some interesting and intense demos of how process would change over time. We all saw process linking with decision management, customer journeys, IoT, process mining, advanced analytics, AI, RPA, Robots, Blockchain, voice and image recognition. There were many dimensions of process evolution practically demonstrated in 30 minute segments. It was clear that process will be involved with significant innovation in the evolving digital world and that transformation is doable in increments. While most of the participants were vendors, there were notable visionary end users like Quicken Loans (the designers of customer journey called "Rocket Mortgage").


Separation of time scales and direct computation of weights in deep neural networks

Dehmamy, Nima, Rohani, Neda, Katsaggelos, Aggelos

arXiv.org Machine Learning

Artificial intelligence is revolutionizing our lives at an ever increasing pace. At the heart of this revolution is the recent advancements in deep neural networks (DNN), learning to perform sophisticated, high-level tasks. However, training DNNs requires massive amounts of data and is very computationally intensive. Gaining analytical understanding of the solutions found by DNNs can help us devise more efficient training algorithms, replacing the commonly used mthod of stochastic gradient descent (SGD). We analyze the dynamics of SGD and show that, indeed, direct computation of the solutions is possible in many cases. We show that a high performing setup used in DNNs introduces a separation of time-scales in the training dynamics, allowing SGD to train layers from the lowest (closest to input) to the highest. We then show that for each layer, the distribution of solutions found by SGD can be estimated using a class-based principal component analysis (PCA) of the layer's input. This finding allows us to forgo SGD entirely and directly derive the DNN parameters using this class-based PCA, which can be well estimated using significantly less data than SGD. We implement these results on image datasets MNIST, CIFAR10 and CIFAR100 and find that, in fact, layers derived using our class-based PCA perform comparable or superior to neural networks of the same size and architecture trained using SGD. We also confirm that the class-based PCA often converges using a fraction of the data required for SGD. Thus, using our method training time can be reduced both by requiring less training data than SGD, and by eliminating layers in the costly backpropagation step of the training.