Goto

Collaborating Authors

 Li, Shihao


PANDORA: Deep graph learning based COVID-19 infection risk level forecasting

arXiv.org Artificial Intelligence

COVID-19 as a global pandemic causes a massive disruption to social stability that threatens human life and the economy. Policymakers and all elements of society must deliver measurable actions based on the pandemic's severity to minimize the detrimental impact of COVID-19. A proper forecasting system is arguably important to provide an early signal of the risk of COVID-19 infection so that the authorities are ready to protect the people from the worst. However, making a good forecasting model for infection risks in different cities or regions is not an easy task, because it has a lot of influential factors that are difficult to be identified manually. To address the current limitations, we propose a deep graph learning model, called PANDORA, to predict the infection risks of COVID-19, by considering all essential factors and integrating them into a geographical network. The framework uses geographical position relations and transportation frequency as higher-order structural properties formulated by higher-order network structures (i.e., network motifs). Moreover, four significant node attributes (i.e., multiple features of a particular area, including climate, medical condition, economy, and human mobility) are also considered. We propose three different aggregators to better aggregate node attributes and structural features, namely, Hadamard, Summation, and Connection. Experimental results over real data show that PANDORA outperforms the baseline method with higher accuracy and faster convergence speed, no matter which aggregator is chosen. We believe that PANDORA using deep graph learning provides a promising approach to get superior performance in infection risk level forecasting and help humans battle the COVID-19 crisis.


State Space Model for New-Generation Network Alternative to Transformers: A Survey

arXiv.org Artificial Intelligence

In the post-deep learning era, the Transformer architecture has demonstrated its powerful performance across pre-trained big models and various downstream tasks. However, the enormous computational demands of this architecture have deterred many researchers. To further reduce the complexity of attention models, numerous efforts have been made to design more efficient methods. Among them, the State Space Model (SSM), as a possible replacement for the self-attention based Transformer model, has drawn more and more attention in recent years. In this paper, we give the first comprehensive review of these works and also provide experimental comparisons and analysis to better demonstrate the features and advantages of SSM. Specifically, we first give a detailed description of principles to help the readers quickly capture the key ideas of SSM. After that, we dive into the reviews of existing SSMs and their various applications, including natural language processing, computer vision, graph, multi-modal and multi-media, point cloud/event stream, time series data, and other domains. In addition, we give statistical comparisons and analysis of these models and hope it helps the readers to understand the effectiveness of different structures on various tasks. Then, we propose possible research points in this direction to better promote the development of the theoretical model and application of SSM. More related works will be continuously updated on the following GitHub: https://github.com/Event-AHU/Mamba_State_Space_Model_Paper_List.


MRIF: Multi-resolution Interest Fusion for Recommendation

arXiv.org Machine Learning

In this paper, we introduce multi-resolution interest interests based on their historical behaviors. Most of recent advances fusion model (MRIF) composed of interest extraction layer, interest in recommender systems mainly focus on modeling users' aggregation layer, and attentional fusion structure, which addresses preferences accurately using deep learning based approaches. There the problem of extracting users' preferences at different temporalranges are two important properties of users' interests, one is that users' and combining multi-resolution interests effectively. The interests are dynamic and evolve over time, the other is that users' main contributions are: interests have different resolutions, or temporal-ranges to be precise, such as long-term and short-term preferences. Existing approaches - We design a new network structure to model the dynamic either use Recurrent Neural Networks (RNNs) to address changes and different temporal-ranges of users' interests, the drifts in users' interests without considering different temporalranges, which yields more accurate prediction results than extracting or design two different networks to model long-term and main interest directly from interaction sequences.