Tong, Hanghang
Logic Query of Thoughts: Guiding Large Language Models to Answer Complex Logic Queries with Knowledge Graphs
Liu, Lihui, Wang, Zihao, Qiu, Ruizhong, Ban, Yikun, Chan, Eunice, Song, Yangqiu, He, Jingrui, Tong, Hanghang
Despite the superb performance in many tasks, large language models (LLMs) bear the risk of generating hallucination or even wrong answers when confronted with tasks that demand the accuracy of knowledge. The issue becomes even more noticeable when addressing logic queries that require multiple logic reasoning steps. On the other hand, knowledge graph (KG) based question answering methods are capable of accurately identifying the correct answers with the help of knowledge graph, yet its accuracy could quickly deteriorate when the knowledge graph itself is sparse and incomplete. It remains a critical challenge on how to integrate knowledge graph reasoning with LLMs in a mutually beneficial way so as to mitigate both the hallucination problem of LLMs as well as the incompleteness issue of knowledge graphs. In this paper, we propose 'Logic-Query-of-Thoughts' (LGOT) which is the first of its kind to combine LLMs with knowledge graph based logic query reasoning. LGOT seamlessly combines knowledge graph reasoning and LLMs, effectively breaking down complex logic queries into easy to answer subquestions. Through the utilization of both knowledge graph reasoning and LLMs, it successfully derives answers for each subquestion. By aggregating these results and selecting the highest quality candidate answers for each step, LGOT achieves accurate results to complex questions. Our experimental findings demonstrate substantial performance enhancements, with up to 20% improvement over ChatGPT.
Heterogeneous Contrastive Learning for Foundation Models and Beyond
Zheng, Lecheng, Jing, Baoyu, Li, Zihao, Tong, Hanghang, He, Jingrui
In the era of big data and Artificial Intelligence, an emerging paradigm is to utilize contrastive self-supervised learning to model large-scale heterogeneous data. Many existing foundation models benefit from the generalization capability of contrastive self-supervised learning by learning compact and high-quality representations without relying on any label information. Amidst the explosive advancements in foundation models across multiple domains, including natural language processing and computer vision, a thorough survey on heterogeneous contrastive learning for the foundation model is urgently needed. In response, this survey critically evaluates the current landscape of heterogeneous contrastive learning for foundation models, highlighting the open challenges and future trends of contrastive learning. In particular, we first present how the recent advanced contrastive learning-based methods deal with view heterogeneity and how contrastive learning is applied to train and fine-tune the multi-view foundation models. Then, we move to contrastive learning methods for task heterogeneity, including pretraining tasks and downstream tasks, and show how different tasks are combined with contrastive learning loss for different purposes. Finally, we conclude this survey by discussing the open challenges and shedding light on the future directions of contrastive learning.
Soft Reasoning on Uncertain Knowledge Graphs
Fei, Weizhi, Wang, Zihao, Yin, Hang, Duan, Yang, Tong, Hanghang, Song, Yangqiu
The further possibilities in data management (Wang et al., 2022; uncertain nature of knowledge is widely observed Ren et al., 2023). in the real world, but does not align seamlessly with the first-order logic underpinning existing Uncertain knowledge is widely observed from the daily studies. To bridge this gap, we study the setting events (Zhang et al., 2020) to the interaction of biological of soft queries on uncertain knowledge, which systems (Szklarczyk et al., 2023). Besides, uncertainty is is motivated by the establishment of soft constraint also particularly pervasive in KGs because KGs are constructed programming. We further propose an MLbased by information extraction models that could introduce approach with both forward inference and errors (Angeli et al., 2015; Ponte & Croft, 2017) backward calibration to answer soft queries on and from large corpses that could be noisy (Carlson et al., large-scale, incomplete, and uncertain knowledge 2010). To represent the uncertain knowledge, confidence graphs. Theoretical discussions present that our values p are associated with triples in many well-established methods share the same complexity as state-ofthe-art KGs (Carlson et al., 2010; Speer et al., 2017; Szklarczyk inference algorithms for first-order queries.
Rethinking the Bounds of LLM Reasoning: Are Multi-Agent Discussions the Key?
Wang, Qineng, Wang, Zihao, Su, Ying, Tong, Hanghang, Song, Yangqiu
Recent progress in LLMs discussion suggests that multi-agent discussion improves the reasoning abilities of LLMs. In this work, we reevaluate this claim through systematic experiments, where we propose a novel group discussion framework to enrich the set of discussion mechanisms. Interestingly, our results show that a single-agent LLM with strong prompts can achieve almost the same performance as the best existing discussion approach on a wide range of reasoning tasks and backbone LLMs. We observe that the multi-agent discussion performs better than a single agent only when there is no demonstration in the prompt. Further study reveals the common interaction mechanisms of LLMs during the discussion.
On the Generalization Capability of Temporal Graph Learning Algorithms: Theoretical Insights and a Simpler Method
Cong, Weilin, Kang, Jian, Tong, Hanghang, Mahdavi, Mehrdad
Temporal graph learning (TGL) has emerged as an important machine learning problem and is widely used in a number of real-world applications, such as traffic prediction [Yuan and Li, 2021, Zhang et al., 2021], knowledge graphs [Cai et al., 2022, Leblay and Chekol, 2018], and recommender systems [Kumar et al., 2019, Rossi et al., 2020, Xu et al., 2020a]. A typical downstream task of temporal graph learning is link prediction, which focuses on predicting future interactions among nodes. For example in an online video recommender system, the user-video clicks can be modeled as a temporal graph whose nodes represent users and videos, and links are associated with timestamps indicating when users click videos. Link prediction between nodes can be used to predict if and when a user is interested in a video. Therefore, designing graph learning models that can capture node evolutionary patterns and accurately predict future links is important. TGL is generally more challenging than static graph learning, thereby requiring more sophisticated algorithms to model the temporal evolutionary patterns [Huang et al., 2023]. In recent years, many TGL algorithms [Kumar et al., 2019, Xu et al., 2020a, Rossi et al., 2020, Sankar et al., 2020, Wang et al., 2021e] have been proposed that leverage memory blocks, self-attention, time-encoding function, recurrent neural networks (RNNs), temporal walks, and message passing to better capture the meaningful structural or temporal patterns. For instance, JODIE [Kumar et al., 2019] maintains a memory block for each node and utilizes an RNN to update the memory blocks upon the occurance of each interaction; TGAT [Xu et al., 2020a] utilizes self-attention message passing to aggregate neighbor information on the temporal graph; TGN [Rossi et al., 2020] combines memory blocks with message passing to allow each node in the temporal graph to have a receptive field that is not limited by the number of message-passing layers; DySAT [Sankar et al., 2020] uses self-attention to capture structural information and uses RNN to capture temporal dependencies; CAW [Wang et al., 2021e] captures temporal dependencies between nodes by performing multiple temporal walks from the root
Trustworthy Graph Neural Networks: Aspects, Methods and Trends
Zhang, He, Wu, Bang, Yuan, Xingliang, Pan, Shirui, Tong, Hanghang, Pei, Jian
Graph neural networks (GNNs) have emerged as a series of competent graph learning methods for diverse real-world scenarios, ranging from daily applications like recommendation systems and question answering to cutting-edge technologies such as drug discovery in life sciences and n-body simulation in astrophysics. However, task performance is not the only requirement for GNNs. Performance-oriented GNNs have exhibited potential adverse effects like vulnerability to adversarial attacks, unexplainable discrimination against disadvantaged groups, or excessive resource consumption in edge computing environments. To avoid these unintentional harms, it is necessary to build competent GNNs characterised by trustworthiness. To this end, we propose a comprehensive roadmap to build trustworthy GNNs from the view of the various computing technologies involved. In this survey, we introduce basic concepts and comprehensively summarise existing efforts for trustworthy GNNs from six aspects, including robustness, explainability, privacy, fairness, accountability, and environmental well-being. Additionally, we highlight the intricate cross-aspect relations between the above six aspects of trustworthy GNNs. Finally, we present a thorough overview of trending directions for facilitating the research and industrialisation of trustworthy GNNs.
ARIEL: Adversarial Graph Contrastive Learning
Feng, Shengyu, Jing, Baoyu, Zhu, Yada, Tong, Hanghang
Contrastive learning is an effective unsupervised method in graph representation learning, and the key component of contrastive learning lies in the construction of positive and negative samples. Previous methods usually utilize the proximity of nodes in the graph as the principle. Recently, the data-augmentation-based contrastive learning method has advanced to show great power in the visual domain, and some works extended this method from images to graphs. However, unlike the data augmentation on images, the data augmentation on graphs is far less intuitive and much harder to provide high-quality contrastive samples, which leaves much space for improvement. In this work, by introducing an adversarial graph view for data augmentation, we propose a simple but effective method, Adversarial Graph Contrastive Learning (ARIEL), to extract informative contrastive samples within reasonable constraints. We develop a new technique called information regularization for stable training and use subgraph sampling for scalability. We generalize our method from node-level contrastive learning to the graph level by treating each graph instance as a super-node. ARIEL consistently outperforms the current graph contrastive learning methods for both node-level and graph-level classification tasks on real-world datasets. We further demonstrate that ARIEL is more robust in the face of adversarial attacks.
Conversational Question Answering with Reformulations over Knowledge Graph
Liu, Lihui, Hill, Blaine, Du, Boxin, Wang, Fei, Tong, Hanghang
conversational question answering (convQA) over knowledge graphs (KGs) involves answering multi-turn natural language questions about information contained in a KG. State-of-the-art methods of ConvQA often struggle with inexplicit question-answer pairs. These inputs are easy for human beings to understand given a conversation history, but hard for a machine to interpret, which can degrade ConvQA performance. To address this problem, we propose a reinforcement learning (RL) based model, CornNet, which utilizes question reformulations generated by large language models (LLMs) to improve ConvQA performance. CornNet adopts a teacher-student architecture where a teacher model learns question representations using human writing reformulations, and a student model to mimic the teacher model's output via reformulations generated by LLMs. The learned question representation is then used by an RL model to locate the correct answer in a KG. Extensive experimental results show that CornNet outperforms state-of-the-art convQA models.
STERLING: Synergistic Representation Learning on Bipartite Graphs
Jing, Baoyu, Yan, Yuchen, Ding, Kaize, Park, Chanyoung, Zhu, Yada, Liu, Huan, Tong, Hanghang
A fundamental challenge of bipartite graph representation learning is how to extract informative node embeddings. Self-Supervised Learning (SSL) is a promising paradigm to address this challenge. Most recent bipartite graph SSL methods are based on contrastive learning which learns embeddings by discriminating positive and negative node pairs. Contrastive learning usually requires a large number of negative node pairs, which could lead to computational burden and semantic errors. In this paper, we introduce a novel synergistic representation learning model (STERLING) to learn node embeddings without negative node pairs. STERLING preserves the unique local and global synergies in bipartite graphs. The local synergies are captured by maximizing the similarity of the inter-type and intra-type positive node pairs, and the global synergies are captured by maximizing the mutual information of co-clusters. Theoretical analysis demonstrates that STERLING could improve the connectivity between different node types in the embedding space. Extensive empirical evaluation on various benchmark datasets and tasks demonstrates the effectiveness of STERLING for extracting node embeddings.
FairGen: Towards Fair Graph Generation
Zheng, Lecheng, Zhou, Dawei, Tong, Hanghang, Xu, Jiejun, Zhu, Yada, He, Jingrui
There have been tremendous efforts over the past decades dedicated to the generation of realistic graphs in a variety of domains, ranging from social networks to computer networks, from gene regulatory networks to online transaction networks. Despite the remarkable success, the vast majority of these works are unsupervised in nature and are typically trained to minimize the expected graph reconstruction loss, which would result in the representation disparity issue in the generated graphs, i.e., the protected groups (often minorities) contribute less to the objective and thus suffer from systematically higher errors. In this paper, we aim to tailor graph generation to downstream mining tasks by leveraging label information and user-preferred parity constraints. In particular, we start from the investigation of representation disparity in the context of graph generative models. To mitigate the disparity, we propose a fairness-aware graph generative model named FairGen. Our model jointly trains a label-informed graph generation module and a fair representation learning module by progressively learning the behaviors of the protected and unprotected groups, from the `easy' concepts to the `hard' ones. In addition, we propose a generic context sampling strategy for graph generative models, which is proven to be capable of fairly capturing the contextual information of each group with a high probability. Experimental results on seven real-world data sets, including web-based graphs, demonstrate that FairGen (1) obtains performance on par with state-of-the-art graph generative models across nine network properties, (2) mitigates the representation disparity issues in the generated graphs, and (3) substantially boosts the model performance by up to 17% in downstream tasks via data augmentation.