Goto

Collaborating Authors

 Ju, Li


PARIC: Probabilistic Attention Regularization for Language Guided Image Classification from Pre-trained Vison Language Models

arXiv.org Artificial Intelligence

Developing robust image classification models that generalize effectively to unseen or out-of-distribution data remains a challenging problem in computer vision. This issue largely arises from biases and limited diversity in training datasets Torralba and Efros [2011]. Standard models trained on such data often prioritize irrelevant background or contextual cues over the discriminative visual features that define each class Ribeiro et al. [2016]. Consequently, these models struggle to generalize to unfamiliar or atypical examples, undermining their reliability and practical utility in real-world applications. Learning robust joint representations for vision and language is an important challenge in modern deep learning research, where the goal is to construct a function f(V, L) that aligns visual data V and linguistic data L into a unified representation capturing shared semantics while preserving modality-specific details; mathematically, this can be expressed as f: V L Z, where Z denotes the joint latent space encoding these semantics, with the primary challenge being to construct f such that it is both expressive and generalizable across diverse input types.


GraphBridge: Towards Arbitrary Transfer Learning in GNNs

arXiv.org Artificial Intelligence

Graph neural networks (GNNs) are conventionally trained on a per-domain, per-task basis. It creates a significant barrier in transferring the acquired knowledge to different, heterogeneous data setups. This paper introduces GraphBridge, a novel framework to enable knowledge transfer across disparate tasks and domains in GNNs, circumventing the need for modifications to task configurations or graph structures. Specifically, GraphBridge allows for the augmentation of any pre-trained GNN with prediction heads and a bridging network that connects the input to the output layer. This architecture not only preserves the intrinsic knowledge of the original model but also supports outputs of arbitrary dimensions. To mitigate the negative transfer problem, GraphBridge merges the source model with a concurrently trained model, thereby reducing the source bias when applied to the target domain. Our method is thoroughly evaluated across diverse transfer learning scenarios, including Graph2Graph, Node2Node, Graph2Node, and graph2point-cloud. Empirical validation, conducted over 16 datasets representative of these scenarios, confirms the framework's capacity for task- and domain-agnostic transfer learning within graph-like data, marking a significant advancement in the field of GNNs. Code is available at https://github.com/jujulili888/GraphBridge.


$\texttt{InfoHier}$: Hierarchical Information Extraction via Encoding and Embedding

arXiv.org Artificial Intelligence

Analyzing large-scale datasets, especially involving complex and high-dimensional data like images, is particularly challenging. While self-supervised learning (SSL) has proven effective for learning representations from unlabeled data, it typically focuses on flat, non-hierarchical structures, missing the multi-level relationships present in many realworld datasets. Hierarchical clustering (HC) can uncover these relationships by organizing data into a tree-like structure, but it often relies on rigid similarity metrics that struggle to capture the complexity of diverse data types. To address these we envision InfoHier, a framework that combines SSL with HC to jointly learn robust latent representations and hierarchical structures. This approach leverages SSL to provide adaptive representations, enhancing HC's ability to capture complex patterns. Simultaneously, it integrates HC loss to refine SSL training, resulting in representations that are more attuned to the underlying information hierarchy. InfoHier has the potential to improve the expressiveness and performance of both clustering and representation learning, offering significant benefits for data analysis, management, and information retrieval.


Accelerating Fair Federated Learning: Adaptive Federated Adam

arXiv.org Artificial Intelligence

Federated learning is a distributed and privacy-preserving approach to train a statistical model collaboratively from decentralized data of different parties. However, when datasets of participants are not independent and identically distributed (non-IID), models trained by naive federated algorithms may be biased towards certain participants, and model performance across participants is non-uniform. This is known as the fairness problem in federated learning. In this paper, we formulate fairness-controlled federated learning as a dynamical multi-objective optimization problem to ensure fair performance across all participants. To solve the problem efficiently, we study the convergence and bias of Adam as the server optimizer in federated learning, and propose Adaptive Federated Adam (AdaFedAdam) to accelerate fair federated learning with alleviated bias. We validated the effectiveness, Pareto optimality and robustness of AdaFedAdam in numerical experiments and show that AdaFedAdam outperforms existing algorithms, providing better convergence and fairness properties of the federated scheme.