Goto

Collaborating Authors

 performance comparison






Trustworthy Machine Learning under Distribution Shifts

Huang, Zhuo

arXiv.org Machine Learning

Machine Learning (ML) has been a foundational topic in artificial intelligence (AI), providing both theoretical groundwork and practical tools for its exciting advancements. From ResNet for visual recognition to Transformer for vision-language alignment, the AI models have achieved superior capability to humans. Furthermore, the scaling law has enabled AI to initially develop general intelligence, as demonstrated by Large Language Models (LLMs). To this stage, AI has had an enormous influence on society and yet still keeps shaping the future for humanity. However, distribution shift remains a persistent ``Achilles' heel'', fundamentally limiting the reliability and general usefulness of ML systems. Moreover, generalization under distribution shift would also cause trust issues for AIs. Motivated by these challenges, my research focuses on \textit{Trustworthy Machine Learning under Distribution Shifts}, with the goal of expanding AI's robustness, versatility, as well as its responsibility and reliability. We carefully study the three common distribution shifts into: (1) Perturbation Shift, (2) Domain Shift, and (3) Modality Shift. For all scenarios, we also rigorously investigate trustworthiness via three aspects: (1) Robustness, (2) Explainability, and (3) Adaptability. Based on these dimensions, we propose effective solutions and fundamental insights, meanwhile aiming to enhance the critical ML problems, such as efficiency, adaptability, and safety.


Appendix for Regularized Softmax Deep Multi-Agent Q-Learning Ling Pan

Neural Information Processing Systems

A darker color represents a larger value. Work done while at University of Oxford. B.2 Algorithm and More Details for Computing the Approximate Softmax Operator The full algorithm for computing the approximate softmax operator is in Algorithm 1.Algorithm 1 Approximate softmax operator E.1 Experimental Setup T asks. SC2.4.6.2.69232, and performance is not always comparable across versions. All experiments are run on P100 GPU.


Reviewer # 1

Neural Information Processing Systems

We thank the reviewers for their positive feedbacks and valuable suggestions. We address their comments below. In short, "pruning-at-initialization" methods have the advantage of less overhead at training time. In contrast, our approach can be adopted even when the model is trained without any consideration of pruning. However, we think providing a self-sufficient proof in the current form is not a bad idea, either.


Supplementary Material Density-driven Regularization for Out-of-distribution Detection A.1 Proof of lemma 1

Neural Information Processing Systems

If Eq.(3) holds, then null Lemma 2. If Eq.(3) holds, then Υ For independent and identically distributed (i.i.d.) random vector Let g ([x, y,z ]) = y/x z. Proposition 1. subtracting a fixed constant from the classification logits leads to the same consistency OOD datasets to verify the effectiveness of the proposed two regularization terms. The result is the average value across all OOD datasets.ablation Fig.4 shows the distribution of log-likelihood values


KAN-AFT: An Interpretable Nonlinear Survival Model Integrating Kolmogorov-Arnold Networks with Accelerated Failure Time Analysis

Jose, Mebin, Francis, Jisha, Kattumannil, Sudheesh Kumar

arXiv.org Machine Learning

Survival analysis relies fundamentally on the semi-parametric Cox Proportional Hazards (CoxPH) model and the parametric Accelerated Failure Time (AFT) model. CoxPH assumes constant hazard ratios, often failing to capture real-world dynamics, while traditional AFT models are limited by rigid distributional assumptions. Although deep learning models like DeepAFT address these constraints by improving predictive accuracy and handling censoring, they inherit the significant challenge of black-box interpretability. The recent introduction of CoxKAN demonstrated the successful integration of Kolmogorov-Arnold Networks (KANs), a novel architecture that yields highly accurate and interpretable symbolic representations, within the CoxPH framework. Motivated by the interpretability gains of CoxKAN, we introduce KAN-AFT (Kolmogorov Arnold Network-based AFT), the first framework to apply KANs to the AFT model. Our primary contributions include: (i) a principled AFT-KAN formulation, (ii) robust optimization strategies for right-censored observations (e.g., Buckley-James and IPCW), and (iii) an interpretability pipeline that converts the learned spline functions into closed-form symbolic equations for survival time. Empirical results on multiple datasets confirm that KAN-AFT achieves performance comparable to or better than DeepAFT, while uniquely providing transparent, symbolic models of the survival process.


How Should We Evaluate Data Deletion in Graph-Based ANN Indexes?

Yamashita, Tomohiro, Amagata, Daichi, Matsui, Yusuke

arXiv.org Artificial Intelligence

Approximate Nearest Neighbor Search (ANNS) has recently gained significant attention due to its many applications, such as Retrieval-Augmented Generation. Such applications require ANNS algorithms that support dynamic data, so the ANNS problem on dynamic data has attracted considerable interest. However, a comprehensive evaluation methodology for data deletion in ANNS has yet to be established. This study proposes an experimental framework and comprehensive evaluation metrics to assess the efficiency of data deletion for ANNS indexes under practical use cases. Specifically, we categorize data deletion methods in graph-based ANNS into three approaches and formalize them mathematically. The performance is assessed in terms of accuracy, query speed, and other relevant metrics. Finally, we apply the proposed evaluation framework to Hierarchical Navigable Small World, one of the state-of-the-art ANNS methods, to analyze the effects of data deletion, and propose Deletion Control, a method which dynamically selects the appropriate deletion method under a required search accuracy.