Goto

Collaborating Authors

Block Sparse Bayesian Learning: A Diversified Scheme

Neural Information Processing Systems

This paper introduces a novel prior called Diversified Block Sparse Prior to characterize the widespread block sparsity phenomenon in real-world data. By allowing diversification on intra-block variance and inter-block correlation matrices, we effectively address the sensitivity issue of existing block sparse learning methods to pre-defined block information, which enables adaptive block estimation while mitigating the risk of overfitting. Based on this, a diversified block sparse Bayesian learning method (DivSBL) is proposed, utilizing EM algorithm and dual ascent method for hyperparameter estimation. Moreover, we establish the global and local optimality theory of our model.


On Robustness of Kernel Clustering

Neural Information Processing Systems

Clustering is an important unsupervised learning problem in machine learning and statistics. Among many existing algorithms, kernel k-means has drawn much research attention due to its ability to find non-linear cluster boundaries and its inherent simplicity. There are two main approaches for kernel k-means: SVD of the kernel matrix and convex relaxations. Despite the attention kernel clustering has received both from theoretical and applied quarters, not much is known about robustness of the methods. In this paper we first introduce a semidefinite programming relaxation for the kernel clustering problem, then prove that under a suitable model specification, both K-SVD and SDP approaches are consistent in the limit, albeit SDP is strongly consistent, i.e. achieves exact recovery, whereas K-SVD is weakly consistent, i.e. the fraction of misclassified nodes vanish. Also the error bounds suggest that SDP is more resilient towards outliers, which we also demonstrate with experiments.


On-Road Object Importance Estimation: A New Dataset and A Model with Multi-Fold Top-Down Guidance Zhixiong Nan 1, and Tao Xiang

Neural Information Processing Systems

This paper addresses the problem of on-road object importance estimation, which utilizes video sequences captured from the driver's perspective as the input. Although this problem is significant for safer and smarter driving systems, the exploration of this problem remains limited. On one hand, publicly-available large-scale datasets are scarce in the community. To address this dilemma, this paper contributes a new large-scale dataset named Traffic Object Importance (TOI). On the other hand, existing methods often only consider either bottom-up feature or single-fold guidance, leading to limitations in handling highly dynamic and diverse traffic scenarios.


ฮฒ-DPO: Direct Preference Optimization with Dynamic ฮฒ Junkang Wu1 Zhengyi Yang 1 Jiancan Wu1

Neural Information Processing Systems

Direct Preference Optimization (DPO) has emerged as a compelling approach for training Large Language Models (LLMs) to adhere to human preferences. However, the performance of DPO is sensitive to the fine-tuning of its trade-off parameter ฮฒ, as well as to the quality of the preference data. We analyze the impact of ฮฒ and data quality on DPO, uncovering that optimal ฮฒ values vary with the informativeness of pairwise data. Addressing the limitations of static ฮฒ values, we introduce a novel framework that dynamically calibrates ฮฒ at the batch level, informed by data quality considerations. Additionally, our method incorporates ฮฒ-guided data filtering to safeguard against the influence of outliers. Through empirical evaluation, we demonstrate that our dynamic ฮฒ adjustment technique significantly improves DPO's performance across a range of models and datasets, offering a more robust and adaptable training paradigm for aligning LLMs with human feedback.


A Functional Extension of Semi-Structured Networks

Neural Information Processing Systems

Semi-structured networks (SSNs) merge the structures familiar from additive models with deep neural networks, allowing the modeling of interpretable partial feature effects while capturing higher-order non-linearities at the same time. A significant challenge in this integration is maintaining the interpretability of the additive model component. Inspired by large-scale biomechanics datasets, this paper explores extending SSNs to functional data. Existing methods in functional data analysis are promising but often not expressive enough to account for all interactions and non-linearities and do not scale well to large datasets. Although the SSN approach presents a compelling potential solution, its adaptation to functional data remains complex. In this work, we propose a functional SSN method that retains the advantageous properties of classical functional regression approaches while also improving scalability. Our numerical experiments demonstrate that this approach accurately recovers underlying signals, enhances predictive performance, and performs favorably compared to competing methods.


Showing versus doing: Teaching by demonstration

Neural Information Processing Systems

People often learn from others' demonstrations, and inverse reinforcement learning (IRL) techniques have realized this capacity in machines. In contrast, teaching by demonstration has been less well studied computationally. Here, we develop a Bayesian model for teaching by demonstration. Stark differences arise when demonstrators are intentionally teaching (i.e.


Regularized Q-Learning

Neural Information Processing Systems

Q-learning is widely used algorithm in reinforcement learning (RL) community. Under the lookup table setting, its convergence is well established. However, its behavior is known to be unstable with the linear function approximation case. This paper develops a new Q-learning algorithm, called RegQ, that converges when linear function approximation is used. We prove that simply adding an appropriate regularization term ensures convergence of the algorithm. Its stability is established using a recent analysis tool based on switching system models. Moreover, we experimentally show that RegQ converges in environments where Q-learning with linear function approximation was known to diverge. An error bound on the solution where the algorithm converges is also given.


Fast yet Safe: Early-Exiting with Risk Control Alexander Timans 1, Tin Hadลพi Veljkoviฤ‡

Neural Information Processing Systems

Scaling machine learning models significantly improves their performance. However, such gains come at the cost of inference being slow and resource-intensive. Early-exit neural networks (EENNs) offer a promising solution: they accelerate inference by allowing intermediate layers to'exit' and produce a prediction early. Yet a fundamental issue with EENNs is how to determine when to exit without severely degrading performance. In other words, when is it'safe' for an EENN to go'fast'? To address this issue, we investigate how to adapt frameworks of risk control to EENNs. Risk control offers a distribution-free, post-hoc solution that tunes the EENN's exiting mechanism so that exits only occur when the output is of sufficient quality. We empirically validate our insights on a range of vision and language tasks, demonstrating that risk control can produce substantial computational savings, all the while preserving user-specified performance goals.


This AI image generator lets you create NSFW art, and it's only 30 for life

Mashable

TL;DR: Create anything, even NSFW art, with a lifetime subscription to Imagiyo for only 29.68. Digital creativity has never been more accessible, yet many of us remember the days when crafting a single image meant wrestling with layers and plugins for hours on end. Now there's a way to generate stunning visuals in seconds, simply by typing a description of what you have in mind. Imagiyo (on sale for life for 29.68) uses Stable Diffusion AI alongside FLUX AI to turn text prompts into high-quality images ready for commercial use, and there aren't many limits to what you can create. What do you want to make first?


Phased Exploration with Greedy Exploitation in Stochastic Combinatorial Partial Monitoring Games

Neural Information Processing Systems

Partial monitoring games are repeated games where the learner receives feedback that might be different from adversary's move or even the reward gained by the learner. Recently, a general model of combinatorial partial monitoring (CPM) games was proposed [1], where the learner's action space can be exponentially large and adversary samples its moves from a bounded, continuous space, according to a fixed distribution.