Well File:

OpenAssistant Conversations - Democratizing Large Language Model Alignment

Neural Information Processing Systems

Aligning large language models (LLMs) with human preferences has proven to drastically improve usability and has driven rapid adoption as demonstrated by ChatGPT. Alignment techniques such as supervised fine-tuning (SFT) and reinforcement learning from human feedback (RLHF) greatly reduce the required skill and domain knowledge to effectively harness the capabilities of LLMs, increasing their accessibility and utility across various domains. However, state-of-the-art alignment techniques like RLHF rely on high-quality human feedback data, which is expensive to create and often remains proprietary. In an effort to democratize research on large-scale alignment, we release OpenAssistant Conversations, a human-generated, human-annotated assistant-style conversation corpus consisting of 161,443 messages in 35 different languages, annotated with 461,292 quality ratings, resulting in over 10,000 complete and fully annotated conversation trees. The corpus is a product of a worldwide crowd-sourcing effort involving over 13,500 volunteers. Models trained on OpenAssistant Conversations show consistent improvements on standard benchmarks over respective base models.


MAC Advice for Facility Location Mechanism Design

Neural Information Processing Systems

Algorithms with predictions are gaining traction across various domains, as a way to surpass traditional worst-case bounds through (machine-learned) advice. We study the canonical problem of k-facility location mechanism design, where the n agents are strategic and might misreport their locations. We receive a prediction for each agent's location, and these predictions are crucially allowed to be only "mostly" and "approximately" correct (M


A Proofs

Neural Information Processing Systems

We will prove it by contradiction. To prove Lemma 2 we will use the following lemma. This is a special case of the simulation lemma (Kearns and Singh, 2002). We will prove it by contradiction. There is a sizeable body of literature that concentrates on the non-stationarity issues arising from having multiple agents learning simultaneously in the same environment (Laurent et al., 2011; In contrast, Foerster et al. (2018a) add an extra term to The works by Lowe et al. (2017) and Foerster The works by de Witt et al. (2020) and Y u et al. (2021) show that Y u et al. attribute the positive empirical results to the clipping parameter Global simulator, observation functions, and joint policy for n 0, ...,N/T do s The bar plots show the total runtime of training for 4M timesteps with the three simulators.




Scalable and Robust Speculative Decoding

Neural Information Processing Systems

As the usage of large language models (LLMs) grows, it becomes increasingly important to serve them quickly and efficiently. While speculative decoding has recently emerged as a promising direction for accelerating LLM serving, existing methods are limited in their ability to scale to larger speculation budgets and adapt to different hyperparameters.


DELTA: Diverse Client Sampling for Fasting Federated Learning Lin Wang 1,2, Tao Lin 4,5

Neural Information Processing Systems

Partial client participation has been widely adopted in Federated Learning (FL) to reduce the communication burden efficiently. However, an inadequate client sampling scheme can lead to the selection of unrepresentative subsets, resulting in significant variance in model updates and slowed convergence. Existing sampling methods are either biased or can be further optimized for faster convergence. In this paper, we present DELTA, an unbiased sampling scheme designed to alleviate these issues. DELTA characterizes the effects of client diversity and local variance, and samples representative clients with valuable information for global model updates. In addition, DELTA is a proven optimal unbiased sampling scheme that minimizes variance caused by partial client participation and outperforms other unbiased sampling schemes in terms of convergence. Furthermore, to address full-client gradient dependence, we provide a practical version of DELTA depending on the available clients' information, and also analyze its convergence. Our results are validated through experiments on both synthetic and real-world datasets.


GSDF: 3DGS Meets SDF for Improved Neural Rendering and Reconstruction

Neural Information Processing Systems

Representing 3D scenes from multiview images remains a core challenge in computer vision and graphics, requiring both reliable rendering and reconstruction, which often conflicts due to the mismatched prioritization of image quality over precise underlying scene geometry. Although both neural implicit surfaces and explicit Gaussian primitives have advanced with neural rendering techniques, current methods impose strict constraints on density fields or primitive shapes, which enhances the affinity for geometric reconstruction at the sacrifice of rendering quality. To address this dilemma, we introduce GSDF, a dual-branch architecture combining 3D Gaussian Splatting (3DGS) and neural Signed Distance Fields (SDF). Our approach leverages mutual guidance and joint supervision during the training process to mutually enhance reconstruction and rendering. Specifically, our method guides the Gaussian primitives to locate near potential surfaces and accelerates the SDF convergence. This implicit mutual guidance ensures robustness and accuracy in both synthetic and real-world scenarios. Experimental results demonstrate that our method boosts the SDF optimization process to reconstruct more detailed geometry, while reducing floaters and blurry edge artifacts in rendering by aligning Gaussian primitives with the underlying geometry.


A Missing Preliminaries

Neural Information Processing Systems

A.1 Fractional Optimal and the Incremental-bang-per-buck Algorithm [SZ79] introduce the notion of Dominated and LP dominated ads and show that they are not used in the fractional optimal solution. For an advertiser i, the rich ads of the advertiser can be dominated in two ways. Moroever, [SZ79] showed that the fractional optimal solution can be obtained through the following incremental-bang-per-buck algorithm. Algorithm 2. Eliminate all the Dominated and LP-dominated rich ads for each advertiser. Otherwise, allocate the advertiser fractionally in the remaining space. This fractional allocation puts a weight x on the previously allocated rich ad of the advertiser and (1 x) on the newly selected rich-ad such that the fractional space equals remaining space plus previously allocated space of the advertiser. We prove some lemmas about the optimal fractional solution.


Simple Mechanisms for Welfare Maximization in Rich Advertising Auctions

Neural Information Processing Systems

Internet ad auctions have evolved from a few lines of text to richer informational layouts that include images, sitelinks, videos, etc. Ads in these new formats occupy varying amounts of space, and an advertiser can provide multiple formats, only one of which can be shown. The seller is now faced with a multi-parameter mechanism design problem. Computing an efficient allocation is computationally intractable, and therefore the standard Vickrey-Clarke-Groves (VCG) auction, while truthful and welfare-optimal, is impractical. In this paper, we tackle a fundamental problem in the design of modern ad auctions. We adopt a "Myersonian" approach and study allocation rules that are monotone both in the bid and set of rich ads.