Goto

Collaborating Authors

 proposal


There's a Long Shot Proposal to Protect California Workers From AI

WIRED

California gubernatorial candidate Tom Steyer is proposing a new jobs guarantee for workers displaced by artificial intelligence. Billionaire California gubernatorial candidate Tom Steyer is rolling out a new proposal that would guarantee jobs with benefits for workers displaced by artificial intelligence . He's the first state-wide candidate to make such a pledge. The plan, which builds on a broader AI policy framework Steyer released in March, promises to make California "the first major economy in the world" to ensure "good-paying" jobs to workers impacted by AI. To do so, Steyer tells WIRED he plans to build off a previous proposal to introduce a "token tax" which would tax big tech companies "a fraction of a cent for every unit of data processed" for AI.


GameStop makes 55.5bn takeover offer for eBay

The Guardian

GameStop's CEO said he could turn eBay into something worth hundreds of billions of dollars. GameStop's CEO said he could turn eBay into something worth hundreds of billions of dollars. GameStop makes $55.5bn takeover offer for eBay Video game retailer's CEO warns that unsolicited bid could turn hostile if it is rebuffed by resale site's board US video games retailer GameStop has offered to buy eBay for $55.5bn (£41bn) in an unsolicited bid that its boss warned could turn hostile if the proposal is rebuffed by eBay's board. GameStop, which has quietly accumulated a 5% stake in eBay, said it was willing to pay $125 a share, split 50-50 between cash and stock. It is an ambitious move by the games company, which catapulted to fame during the meme-stock craze of 2021 but is worth far less than its takeover target.


Bayesian experimental design: grouped geometric pooled posterior via ensemble Kalman methods

Yang, Huchen, Dong, Xinghao, Wu, Jinlong

arXiv.org Machine Learning

Bayesian experimental design (BED) for complex physical systems is often limited by the nested inference required to estimate the expected information gain (EIG) or its gradients. Each outer sample induces a different posterior, creating a large and heterogeneous set of inference targets. Existing methods have to sacrifice either accuracy or efficiency: they either perform per-outer-sample posterior inference, which yields higher fidelity but at prohibitive computational cost, or amortize the inner inference across all outer samples for computational reuse, at the risk of degraded accuracy under posterior heterogeneity. To improve accuracy and maintain cost at the amortized level, we propose a grouped geometric pooled posterior framework that partitions outer samples into groups and constructs a pooled proposal for each group. While such grouping strategy would normally require generating separate proposal samples for different groups, our tailored ensemble Kalman inversion (EKI) formulation generates these samples without extra forward-model evaluation cost. We also introduce a conservative diagnostic to assess importance-sampling quality to guide grouping. This grouping strategy improves within-group proposal-target alignment, yielding more accurate and stable estimators while keeping the cost comparable to amortized approaches. We evaluate the performance of our method on both Gaussian-linear and high-dimensional network-based model discrepancy calibration problems.


A proposal for PU classification under Non-SCAR using clustering and logistic model

Furmanczyk, Konrad, Paczutkowski, Kacper

arXiv.org Machine Learning

The present study aims to investigate a cluster cleaning algorithm that is both computationally simple and capable of solving the PU classification when the SCAR condition is unsatisfied. A secondary objective of this study is to determine the robustness of the LassoJoint method to perturbations of the SCAR condition. In the first step of our algorithm, we obtain cleaning labels from 2-means clustering. Subsequently, we perform logistic regression on the cleaned data, assigning positive labels from the cleaning algorithm with additional true positive observations. The remaining observations are assigned the negative label. The proposed algorithm is evaluated by comparing 11 real data sets from machine learning repositories and a synthetic set. The findings obtained from this study demonstrate the efficacy of the clustering algorithm in scenarios where the SCAR condition is violated and further underscore the moderate robustness of the LassoJoint algorithm in this context.


How to Approximate Inference with Subtractive Mixture Models

Zellinger, Lena, Branchini, Nicola, De Smet, Lennert, Elvira, Víctor, Malkin, Nikolay, Vergari, Antonio

arXiv.org Machine Learning

Classical mixture models (MMs) are widely used tractable proposals for approximate inference settings such as variational inference (VI) and importance sampling (IS). Recently, mixture models with negative coefficients, called subtractive mixture models (SMMs), have been proposed as a potentially more expressive alternative. However, how to effectively use SMMs for VI and IS is still an open question as they do not provide latent variable semantics and therefore cannot use sampling schemes for classical MMs. In this work, we study how to circumvent this issue by designing several expectation estimators for IS and learning schemes for VI with SMMs, and we empirically evaluate them for distribution approximation. Finally, we discuss the additional challenges in estimation stability and learning efficiency that they carry and propose ways to overcome them. Code is available at: https://github.com/april-tools/delta-vi.


Graph Energy Matching: Transport-Aligned Energy-Based Modeling for Graph Generation

Balcerak, Michal, Shit, Suprosana, Prabhakar, Chinmay, Kaltenbach, Sebastian, Albergo, Michael S., Du, Yilun, Menze, Bjoern

arXiv.org Machine Learning

Energy-based models for discrete domains, such as graphs, explicitly capture relative likelihoods, naturally enabling composable probabilistic inference tasks like conditional generation or enforcing constraints at test-time. However, discrete energy-based models typically struggle with efficient and high-quality sampling, as off-support regions often contain spurious local minima, trapping samplers and causing training instabilities. This has historically resulted in a fidelity gap relative to discrete diffusion models. We introduce Graph Energy Matching (GEM), a generative framework for graphs that closes this fidelity gap. Motivated by the transport map optimization perspective of the Jordan-Kinderlehrer-Otto (JKO) scheme, GEM learns a permutation-invariant potential energy that simultaneously provides transport-aligned guidance from noise toward data and refines samples within regions of high data likelihood. Further, we introduce a sampling protocol that leverages an energy-based switch to seamlessly bridge: (i) rapid, gradient-guided transport toward high-probability regions to (ii) a mixing regime for exploration of the learned graph distribution. On molecular graph benchmarks, GEM matches or exceeds strong discrete diffusion baselines. Beyond sample quality, explicit modeling of relative likelihood enables targeted exploration at inference time, facilitating compositional generation, property-constrained sampling, and geodesic interpolation between graphs.


Explainable cluster analysis: a bagging approach

Quetti, Federico Maria, Ballante, Elena, Figini, Silvia, Giudici, Paolo

arXiv.org Machine Learning

A major limitation of clustering approaches is their lack of explainability: methods rarely provide insight into which features drive the grouping of similar observations. To address this limitation, we propose an ensemble-based clustering framework that integrates bagging and feature dropout to generate feature importance scores, in analogy with feature importance mechanisms in supervised random forests. By leveraging multiple bootstrap resampling schemes and aggregating the resulting partitions, the method improves stability and robustness of the cluster definition, particularly in small-sample or noisy settings. Feature importance is assessed through an information-theoretic approach: at each step, the mutual information between each feature and the estimated cluster labels is computed and weighted by a measure of clustering validity to emphasize well-formed partitions, before being aggregated into a final score. The method outputs both a consensus partition and a corresponding measure of feature importance, enabling a unified interpretation of clustering structure and variable relevance. Its effectiveness is demonstrated on multiple simulated and real-world datasets.


I Believe in one God, and It's Not a Computer

Mother Jones

How the data center boom plunged one small Pennsylvania town into chaos. Valley View Estates is set to be surrounded by data centers. Get your news from a source that's not owned and controlled by oligarchs. "I don't like to see anyone upset," said Nick Farris of Provident Real Estate Advisors. He was sitting in the front of a crowd of roughly 150 inside Valley View High School's auditorium in Archbald, a town of about 7,500, huddled between two mountain ranges in Pennsylvania's Lackawanna Valley. Farris was there to represent the developer for Project Scott, one of many data center campuses coming to town. "I think that this is the best data center site in this area of the country, by far." The audience had been fairly quiet, bundled in thick coats against the late January cold. But as Farris spoke about data centers as a boon for communities, they began to laugh, drawing a rebuke from town officials. "What about the children?" someone shouted from the crowd. The children were watching from the walls; long banners of Valley View Performing Arts students hanging around the auditorium like championship pennants. Project Scott and four other data facilities will sit just a few thousand feet from the middle and high schools. He was referring to Lockheed Martin's 350,000-square-foot Missiles and Fire Control facility directly next to the high school, parts of which are highly contaminated . "That sucks too!" another attendee yelled back.


A-NICE-MC: Adversarial Training for MCMC

Neural Information Processing Systems

Existing Markov Chain Monte Carlo (MCMC) methods are either based on general-purpose and domain-agnostic schemes, which can lead to slow convergence, or require hand-crafting of problem-specific proposals by an expert. We propose A-NICE-MC, a novel method to train flexible parametric Markov chain kernels to produce samples with desired properties. First, we propose an efficient likelihood-free adversarial training method to train a Markov chain and mimic a given data distribution. Then, we leverage flexible volume preserving flows to obtain parametric kernels for MCMC. Using a bootstrap approach, we show how to train efficient Markov Chains to sample from a prescribed posterior distribution by iteratively improving the quality of both the model and the samples. A-NICE-MC provides the first framework to automatically design efficient domain-specific MCMC proposals. Empirical results demonstrate that A-NICE-MC combines the strong guarantees of MCMC with the expressiveness of deep neural networks, and is able to significantly outperform competing methods such as Hamiltonian Monte Carlo.


Dynamic Importance Sampling for Anytime Bounds of the Partition Function

Neural Information Processing Systems

Computing the partition function is a key inference task in many graphical models. In this paper, we propose a dynamic importance sampling scheme that provides anytime finite-sample bounds for the partition function. Our algorithm balances the advantages of the three major inference strategies, heuristic search, variational bounds, and Monte Carlo methods, blending sampling with search to refine a variationally defined proposal. Our algorithm combines and generalizes recent work on anytime search and probabilistic bounds of the partition function. By using an intelligently chosen weighted average over the samples, we construct an unbiased estimator of the partition function with strong finite-sample confidence intervals that inherit both the rapid early improvement rate of sampling and the long-term benefits of an improved proposal from search. This gives significantly improved anytime behavior, and more flexible trade-offs between memory, time, and solution quality. We demonstrate the effectiveness of our approach empirically on real-world problem instances taken from recent UAI competitions.