Goto

Collaborating Authors

 Romania


Local Linearity of LLMs Enables Activation Steering via Model-Based Linear Optimal Control

Skifstad, Julian, Yang, Xinyue Annie, Chou, Glen

arXiv.org Machine Learning

Inference-time LLM alignment methods, particularly activation steering, offer an alternative to fine-tuning by directly modifying activations during generation. Existing methods, however, often rely on non-anticipative interventions that ignore how perturbations propagate through transformer layers and lack online error feedback, resulting in suboptimal, open-loop control. To address this, we show empirically that, despite the nonlinear structure of transformer blocks, layer-wise dynamics across multiple LLM architectures and scales are well-approximated by locally-linear models. Exploiting this property, we model LLM inference as a linear time-varying dynamical system and adapt the classical linear quadratic regulator to compute feedback controllers using layer-wise Jacobians, steering activations toward desired semantic setpoints in closed-loop with minimal computational overhead and no offline training. We also derive theoretical bounds on setpoint tracking error, enabling formal guarantees on steering performance. Using a novel adaptive semantic feature setpoint signal, our method yields robust, fine-grained behavior control across models, scales, and tasks, including state-of-the-art modulation of toxicity, truthfulness, refusal, and arbitrary concepts, surpassing baseline steering methods. Our code is available at: https://github.com/trustworthyrobotics/lqr-activation-steering


Can you best a math Olympiad? Test your skills with the world's largest database of problems.

Popular Science

MathNet contains 30,000 free math problems collected over half a century. More information Adding us as a Preferred Source in Google by using this link indicates that you would like to see more of our content in Google News results. The International Mathematical Olympiad was first held in Romania in 1959. Breakthroughs, discoveries, and DIY tips sent six days a week. In 1959, countries around the world sent their most talented students to Romania to compete in the first-ever International Mathematical Olympiad (IMO).


Experimental Design for Missing Physics

Strouwen, Arno, Micluţa-Câmpeanu, Sebastián

arXiv.org Machine Learning

For most process systems, knowledge of the model structure is incomplete. This missing physics must then be learned from experimental data. Recently, a combination of universal differential equations and symbolic regression has become a popular tool to discover these missing physics. Universal differential equations employ neural networks to represent missing parts of the model structure, and symbolic regression aims to make these neural networks interpretable. These machine learning techniques require high-quality data to successfully recover the true model structure. To gather such informative data, a sequential experimental design technique is developed which is based on optimally discriminating between the plausible model structures suggested by symbolic regression. This technique is then applied to discovering the missing physics of a bioreactor.


RG-TTA: Regime-Guided Meta-Control for Test-Time Adaptation in Streaming Time Series

Kumar, Indar, Tiwari, Akanksha, Jasti, Sai Krishna, Lade, Ankit Hemant

arXiv.org Machine Learning

Test-time adaptation (TTA) enables neural forecasters to adapt to distribution shifts in streaming time series, but existing methods apply the same adaptation intensity regardless of the nature of the shift. We propose Regime-Guided Test-Time Adaptation (RG-TTA), a meta-controller that continuously modulates adaptation intensity based on distributional similarity to previously-seen regimes. Using an ensemble of Kolmogorov-Smirnov, Wasserstein-1, feature-distance, and variance-ratio metrics, RG-TTA computes a similarity score for each incoming batch and uses it to (i) smoothly scale the learning rate -- more aggressive for novel distributions, conservative for familiar ones -- and (ii) control gradient effort via loss-driven early stopping rather than fixed budgets, allowing the system to allocate exactly the effort each batch requires. As a supplementary mechanism, RG-TTA gates checkpoint reuse from a regime memory, loading stored specialist models only when they demonstrably outperform the current model (loss improvement >= 30%). RG-TTA is model-agnostic and strategy-composable: it wraps any forecaster exposing train/predict/save/load interfaces and enhances any gradient-based TTA method. We demonstrate three compositions -- RG-TTA, RG-EWC, and RG-DynaTTA -- and evaluate 6 update policies (3 baselines + 3 regime-guided variants) across 4 compact architectures (GRU, iTransformer, PatchTST, DLinear), 14 datasets (6 real-world multivariate benchmarks + 8 synthetic regime scenarios), and 4 forecast horizons (96, 192, 336, 720) under a streaming evaluation protocol with 3 random seeds (672 experiments total). Regime-guided policies achieve the lowest MSE in 156 of 224 seed-averaged experiments (69.6%), with RG-EWC winning 30.4% and RG-TTA winning 29.0%. Overall, RG-TTA reduces MSE by 5.7% vs TTA while running 5.5% faster; RG-EWC reduces MSE by 14.1% vs standalone EWC.


Deep Adaptive Model-Based Design of Experiments

Strouwen, Arno, Micluţa-Câmpeanu, Sebastian

arXiv.org Machine Learning

Model-based design of experiments (MBDOE) is essential for efficient parameter estimation in nonlinear dynamical systems. However, conventional adaptive MBDOE requires costly posterior inference and design optimization between each experimental step, precluding real-time applications. We address this by combining Deep Adaptive Design (DAD), which amortizes sequential design into a neural network policy trained offline, with differentiable mechanistic models. For dynamical systems with known governing equations but uncertain parameters, we extend sequential contrastive training objectives to handle nuisance parameters and propose a transformer-based policy architecture that respects the temporal structure of dynamical systems. We demonstrate the approach on four systems of increasing complexity: a fed-batch bioreactor with Monod kinetics, a Haldane bioreactor with uncertain substrate inhibition, a two-compartment pharmacokinetic model with nuisance clearance parameters, and a DC motor for real-time deployment.


Algorithmic warm starts for Hamiltonian Monte Carlo

Zhang, Matthew S., Altschuler, Jason M., Chewi, Sinho

arXiv.org Machine Learning

Generating samples from a continuous probability density is a central algorithmic problem across statistics, engineering, and the sciences. For high-dimensional settings, Hamiltonian Monte Carlo (HMC) is the default algorithm across mainstream software packages. However, despite the extensive line of work on HMC and its widespread empirical success, it remains unclear how many iterations of HMC are required as a function of the dimension $d$. On one hand, a variety of results show that Metropolized HMC converges in $O(d^{1/4})$ iterations from a warm start close to stationarity. On the other hand, Metropolized HMC is significantly slower without a warm start, e.g., requiring $Ω(d^{1/2})$ iterations even for simple target distributions such as isotropic Gaussians. Finding a warm start is therefore the computational bottleneck for HMC. We resolve this issue for the well-studied setting of sampling from a probability distribution satisfying strong log-concavity (or isoperimetry) and third-order derivative bounds. We prove that \emph{non-Metropolized} HMC generates a warm start in $\tilde{O}(d^{1/4})$ iterations, after which we can exploit the warm start using Metropolized HMC. Our final complexity of $\tilde{O}(d^{1/4})$ is the fastest algorithm for high-accuracy sampling under these assumptions, improving over the prior best of $\tilde{O}(d^{1/2})$. This closes the long line of work on the dimensional complexity of MHMC for such settings, and also provides a simple warm-start prescription for practical implementations.


Privacy-Preserving Reinforcement Learning from Human Feedback via Decoupled Reward Modeling

Cho, Young Hyun, Sun, Will Wei

arXiv.org Machine Learning

Preference-based fine-tuning has become an important component in training large language models, and the data used at this stage may contain sensitive user information. A central question is how to design a differentially private pipeline that is well suited to the distinct structure of reinforcement learning from human feedback. We propose a privacy-preserving framework that imposes differential privacy only on reward learning and derives the final policy from the resulting private reward model. Theoretically, we study the suboptimality gap and show that privacy contributes an additional additive term beyond the usual non-private statistical error. We also establish a minimax lower bound and show that the dominant term changes with sample size and privacy level, which in turn characterizes regimes in which the upper bound is rate-optimal up to logarithmic factors. Empirically, synthetic experiments confirm the scaling predicted by the theory, and experiments on the Anthropic HH-RLHF dataset using the Gemma-2B-IT model show stronger private alignment performance than existing differentially private baseline methods across privacy budgets.


How Doodles Became the Dog du Jour

The New Yorker

Poodle crossbreeds have grown overwhelmingly popular, sparking controversy in dog parks and kennel clubs alike. The features of doodles such as Peaches (above), a goldendoodle, have become the canine equivalent of Instagram face. Meet the Breeds, the American Kennel Club's annual showcase of purebred dogs, took place over two eye-wateringly cold days in early February at the Javits Center, in Manhattan. About a hundred and fifty of the two hundred and five varieties recognized as official breeds by the A.K.C., the long-standing authority in the U.S. dog world, were in attendance for the public to ogle, fondle, and coo "So cute!" to, including the basset fauve de Bretagne, a hunting hound from France that's one of three newly recognized breeds recently allowed into the purebred pantheon. Some of the dogs had competed in the Westminster Kennel Club Dog Show earlier in the week, and past champions had their ribbons on display. In spite of the frigid weather, pavilions hosting the more popular breeds--the pug, the Doberman pinscher, the Great Dane, the St. Bernard--were packed. Lesser-known varieties, such as the saluki, the Löwchen, and the Lapponian herder, drew sparser crowds. There were exhibition spaces for each breed, and on the back walls were three adjectives supposedly describing that particular type of dog's temperament. There is, in fact, no evidence that temperament is consistent within a breed, but the idea is deeply rooted in dogdom. I stopped to caress the velvety ear leather of a pharaoh hound ("Friendly, Smart, Noble"), a sprinting breed once used to hunt rabbits in Malta; accept kisses from a Portuguese water dog, bred to assist with retrieving tackle ("Affectionate, Adventurous, Athletic"); and have my photograph taken with a Leonberger, a German breed from the town of Leonberg, in southwest Germany ("Friendly, Gentle, Playful"). No one was supposed to be openly selling dogs, but, if you asked, the breeders would share their information. Excluding what are known as companion dogs, like the Leonberger, most of the animals at the show were designed for a purpose that is no longer required of them. In Great Britain, foxhounds are legally barred from chasing foxes. Consider the fate of the otterhound, an ancient variety with a noble heritage which was once used in the U.K. to hunt river otters, which were prized for their thick fur and disliked by wealthy landowners because they ate fish in their stocked ponds.