Goto

Collaborating Authors

 United States Government


Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices

Neural Information Processing Systems

We present an extension of the conditional gradient method to problems whose feasible sets are convex cones. We provide a convergence analysis for the method and for variants with nonconvex objectives, and we extend the analysis to practical cases with effective line search strategies. For the specific case of the positive semidefinite cone, we present a memory-efficient version based on randomized matrix sketches and advocate a heuristic greedy step that greatly improves its practical performance. Numerical results on phase retrieval and matrix completion problems indicate that our method can offer substantial advantages over traditional conditional gradient and Burer-Monteiro approaches.


Conic Descent and its Application to Memory-efficient Optimization over Positive Semidefinite Matrices

Neural Information Processing Systems

We present an extension of the conditional gradient method to problems whose feasible sets are convex cones. We provide a convergence analysis for the method and for variants with nonconvex objectives, and we extend the analysis to practical cases with effective line search strategies. For the specific case of the positive semidefinite cone, we present a memory-efficient version based on randomized matrix sketches and advocate a heuristic greedy step that greatly improves its practical performance. Numerical results on phase retrieval and matrix completion problems indicate that our method can offer substantial advantages over traditional conditional gradient and Burer-Monteiro approaches.


Adversarial training for free!

Neural Information Processing Systems

Adversarial training, in which a network is trained on adversarial examples, is one of the few defenses against adversarial attacks that withstands strong attacks. Unfortunately, the high cost of generating strong adversarial examples makes standard adversarial training impractical on large-scale problems like ImageNet. We present an algorithm that eliminates the overhead cost of generating adversarial examples by recycling the gradient information computed when updating model parameters. Our "free" adversarial training algorithm achieves comparable robustness to PGD adversarial training on the CIFAR-10 and CIFAR-100 datasets at negligible additional cost compared to natural training, and can be 7 to 30 times faster than other strong adversarial training methods. Using a single workstation with 4 P100 GPUs and 2 days of runtime, we can train a robust model for the large-scale ImageNet classification task that maintains 40% accuracy against PGD attacks.


A Scalable Approach for Privacy-Preserving Collaborative Machine Learning

Neural Information Processing Systems

We consider a collaborative learning scenario in which multiple data-owners wish to jointly train a logistic regression model, while keeping their individual datasets private from the other parties. We propose COPML, a fully-decentralized training framework that achieves scalability and privacy-protection simultaneously. The key idea of COPML is to securely encode the individual datasets to distribute the computation load effectively across many parties and to perform the training computations as well as the model updates in a distributed manner on the securely encoded data. We provide the privacy analysis of COPML and prove its convergence. Furthermore, we experimentally demonstrate that COPML can achieve significant speedup in training over the benchmark protocols. Our protocol provides strong statistical privacy guarantees against colluding parties (adversaries) with unbounded computational power, while achieving up to 16 speedup in the training time against the benchmark protocols.



Refining Language Models with Compositional Explanations

Neural Information Processing Systems

Pre-trained language models have been successful on text classification tasks, but are prone to learning spurious correlations from biased datasets, and are thus vulnerable when making inferences in a new domain. Prior work reveals such spurious patterns via post-hoc explanation algorithms which compute the importance of input features. Further, the model is regularized to align the importance scores with human knowledge, so that the unintended model behaviors are eliminated. However, such a regularization technique lacks flexibility and coverage, since only importance scores towards a pre-defined list of features are adjusted, while more complex human knowledge such as feature interaction and pattern generalization can hardly be incorporated. In this work, we propose to refine a learned language model for a target domain by collecting human-provided compositional explanations regarding observed biases. By parsing these explanations into executable logic rules, the human-specified refinement advice from a small set of explanations can be generalized to more training examples. We additionally introduce a regularization term allowing adjustments for both importance and interaction of features to better rectify model behavior.


Efficient Probabilistic Inference in the Quest for Physics Beyond the Standard Model Atฤฑlฤฑm GรผneลŸ Baydin, 1

Neural Information Processing Systems

We present a novel probabilistic programming framework that couples directly to existing large-scale simulators through a cross-platform probabilistic execution protocol, which allows general-purpose inference engines to record and control random number draws within simulators in a language-agnostic way. The execution of existing simulators as probabilistic programs enables highly interpretable posterior inference in the structured model defined by the simulator code base. We demonstrate the technique in particle physics, on a scientifically accurate simulation of the ฯ„ (tau) lepton decay, which is a key ingredient in establishing the properties of the Higgs boson. Inference efficiency is achieved via inference compilation where a deep recurrent neural network is trained to parameterize proposal distributions and control the stochastic simulator in a sequential importance sampling scheme, at a fraction of the computational cost of a Markov chain Monte Carlo baseline.


Synthetic Design: An Optimization Approach to Experimental Design with Synthetic Controls

Neural Information Processing Systems

We investigate the optimal design of experimental studies that have pre-treatment outcome data available. The average treatment effect is estimated as the difference between the weighted average outcomes of the treated and control units. A number of commonly used approaches fit this formulation, including the difference-inmeans estimator and a variety of synthetic-control techniques. We propose several methods for choosing the set of treated units in conjunction with the weights. Observing the NP-hardness of the problem, we introduce a mixed-integer programming formulation which selects both the treatment and control sets and unit weightings. We prove that these proposed approaches lead to qualitatively different experimental units being selected for treatment. We use simulations based on publicly available data from the US Bureau of Labor Statistics that show improvements in terms of mean squared error and statistical power when compared to simple and commonly used alternatives such as randomized trials.


The Morning After: Everything Samsung announced this week (and future devices teased)

Engadget

Welcome to a new newsletter, with a bit of a new direction. While our mid-week edition tackles news specifics, this end-of-the-week missive combines the biggest news with more context, more things to read and watch, recommendations, easter eggs, inside baseball and stuff that interests our readers, alongside the breaking news, reviews and features you expect from Engadget. We'd love your feedback on what you'd like to see covered in these meatier editions -- hit me up at tma(at)engadget.com. Luckily for me, we kick things off with Samsung's big Unpacked event, launching three new phones and teasing two -- yes, two! -- more coming soon. Everything Samsung announced, including prices and launch dates (February 8 -- I'll save you a click), we collated here, but it was largely a fallow year for Galaxy S hardware, barring a substantially more powerful chip.


Dating Apps Promise to Remain a Rare Haven Following Trump's Executive Order

WIRED

Mere moments after his swearing in Monday, President Donald Trump made a proclamation to attendees of his inauguration: "It shall henceforth be the policy of the United States government that there are only two genders: male and female." Trump then signed an executive order disparaging what the White House called "gender ideology" and claiming that a person's sex is "not changeable and [is] grounded in fundamental and incontrovertible reality." Trump's order, which was widely seen as an unscientific attempt to roll back the rights of transgender and gender-expansive people, also instructs federal agencies "to require that government-issued identification documents, including passports, visas, and Global Entry cards, accurately reflect the holder's sex," rather than their gender identity. It was one of 78 orders signed on Monday, some of which were part of Trump's attempts to end Biden-era policies that "socially engineer race and gender into every aspect of public and private life." While the executive order only affects federal policy, the broader implications are vast.