Plotting

 Technology


We Bought a 'Peeing' Robot Attack Dog From Temu. It Was Even Weirder Than Expected

WIRED

In my 15 years of reviewing tech, this pellet-firing, story-telling, pretend-urinating robot attack dog is easily the strangest thing I've ever tested. Arriving in a slightly battered box following a series of questionable decisions on Temu, I'm immediately drawn to the words "FIRE BULLETS PET" emblazoned on the box. And there, resting behind the protective plastic window with all the innocence of a newborn lamb, lies the plastic destroyer of worlds that my four-and-a-half-year-old immediately (and inexplicably), names Clippy. Clippy is a robot dog. And he (my son assures me that it's a he), is clearly influenced by the remarkable, and somewhat terrifying, robotic canine creations of Boston Dynamics--a renowned company that's leading the robot revolution.


Lean Workbook: A large-scale Lean problem set formalized from natural language math problems

Neural Information Processing Systems

Large language models have demonstrated impressive capabilities across various natural language processing tasks, especially in solving mathematical problems. However, large language models are not good at math theorem proving using formal languages like Lean. A significant challenge in this area is the scarcity of training data available in these formal languages. To address this issue, we propose a novel pipeline that iteratively generates and filters synthetic data to translate natural language mathematical problems into Lean 4 statements, and vice versa. Our results indicate that the synthetic data pipeline can provide useful training data and improve the performance of LLMs in translating and understanding complex mathematical problems and proofs. Our final dataset contains about 57K formal-informal question pairs along with searched proof from the math contest forum and 21 new IMO questions.


Ask, Attend, Attack: An Effective Decision-Based Black-Box Targeted Attack for Image-to-Text Models, and Min Jiang

Neural Information Processing Systems

While image-to-text models have demonstrated significant advancements in various vision-language tasks, they remain susceptible to adversarial attacks. Existing white-box attacks on image-to-text models require access to the architecture, gradients, and parameters of the target model, resulting in low practicality. Although the recently proposed gray-box attacks have improved practicality, they suffer from semantic loss during the training process, which limits their targeted attack performance. To advance adversarial attacks of image-to-text models, this paper focuses on a challenging scenario: decision-based black-box targeted attacks where the attackers only have access to the final output text and aim to perform targeted attacks. Specifically, we formulate the decision-based black-box targeted attack as a large-scale optimization problem.


Off-Policy Evaluation via Off-Policy Classification Alex Irpan

Neural Information Processing Systems

In this work, we consider the problem of model selection for deep reinforcement learning (RL) in real-world environments. Typically, the performance of deep RL algorithms is evaluated via on-policy interactions with the target environment. However, comparing models in a real-world environment for the purposes of early stopping or hyperparameter tuning is costly and often practically infeasible. This leads us to examine off-policy policy evaluation (OPE) in such settings. We focus on OPE for value-based methods, which are of particular interest in deep RL, with applications like robotics, where off-policy algorithms based on Q-function estimation can often attain better sample complexity than direct policy optimization. Existing OPE metrics either rely on a model of the environment, or the use of importance sampling (IS) to correct for the data being off-policy.


stochastic case, and we present an empirical validation of our method on stochastic tasks

Neural Information Processing Systems

All Reviewers: Thank you for the review. This simplification was only used for the theoretical analysis. While we cannot include a full proof for the stochastic case, a proof sketch follows. Stochastic dynamics only influence the lower bound on return. We modify the Tree environment to execute a random action instead of the policy's action with probability With small probability, the environment repeats the previous action instead of the policy's action.


Dual Variational Generation for Low Shot Heterogeneous Face Recognition

Neural Information Processing Systems

Heterogeneous Face Recognition (HFR) is a challenging issue because of the large domain discrepancy and a lack of heterogeneous data. This paper considers HFR as a dual generation problem, and proposes a novel Dual Variational Generation (DVG) framework. It generates large-scale new paired heterogeneous images with the same identity from noise, for the sake of reducing the domain gap of HFR. Specifically, we first introduce a dual variational autoencoder to represent a joint distribution of paired heterogeneous images. Then, in order to ensure the identity consistency of the generated paired heterogeneous images, we impose a distribution alignment in the latent space and a pairwise identity preserving in the image space. Moreover, the HFR network reduces the domain discrepancy by constraining the pairwise feature distances between the generated paired heterogeneous images. Extensive experiments on four HFR databases show that our method can significantly improve state-of-the-art results.



Diffusion Models With Learned Adaptive Noise

Neural Information Processing Systems

Diffusion models have gained traction as powerful algorithms for synthesizing highquality images. Central to these algorithms is the diffusion process, a set of equations which maps data to noise in a way that can significantly affect performance. In this paper, we explore whether the diffusion process can be learned from data. Our work is grounded in Bayesian inference and seeks to improve log-likelihood estimation by casting the learned diffusion process as an approximate variational posterior that yields a tighter lower bound (ELBO) on the likelihood.


Quality Aware Generative Adversarial Networks

Neural Information Processing Systems

Generative Adversarial Networks (GANs) have become a very popular tool for implicitly learning high-dimensional probability distributions. Several improvements have been made to the original GAN formulation to address some of its shortcomings like mode collapse, convergence issues, entanglement, poor visual quality etc. While a significant effort has been directed towards improving the visual quality of images generated by GANs, it is rather surprising that objective image quality metrics have neither been employed as cost functions nor as regularizers in GAN objective functions. In this work, we show how a distance metric that is a variant of the Structural SIMilarity (SSIM) index (a popular full-reference image quality assessment algorithm), and a novel quality aware discriminator gradient penalty function that is inspired by the Natural Image Quality Evaluator (NIQE, a popular no-reference image quality assessment algorithm) can each be used as excellent regularizers for GAN objective functions. Specifically, we demonstrate state-ofthe-art performance using the Wasserstein GAN gradient penalty (WGAN-GP) framework over CIFAR-10, STL10 and CelebA datasets.


Deep Active Learning with a Neural Architecture Search

Neural Information Processing Systems

We consider active learning of deep neural networks. Most active learning works in this context have focused on studying effective querying mechanisms and assumed that an appropriate network architecture is a priori known for the problem at hand. We challenge this assumption and propose a novel active strategy whereby the learning algorithm searches for effective architectures on the fly, while actively learning. We apply our strategy using three known querying techniques (softmax response, MC-dropout, and coresets) and show that the proposed approach overwhelmingly outperforms active learning using fixed architectures.