Data augmentation for efficient learning from parametric experts
We present a simple, yet powerful data-augmentation technique to enable data-efficient learning from parametric experts for reinforcement and imitation learning. We focus on what we call the policy cloning setting, in which we use online or offline queries of an expert or expert policy to inform the behavior of a student policy. This setting arises naturally in a number of problems, for instance as variants of behavior cloning, or as a component of other algorithms such as DAGGER, policy distillation or KL-regularized RL. Our approach, augmented policy cloning (APC), uses synthetic states to induce feedback-sensitivity in a region around sampled trajectories, thus dramatically reducing the environment interactions required for successful cloning of the expert. We achieve highly data-efficient transfer of behavior from an expert to a student policy for high-degrees-of-freedom control problems.
PAC-Bayes Compression Bounds So Tight That They Can Explain Generalization
While there has been progress in developing non-vacuous generalization bounds for deep neural networks, these bounds tend to be uninformative about why deep learning works. In this paper, we develop a compression approach based on quantizing neural network parameters in a linear subspace, profoundly improving on previous results to provide state-of-the-art generalization bounds on a variety of tasks, including transfer learning. We use these tight bounds to better understand the role of model size, equivariance, and the implicit biases of optimization, for generalization in deep learning. Notably, we find large models can be compressed to a much greater extent than previously known, encapsulating Occam's razor.
A Unified Sequence Interface for Vision Tasks
While language tasks are naturally expressed in a single, unified, modeling framework, i.e., generating sequences of tokens, this has not been the case in computer vision. As a result, there is a proliferation of distinct architectures and loss functions for different vision tasks. In this work we show that a diverse set of "core" computer vision tasks can also be unified if formulated in terms of a shared pixel-to-sequence interface. We focus on four tasks, namely, object detection, instance segmentation, keypoint detection, and image captioning, all with diverse types of outputs, e.g., bounding boxes or dense masks. Despite that, by formulating the output of each task as a sequence of discrete tokens with a unified interface, we show that one can train a neural network with a single model architecture and loss function on all these tasks, with no task-specific customization.
DoorDash dives into delicious drone deliveries
And in so many ways, it kinda sucks. A new graphics card costs more than a mortgage payment because billionaires are sucking up all the GPUs to boil the planet and make Hayao Miyazaki cry at the same time, and I still don't have a Marty McFly hoverboard. But at least I can order fast food that literally flies to my door. In fact, I could order a flying curry delivery if I lived in Charlotte, North Carolina--specifically, within four miles of the Arboretum Shopping Center--where DoorDash is now offering food deliveries via drone. You can choose from a limited selection of local eateries, including Panera Bread, Matcha Cafe Maiko, and Joa Korean.
Infinite Recommendation Networks: A Data-Centric Approach
We leverage the Neural Tangent Kernel and its equivalence to training infinitely-wide neural networks to devise \infty -AE: an autoencoder with infinitely-wide bottleneck layers. The outcome is a highly expressive yet simplistic recommendation model with a single hyper-parameter and a closed-form solution. Leveraging \infty -AE's simplicity, we also develop Distill-CF for synthesizing tiny, high-fidelity data summaries which distill the most important knowledge from the extremely large and sparse user-item interaction matrix for efficient and accurate subsequent data-usage like model training, inference, architecture search, etc. This takes a data-centric approach to recommendation, where we aim to improve the quality of logged user-feedback data for subsequent modeling, independent of the learning algorithm. We particularly utilize the concept of differentiable Gumbel-sampling to handle the inherent data heterogeneity, sparsity, and semi-structuredness, while being scalable to datasets with hundreds of millions of user-item interactions.
What to Know About the Apple Class Action Lawsuit Settlement--and How You Can File a Claim
Apple users--specifically those who use Siri through products such as Macbooks, iPhones, and Apple TVs--may be entitled to make a claim after Apple's class action lawsuit settlement, worth 95 million dollars, regarding the voice-activated assistant. The settlement comes from a lawsuit filed in 2021 by Californian Fumiko Lopez, who claimed that Apple, via Siri, conducted "unlawful and intentional interception and recording of individuals' confidential communications without their consent and subsequent unauthorized disclosure of those communications." "Apple intentionally, willfully, and knowingly violated consumers' privacy rights, including within the sanctity of consumers' own homes where they have the greatest expectation of privacy," the lawsuit stated. "Plaintiffs and Class Members would not have bought their Siri Devices, or would have paid less for them, if they had known Apple was intercepting, recording, disclosing, and otherwise misusing their conversations without consent or authorization." In 2019, Apple published a statement titled "Improving Siri's privacy protections," in which they said they hadn't "been fully living up" to their "high ideals" and vowed to issue improvements.
Separations in the Representational Capabilities of Transformers and Recurrent Architectures
Transformer architectures have been widely adopted in foundation models. Due to their high inference costs, there is renewed interest in exploring the potential of efficient recurrent architectures (RNNs). In this paper, we analyze the differences in the representational capabilities of Transformers and RNNs across several tasks of practical relevance, including index lookup, nearest neighbor, recognizing bounded Dyck languages, and string equality. For the tasks considered, our results show separations based on the size of the model required for different architectures. For example, we show that a one-layer Transformer of logarithmic width can perform index lookup, whereas an RNN requires a hidden state of linear size.
Major blow to Elon Musk as billionaire could be forced to cancel long-awaited dream
Elon Musk's Tesla plans to roll-out self-driving'robotaxis' in just a few weeks, but auto safety officials may force the billionaire to cancel his long-awaited dream Tesla was set to launch the service next month in Austin, Texas, unleashing taxis powered by its Full Self-Driving (FSD) program. The National Highway Traffic Safety Administration (NHTSA) recently caught wind of Musk's upcoming rollout and sent the company a letter to gather additional information. The NHTSA wants to ' understand how Tesla plans to evaluate its vehicles and driving automation technologies for use on public roads' before the robotaxis are unleashed on busy Austin streets. The agency highlighted its investigations into four crashes and a pedestrian linked to Tesla's FDS. The blow has led to Musk's critics suggesting he will have to put a pin in his plans.
I test a lot of AI coding tools, and this stunning new OpenAI release just saved me days of work
Last week, OpenAI quietly dropped a programming bombshell post on X/Twitter. It turns out you can now connect GitHub repos to Deep Research in ChatGPT. What makes this particularly interesting is that you can put ChatGPT Deep Research to work scanning that repo for all sorts of yummy nuggets of information. GitHub is an online resource owned by Microsoft that holds an enormous number of programming projects, both open source and proprietary. It's used by teams to coordinate and track development.
Google DeepMind's AI Agent Dreams Up Algorithms Beyond Human Expertise
A key question in artificial intelligence is how often models go beyond just regurgitating and remixing what they have learned and produce truly novel ideas or insights. A new project from Google DeepMind shows that with a few clever tweaks these models can at least surpass human expertise designing certain types of algorithms--including ones that are useful for advancing AI itself. The company's latest AI project, called AlphaEvolve, combines the coding skills of its Gemini AI model with a method for testing the effectiveness of new algorithms and an evolutionary method for producing new designs. AlphaEvolve came up with more efficient algorithms for several kinds of computation, including a method for calculations involving matrices that betters an approach called the Strassen algorithm that has been relied upon for 56 years. The new approach improves the computational efficiency by reducing the number of calculations required to produce a result.