Goto

Collaborating Authors

 pecan


MinatoLoader: Accelerating Machine Learning Training Through Efficient Data Preprocessing

Nouaji, Rahma, Bitchebe, Stella, Macedo, Ricardo, Balmau, Oana

arXiv.org Artificial Intelligence

Data loaders are used by Machine Learning (ML) frameworks like PyTorch and TensorFlow to apply transformations to data before feeding it into the accelerator. This operation is called data preprocessing. Data preprocessing plays an important role in the ML training workflow because if it is inefficiently pipelined with the training, it can yield high GPU idleness, resulting in important training delays. Unfortunately, existing data loaders turn out to waste GPU resources, with $76\%$ GPU idleness when using the PyTorch data loader, for example. One key source of inefficiency is the variability in preprocessing time across samples within the same dataset. Existing data loaders are oblivious to this variability, and they construct batches without any consideration of slow or fast samples. In this case, the entire batch is delayed by a single slow sample, stalling the training pipeline and resulting in head-of-line blocking. To address these inefficiencies, we present MinatoLoader, a general-purpose data loader for PyTorch that accelerates training and improves GPU utilization. MinatoLoader is designed for a single-server setup, containing multiple GPUs. It continuously prepares data in the background and actively constructs batches by prioritizing fast-to-preprocess samples, while slower samples are processed in parallel. We evaluate MinatoLoader on servers with V100 and A100 GPUs. On a machine with four A100 GPUs, MinatoLoader improves the training time of a wide range of workloads by up to $7.5\times$ ($3.6\times$ on average) over PyTorch DataLoader and Pecan, and up to $3\times$ ($2.2\times$ on average) over DALI. It also increases average GPU utilization from 46.4\% with PyTorch to 90.45\%, while preserving model accuracy and enabling faster convergence.


PECAN: Personalizing Robot Behaviors through a Learned Canonical Space

Nemlekar, Heramb, Sanchez, Robert Ramirez, Losey, Dylan P.

arXiv.org Artificial Intelligence

Robots should personalize how they perform tasks to match the needs of individual human users. Today's robot achieve this personalization by asking for the human's feedback in the task space. For example, an autonomous car might show the human two different ways to decelerate at stoplights, and ask the human which of these motions they prefer. This current approach to personalization is indirect: based on the behaviors the human selects (e.g., decelerating slowly), the robot tries to infer their underlying preference (e.g., defensive driving). By contrast, our paper develops a learning and interface-based approach that enables humans to directly indicate their desired style. We do this by learning an abstract, low-dimensional, and continuous canonical space from human demonstration data. Each point in the canonical space corresponds to a different style (e.g., defensive or aggressive driving), and users can directly personalize the robot's behavior by simply clicking on a point. Given the human's selection, the robot then decodes this canonical style across each task in the dataset -- e.g., if the human selects a defensive style, the autonomous car personalizes its behavior to drive defensively when decelerating, passing other cars, or merging onto highways. We refer to our resulting approach as PECAN: Personalizing Robot Behaviors through a Learned Canonical Space. Our simulations and user studies suggest that humans prefer using PECAN to directly personalize robot behavior (particularly when those users become familiar with PECAN), and that users find the learned canonical space to be intuitive and consistent. See videos here: https://youtu.be/wRJpyr23PKI


One-Shot Safety Alignment for Large Language Models via Optimal Dualization

Huang, Xinmeng, Li, Shuo, Dobriban, Edgar, Bastani, Osbert, Hassani, Hamed, Ding, Dongsheng

arXiv.org Machine Learning

The growing safety concerns surrounding Large Language Models (LLMs) raise an urgent need to align them with diverse human preferences to simultaneously enhance their helpfulness and safety. A promising approach is to enforce safety constraints through Reinforcement Learning from Human Feedback (RLHF). For such constrained RLHF, common Lagrangian-based primal-dual policy optimization methods are computationally expensive and often unstable. This paper presents a dualization perspective that reduces constrained alignment to an equivalent unconstrained alignment problem. We do so by pre-optimizing a smooth and convex dual function that has a closed form. This shortcut eliminates the need for cumbersome primal-dual policy iterations, thus greatly reducing the computational burden and improving training stability. Our strategy leads to two practical algorithms in model-based and preference-based scenarios (MoCAN and PeCAN, respectively). A broad range of experiments demonstrate the effectiveness of our methods.


PECAN: A Deterministic Certified Defense Against Backdoor Attacks

Zhang, Yuhao, Albarghouthi, Aws, D'Antoni, Loris

arXiv.org Artificial Intelligence

Neural networks are vulnerable to backdoor poisoning attacks, where the attackers maliciously poison the training set and insert triggers into the test input to change the prediction of the victim model. Existing defenses for backdoor attacks either provide no formal guarantees or come with expensive-to-compute and ineffective probabilistic guarantees. We present PECAN, an efficient and certified approach for defending against backdoor attacks. The key insight powering PECAN is to apply off-the-shelf test-time evasion certification techniques on a set of neural networks trained on disjoint partitions of the data. We evaluate PECAN on image classification and malware detection datasets. Our results demonstrate that PECAN can (1) significantly outperform the state-of-the-art certified backdoor defense, both in defense strength and efficiency, and (2) on real back-door attacks, PECAN can reduce attack success rate by order of magnitude when compared to a range of baselines from the literature.


PECAN: Leveraging Policy Ensemble for Context-Aware Zero-Shot Human-AI Coordination

Lou, Xingzhou, Guo, Jiaxian, Zhang, Junge, Wang, Jun, Huang, Kaiqi, Du, Yali

arXiv.org Artificial Intelligence

Zero-shot human-AI coordination holds the promise of collaborating with humans without human data. Prevailing methods try to train the ego agent with a population of partners via self-play. However, these methods suffer from two problems: 1) The diversity of a population with finite partners is limited, thereby limiting the capacity of the trained ego agent to collaborate with a novel human; 2) Current methods only provide a common best response for every partner in the population, which may result in poor zero-shot coordination performance with a novel partner or humans. To address these issues, we first propose the policy ensemble method to increase the diversity of partners in the population, and then develop a context-aware method enabling the ego agent to analyze and identify the partner's potential policy primitives so that it can take different actions accordingly. In this way, the ego agent is able to learn more universal cooperative behaviors for collaborating with diverse partners. We conduct experiments on the Overcooked environment, and evaluate the zero-shot human-AI coordination performance of our method with both behavior-cloned human proxies and real humans. The results demonstrate that our method significantly increases the diversity of partners and enables ego agents to learn more diverse behaviors than baselines, thus achieving state-of-the-art performance in all scenarios. We also open-source a human-AI coordination study framework on the Overcooked for the convenience of future studies.


How the economic downturn is affecting the AI sector

#artificialintelligence

The impact of the economic slowdown for vendors depends on the type of vendor. Some large technology companies, including Google, have frozen nonessential hiring. However, some vendors have yet to feel the impact of the economic slowdown. Pecan AI, a vendor that provides a deep learning platform designed to build predictive models for enterprises, has raised about $100 million in the past year. The Israel-based company raised $66 million in its last funding round in February.


Thought Leaders in Artificial Intelligence: Zohar Bronfman, CEO of Pecan.ai (Part 1)

#artificialintelligence

This is a terrific PaaS company in the making with substantial predictive capabilities. Sramana Mitra: Let's start by introducing our audience to yourself as well as Pecan.ai. Prior to founding Pecan along with my co-founder Noam, I spent most of my days in the academia. I did two PhDs in parallel in supplementary fields. I did one in the field of computational neuroscience, and the other in the field of philosophy.


Pecan.ai launches with $11M Series A to automate machine learning – TechCrunch

#artificialintelligence

Pecan.ai, a startup that wants to help business analysts build machine learning models in an automated fashion, emerged from stealth today and announced an $11 million Series A. The round was led by Dell Technologies Capital and S Capital. Along with a previously unannounced $4 million seed round, the company has raised a total of $15 million. CEO Zohar Bronfman says he and co-founder Noam Brezis, whom he has known for more than a decade, started the company with the goal of building an automated machine learning platform. They observed that much of the work involved in building machine learning models is about getting data in a form that the algorithm can consume, something they've automated in Pecan. "The innovative thing about Pecan is that we do all of the data preparation and data, engineering and data processing, and [complete the] various technical steps [for you]," Bronfman explained.


Siri and Alexa walk into a bar: How AI assistants found their funny bone

Engadget

Virtual assistants like Siri and Alexa are useful for a lot of things, like telling you what the weather's going to be or reminding you of an upcoming calendar appointment. But they can be entertaining too, providing the occasional fun fact or playing that hit song from Beyonce. Or, when you want a little levity in an otherwise crappy day, telling some really corny dad jokes. You've probably heard them before. Ask Siri, Alexa, Google or Cortana to tell you a joke, and you'll likely hear something like this: "I couldn't figure out why a baseball kept getting larger. "What do you call a talking dinosaur?