Goto

Collaborating Authors

 ferret


Cloning isn't just for celebrity pets like Tom Brady's dog

MIT Technology Review

Yes, you can pay $50,000 to clone a pet. But others are using the technology to rescue endangered species. This week, we heard that Tom Brady had his dog cloned. The former quarterback revealed that his Junie is actually a clone of Lua, a pit bull mix that died in 2023. Brady's announcement follows those of celebrities like Paris Hilton and Barbra Streisand, who also famously cloned their pet dogs. But some believe there are better ways to make use of cloning technologies.


Ferret: An Efficient Online Continual Learning Framework under Varying Memory Constraints

Zhou, Yuhao, Tian, Yuxin, Lv, Jindi, Shi, Mingjia, Li, Yuanxi, Ye, Qing, Zhang, Shuhao, Lv, Jiancheng

arXiv.org Artificial Intelligence

In the realm of high-frequency data streams, achieving real-time learning within varying memory constraints is paramount. This paper presents Ferret, a comprehensive framework designed to enhance online accuracy of Online Continual Learning (OCL) algorithms while dynamically adapting to varying memory budgets. Ferret employs a fine-grained pipeline parallelism strategy combined with an iterative gradient compensation algorithm, ensuring seamless handling of high-frequency data with minimal latency, and effectively counteracting the challenge of stale gradients in parallel training. To adapt to varying memory budgets, its automated model partitioning and pipeline planning optimizes performance regardless of memory limitations. Extensive experiments across 20 benchmarks and 5 integrated OCL algorithms show Ferret's remarkable efficiency, achieving up to 3.7$\times$ lower memory overhead to reach the same online accuracy compared to competing methods. Furthermore, Ferret consistently outperforms these methods across diverse memory budgets, underscoring its superior adaptability. These findings position Ferret as a premier solution for efficient and adaptive OCL framework in real-time environments.


Ferret: Federated Full-Parameter Tuning at Scale for Large Language Models

Shu, Yao, Hu, Wenyang, Ng, See-Kiong, Low, Bryan Kian Hsiang, Yu, Fei Richard

arXiv.org Artificial Intelligence

Large Language Models (LLMs) have become indispensable in numerous real-world applications. Unfortunately, fine-tuning these models at scale, especially in federated settings where data privacy and communication efficiency are critical, presents significant challenges. Existing methods often resort to parameter-efficient fine-tuning (PEFT) to mitigate communication overhead, but this typically comes at the cost of model accuracy. To address these limitations, we propose federated full-parameter tuning at scale for LLMs (Ferret), the first first-order method with shared randomness to enable scalable full-parameter tuning of LLMs across decentralized data sources while maintaining competitive model accuracy. Ferret accomplishes this through three aspects: (1) it employs widely applied first-order methods for efficient local updates; (2) it projects these updates into a low-dimensional space to considerably reduce communication overhead; and (3) it reconstructs local updates from this low-dimensional space with shared randomness to facilitate effective full-parameter global aggregation, ensuring fast convergence and competitive final performance. Our rigorous theoretical analyses and insights along with extensive experiments, show that Ferret significantly enhances the scalability of existing federated full-parameter tuning approaches by achieving high computational efficiency, reduced communication overhead, and fast convergence, all while maintaining competitive model accuracy. Our implementation is available at https://github.com/allen4747/Ferret.


A Survey of Temporal Credit Assignment in Deep Reinforcement Learning

Pignatelli, Eduardo, Ferret, Johan, Geist, Matthieu, Mesnard, Thomas, van Hasselt, Hado, Toni, Laura

arXiv.org Artificial Intelligence

The Credit Assignment Problem (CAP) refers to the longstanding challenge of Reinforcement Learning (RL) agents to associate actions with their long-term consequences. Solving the CAP is a crucial step towards the successful deployment of RL in the real world since most decision problems provide feedback that is noisy, delayed, and with little or no information about the causes. These conditions make it hard to distinguish serendipitous outcomes from those caused by informed decision-making. However, the mathematical nature of credit and the CAP remains poorly understood and defined. In this survey, we review the state of the art of Temporal Credit Assignment (CA) in deep RL. We propose a unifying formalism for credit that enables equitable comparisons of state of the art algorithms and improves our understanding of the trade-offs between the various methods. We cast the CAP as the problem of learning the influence of an action over an outcome from a finite amount of experience. We discuss the challenges posed by delayed effects, transpositions, and a lack of action influence, and analyse how existing methods aim to address them. Finally, we survey the protocols to evaluate a credit assignment method, and suggest ways to diagnoses the sources of struggle for different credit assignment methods. Overall, this survey provides an overview of the field for new-entry practitioners and researchers, it offers a coherent perspective for scholars looking to expedite the starting stages of a new study on the CAP, and it suggests potential directions for future research


Ferret: Refer and Ground Anything Anywhere at Any Granularity

You, Haoxuan, Zhang, Haotian, Gan, Zhe, Du, Xianzhi, Zhang, Bowen, Wang, Zirui, Cao, Liangliang, Chang, Shih-Fu, Yang, Yinfei

arXiv.org Artificial Intelligence

We introduce Ferret, a new Multimodal Large Language Model (MLLM) capable of understanding spatial referring of any shape or granularity within an image and accurately grounding open-vocabulary descriptions. To unify referring and grounding in the LLM paradigm, Ferret employs a novel and powerful hybrid region representation that integrates discrete coordinates and continuous features jointly to represent a region in the image. To extract the continuous features of versatile regions, we propose a spatial-aware visual sampler, adept at handling varying sparsity across different shapes. Consequently, Ferret can accept diverse region inputs, such as points, bounding boxes, and free-form shapes. To bolster the desired capability of Ferret, we curate GRIT, a comprehensive refer-and-ground instruction tuning dataset including 1.1M samples that contain rich hierarchical spatial knowledge, with 95K hard negative data to promote model robustness. The resulting model not only achieves superior performance in classical referring and grounding tasks, but also greatly outperforms existing MLLMs in region-based and localization-demanded multimodal chatting. Our evaluations also reveal a significantly improved capability of describing image details and a remarkable alleviation in object hallucination. Code and data will be available at https://github.com/apple/ml-ferret


ferret: a Framework for Benchmarking Explainers on Transformers

Attanasio, Giuseppe, Pastor, Eliana, Di Bonaventura, Chiara, Nozza, Debora

arXiv.org Artificial Intelligence

As Transformers are increasingly relied upon to solve complex NLP problems, there is an increased need for their decisions to be humanly interpretable. While several explainable AI (XAI) techniques for interpreting the outputs of transformer-based models have been proposed, there is still a lack of easy access to using and comparing them. We introduce ferret, a Python library to simplify the use and comparisons of XAI methods on transformer-based classifiers. With ferret, users can visualize and compare transformers-based models output explanations using state-of-the-art XAI methods on any free-text or existing XAI corpora. Moreover, users can also evaluate ad-hoc XAI metrics to select the most faithful and plausible explanations. To align with the recently consolidated process of sharing and using transformers-based models from Hugging Face, ferret interfaces directly with its Python library. In this paper, we showcase ferret to benchmark XAI methods used on transformers for sentiment analysis and hate speech detection. We show how specific methods provide consistently better explanations and are preferable in the context of transformer models.


Exclusive Interview with Rob Loughan, Founder and CEO of Ferret

#artificialintelligence

The good news is that we have the AI technology to uncover the rampant investor fraud that has gotten exponentially worse over the last two years because of the convergence of several factors: the pandemic (which made remote meetings the norm), the rise of cryptocurrency (which by definition has no institutional safeguards) and easy Wall Street money (which means deals are being done in a matter of days without the due diligence that would usually protect the investor). While the technology is there to fix the problem, it's not being adopted quickly enough by investors.


Rapidly evolving bits of DNA helped develop the human brain

New Scientist

Many of the fastest-evolving sections of the human genome are involved in brain development. These rapidly changing segments of DNA may have played key roles in the evolution of the human brain and in our cognitive abilities. Chris Walsh at Boston Children's Hospital in Massachusetts and his colleagues studied sections of the human genome dubbed "human accelerated regions" (HARs). These stretches of DNA are virtually identical in many other mammals that have been studied, suggesting they have important functions – but they differ in humans, implying our evolution has changed them. Previous studies have identified 3171 possible HARs, but Walsh says it is unlikely that they are all important.


Commonsense Knowledge Mining from Pretrained Models

Feldman, Joshua, Davison, Joe, Rush, Alexander M.

arXiv.org Artificial Intelligence

Inferring commonsense knowledge is a key challenge in natural language processing, but due to the sparsity of training data, previous work has shown that supervised methods for commonsense knowledge mining underperform when evaluated on novel data. In this work, we develop a method for generating commonsense knowledge using a large, pre-trained bidirectional language model. By transforming relational triples into masked sentences, we can use this model to rank a triple's validity by the estimated pointwise mutual information between the two entities. Since we do not update the weights of the bidirectional model, our approach is not biased by the coverage of any one commonsense knowledge base. Though this method performs worse on a test set than models explicitly trained on a corresponding training set, it outperforms these methods when mining commonsense knowledge from new sources, suggesting that unsupervised techniques may generalize better than current supervised approaches.


Book review: The Master Algorithm by Pedro Domingos

#artificialintelligence

I first came across this book when I was reading analysts' review of President Xi Jinping's New Years' address during the turn of the year and this book was apparently one of the two books on AI and robotics that was on the Chinese President's bookshelf. Piqued by this revelation, I then subsequently learnt that this book was also on Bill Gates' recommended reading list. The book's full title, "The Master Algorithm – How the Quest for the Ultimate Learning Machine will Remake our World," provided the necessary hyperbole that helped me make my decision to read it. Whilst I had some rudimentary of what algorithms do, how AI will impact the world we live in, and how machine learning is being used across various industries from healthcare, to education to security. At the heart of machine learning is the ability of learners to use algorithms to collate data, create meaningful and actionable insights from the data and determine or execute next steps or tasks.