new instruction
Foundations of Large Language Models
The development of neural sequence models, such as Transformers [Vaswani et al., 2017], along with the improvements in large-scale self-supervised learning, has opened the door to universal language understanding and generation. This achievement is largely motivated by pre-training: we separate common components from many neural network-based systems, and then train them on huge amounts of unlabeled data using self-supervision. These pre-trained models serve as foundation models that can be easily adapted to different tasks via fine-tuning or prompting. As a result, the paradigm of NLP has been enormously changed. In many cases, large-scale supervised learning for specific tasks is no longer required, and instead, we only need to adapt pre-trained foundation models.
- North America > United States > California > Los Angeles County > Los Angeles (0.04)
- Asia > China (0.04)
- South America > Argentina (0.04)
- (10 more...)
- Workflow (1.00)
- Research Report > Promising Solution (1.00)
- Overview (1.00)
- Research Report > New Finding (0.92)
- Leisure & Entertainment > Sports (1.00)
- Law (1.00)
- Information Technology > Security & Privacy (1.00)
- (6 more...)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Machine Translation (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- (10 more...)
Genetic Instruct: Scaling up Synthetic Generation of Coding Instructions for Large Language Models
Majumdar, Somshubra, Noroozi, Vahid, Narenthiran, Sean, Ficek, Aleksander, Balam, Jagadeesh, Ginsburg, Boris
Large Language Models (LLMs) rely on instruction samples for alignment, but creating these datasets poses challenges, particularly in expert-dependent tasks like coding, which can be cost-prohibitive. One approach to mitigate these challenges is synthesizing data using another LLM. In this paper, we introduce a scalable method for generating synthetic instructions to enhance the code generation capability of LLMs. The proposed algorithm, Genetic-Instruct, mimics evolutionary processes, utilizing self-instruction to create numerous synthetic samples from a limited number of seeds. Genetic-Instruct is designed for efficient scaling of the generation process. Fine-tuning multiple coding LLMs with the synthetic samples demonstrates a significant improvement in their code generation accuracy compared to the baselines.
- North America > Canada > Ontario > Toronto (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > California > Santa Clara County > Santa Clara (0.04)
FewFedPIT: Towards Privacy-preserving and Few-shot Federated Instruction Tuning
Zhang, Zhuo, Zhang, Jingyuan, Huang, Jintao, Qu, Lizhen, Zhang, Hongzhi, Wang, Qifan, Zhou, Xun, Xu, Zenglin
Instruction tuning has been identified as a crucial technique for optimizing the performance of large language models (LLMs) in generating human-aligned responses. Nonetheless, gathering diversified and superior-quality instruction data for such tuning presents notable obstacles, especially in domains with rigid privacy provisions. Federated instruction tuning (FedIT) has emerged as a promising solution, by consolidating collaborative training across multiple data owners, thereby resulting in a privacy-preserving learning model. However, FedIT encounters limitations such as scarcity of instructional data and risk of exposure to training data extraction attacks. In this paper, we propose a novel federated algorithm, FewFedPIT, designed to simultaneously enhance privacy protection and model performance of federated few-shot learning. FewFedPITcomprises three vital components on the client side: (1) synthetic data generation, which utilizes LLMs' in-context learning capacity to generate synthetic data autonomously, thus expanding the local database; (2) parameter isolation training, which individually updates the public parameters in the synthetic data and the private parameters in the local data, consequently mitigating the noise impact of the synthetic data; (3) local aggregation sharing, which mixes public and private parameters before uploading, effectively preventing data extraction attacks. Extensive experiments on three open-source datasets demonstrate the effectiveness of FewFedPITin, enhancing privacy preservation and improving federated few-shot performance.
- Asia > Myanmar > Tanintharyi Region > Dawei (0.04)
- Asia > China > Guangdong Province > Shenzhen (0.04)
- Oceania > Australia > Victoria > Melbourne (0.04)
- (3 more...)
- Information Technology > Security & Privacy (1.00)
- Health & Medicine (0.93)
- Education (0.93)
Accelerating PyTorch Transformers with Intel Sapphire Rapids - part 1
About a year ago, we showed you how to distribute the training of Hugging Face transformers on a cluster or third-generation Intel Xeon Scalable CPUs (aka Ice Lake). Recently, Intel has launched the fourth generation of Xeon CPUs, code-named Sapphire Rapids, with exciting new instructions that speed up operations commonly found in deep learning models. In this post, you will learn how to accelerate a PyTorch training job with a cluster of Sapphire Rapids servers running on AWS. We will use the Intel oneAPI Collective Communications Library (CCL) to distribute the job, and the Intel Extension for PyTorch (IPEX) library to automatically put the new CPU instructions to work. As both libraries are already integrated with the Hugging Face transformers library, we will be able to run our sample scripts out of the box without changing a line of code.
Google Develops Code-Writing AI To Help Robots Learn New Tasks - AI Summary
Google researchers believe that natural language processing and AI will enable robots to create their code to take action against new instructions. Google is currently testing a system that allows robots to write their code, follow instructions, and complete tasks. This is a way to streamline reprogramming policies for every new task. It can be tedious and time-consuming, and it requires domain experts. Staff could interact with robots on smart factory floors using simple commands without needing to write complicated code. Google researchers developed language modeling programs called Code as Policies (CaP). This code-writing AI system can generate new code for new instructions. Google robotics scientists stated in a blog post that “Given natural languages, instructions, current language models can write not only generic code, but, as we have discovered, code capable of controlling robot actions.” Google researchers have combined large language models with everyday robots to better respond to complex and abstract human requests. According to the Google team, CaP will enable a single system
Google Developed A New Robot That Can Code Itself
Google researchers believe that natural language processing and AI will enable robots to create their code to take action against new instructions. Google is currently testing a system that allows robots to write their code, follow instructions, and complete tasks. This is a way to streamline reprogramming policies for every new task. It can be tedious and time-consuming, and it requires domain experts. Staff could interact with robots on smart factory floors using simple commands without needing to write complicated code.
Robots That Write Their Own Code
A common approach used to control robots is to program them with code to detect objects, sequencing commands to move actuators, and feedback loops to specify how the robot should perform a task. While these programs can be expressive, re-programming policies for each new task can be time consuming, and requires domain expertise. What if when given instructions from people, robots could autonomously write their own code to interact with the world? It turns out that the latest generation of language models, such as PaLM, are capable of complex reasoning and have also been trained on millions of lines of code. Given natural language instructions, current language models are highly proficient at writing not only generic code but, as we've discovered, code that can control robot actions as well.
Classes and Instances · Crafting Interpreters
If you see a mistake, find something unclear, or have a suggestion, please let me know. Don't worry, I won't spam you.) Caring too much for objects can destroy you. Only--if you care for a thing enough, it takes on a life of its own, doesn't it? And isn't the whole point of things--beautiful things--that they connect you to some larger beauty? The last area left to implement in clox is object-oriented programming. OOP is a bundle of intertwined features: classes, instances, fields, methods, initializers, and inheritance. Using relatively high-level Java, we packed all that into two chapters. Now that we're coding in C, which feels like building a model of the Eiffel tower out of toothpicks, we'll devote three chapters to covering the same territory.
Arm Adds Muscle To Machine Learning, Embraces Bfloat16
Arm Holdings has announced that the next revision of its ArmV8-A architecture will include support for bfloat16, a floating point format that is increasingly being used to accelerate machine learning applications. It joins Google, Intel, and a handful of startups, all of whom are etching bfloat16 into their respective silicon. Bfloat16, aka 16-bit "brain floating point, was invented by Google and first implemented in its third-generation Tensor Processing Unit (TPU). Intel thought highly enough of the format to incorporate bfloat16 in its future "Cooper Lake" Xeon SP processors, as well in its upcoming "Spring Crest" neural network processors. Wave Computing, Habana Labs, and Flex Logix followed suit with their custom AI processors. The main idea of bfloat16 to provide a 16-bit floating point format that has the same dynamic range as a standard IEEE-FP32, but with less accuracy.
AVX-512 Vector Neural Network Instructions (VNNI) - x86 - WikiChip
AVX-512 Vector Neural Network Instructions (AVX512 VNNI) is an x86 extension, part of the AVX-512, designed to accelerate convolutional neural network-based algorithms. The AVX512 VNNI x86 extension extends AVX-512 Foundation by introducing four new instructions for accelerating inner convolutional neural network loops. The major motivation behind the AVX512 VNNI extension is the observation that many tight convolutional neural network loops require the repeated multiplication of two 16-bit values or two 8-bit values and accumulate the result to a 32-bit accumulator. Using the foundation AVX-512, for 16-bit, this is possible using two instructions - VPMADDWD which is used to multiply two 16-bit pairs and add them together followed a VPADDD which adds the accumulate value. Likewise, for 8-bit values, three instructions are needed - VPMADDUBSW which is used to multiply two 8-bit pairs and add them together, followed by a VPMADDWD with the value 1 in order to simply up-convert the 16-bit values to 32-bit values, followed by the VPADDD instruction which adds the result to an accumulator.