heft
HEFT: A Coarse-to-Fine Hierarchy for Enhancing the Efficiency and Accuracy of Language Model Reasoning
The adaptation of large language models (LLMs) to specialized reasoning tasks is fundamentally constrained by computational resources. Parameter-Efficient Fine-Tuning (PEFT) methods have emerged as a powerful solution, yet the landscape of these techniques is diverse, with distinct methods operating in either the model's weight space or its representation space. This paper investigates the hypothesis that a synergistic combination of these paradigms can unlock superior performance and efficiency. We introduce HEFT (Hierarchical Efficient Fine-Tuning), a novel hierarchical adaptation strategy that composes two distinct PEFT methods in a coarse-to-fine manner: first, a broad, foundational adaptation in the weight space using Low-Rank Adaptation (LoRA), followed by a precise, surgical refinement of internal activations using Representation Fine-Tuning (ReFT). We evaluate this approach by fine-tuning a Llama-2-7B model on the BoolQ benchmark, a challenging dataset for inferential reasoning. Our results reveal a profound synergistic effect. A model fine-tuned for only three epochs with our HEFT strategy achieves an accuracy of 85.17\%, exceeding the performance of models trained for 20 epochs with either LoRA-only (85.05\%) or ReFT-only (83.36\%) methodologies. This work demonstrates that the thoughtful composition of PEFT methods is a potent algorithmic innovation, offering a more efficient and effective path toward advancing the reasoning capabilities of language models. By achieving superior results with a fraction of the computational budget, our findings present a principled approach to overcoming the obstacles inherent in adapting large-scale models for complex cognitive tasks.
- North America > United States > Wisconsin > Dane County > Madison (0.14)
- Europe > Romania > Sud - Muntenia Development Region > Giurgiu County > Giurgiu (0.04)
- Asia > China > Ningxia Hui Autonomous Region > Yinchuan (0.04)
DeepSoCS: A Neural Scheduler for Heterogeneous System-on-Chip (SoC) Resource Scheduling
Sung, Tegg Taekyong, Ha, Jeongsoo, Kim, Jeewoo, Yahja, Alex, Sohn, Chae-Bong, Ryu, Bo
In this paper, we~present a novel scheduling solution for a class of System-on-Chip (SoC) systems where heterogeneous chip resources (DSP, FPGA, GPU, etc.) must be efficiently scheduled for continuously arriving hierarchical jobs with their tasks represented by a directed acyclic graph. Traditionally, heuristic algorithms have been widely used for many resource scheduling domains, and Heterogeneous Earliest Finish Time (HEFT) has been a dominating state-of-the-art technique across a broad range of heterogeneous resource scheduling domains over many years. Despite their long-standing popularity, HEFT-like algorithms are known to be vulnerable to a small amount of noise added to the environment. Our Deep Reinforcement Learning (DRL)-based SoC Scheduler (DeepSoCS), capable of learning the "best" task ordering under dynamic environment changes, overcomes the brittleness of rule-based schedulers such as HEFT with significantly higher performance across different types of jobs. We~describe a DeepSoCS design process using a real-time heterogeneous SoC scheduling emulator, discuss major challenges, and present two novel neural network design features that lead to outperforming HEFT: (i) hierarchical job- and task-graph embedding; and (ii) efficient use of real-time task information in the state space. Furthermore, we~introduce effective techniques to address two fundamental challenges present in our environment: delayed consequences and joint actions. Through an extensive simulation study, we~show that our DeepSoCS exhibits the significantly higher performance of job execution time than that of HEFT with a higher level of robustness under realistic noise conditions. We~conclude with a discussion of the potential improvements for our DeepSoCS neural scheduler.
- North America > United States (1.00)
- Asia (0.93)
- Europe (0.68)
- Semiconductors & Electronics (0.85)
- Energy > Oil & Gas > Upstream (0.35)
There's No Homunculus In Our Brain Who Guides Us - Issue 81: Maps
In the early 1980s, the psychologist Harry Heft put a 16 mm camera in the back of a sports car and made a movie. It consisted of a continuous shot of a residential neighborhood in Granville, Ohio, where Heft was a professor at Denison University. It didn't have a plot or actors, but it did have a simple narrative: The car started moving at 5 miles per hour and made nine turns from one street to another and then came to a stop after traveling just under a mile. One showed just the vistas along the route, the expansive layout of environmental features, such as a group of houses or trees seen from a distance. The second film showed the transitions of the route, the parts between each vista where the view is occluded by, say, a turn in the road or the crest of a hill.
- North America > United States > Ohio (0.24)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.14)
- North America > United States > New York (0.04)
- (4 more...)
- Health & Medicine > Therapeutic Area > Neurology (1.00)
- Media (0.91)
Amazon's Kindle Oasis Puts Function Over Form
Where many companies are obsessed with making the latest, greatest smartphone, Amazon has found success with other modern devices. Along with the Amazon Echo and Echo Dot line of personal assistants, the former bookseller shines when it comes to giving you a modern way to read those books. The Kindle line started in 2007, but here in 2016, the line of e-readers serves the same function: offering a convenient way to consume your ever-growing book collection. The latest iteration of the e-reader, Amazon's Kindle Oasis, continues to pursue that mission--with longer battery life, a squatter form-factor and a very high price tag. The first thing you'll notice about the Kindle Oasis is its asymmetrical shape.