new approach
Extropic Aims to Disrupt the Data Center Bonanza
A startup hopes to challenge Nvidia, AMD, and Intel with a chip that wrangles probabilities rather than ones and zeros. Extropic claims its exotic new chip, called XTR-0, could be thousands of times more energy efficient than existing chips when scaled up. Extropic, a startup developing an exotic new kind of computer chip that handles probabilistic bits, has produced its first working hardware along with proof that more advanced systems will tackle useful tasks in artificial intelligence and scientific research. The startup's chips work in a fundamentally different way to chips from Nvidia, AMD, and others, and promise to be thousands of times more energy efficient when scaled up. With AI companies pouring billions of dollars into building datacenters, a completely new approach could offer a far less costly alternative to vast arrays of conventional chips.
- North America > United States > New York (0.05)
- North America > United States > New Mexico (0.05)
- North America > United States > California (0.05)
- (4 more...)
Inside the making of a world-class corn maze
In Indiana, Exploration Acres found a way to keep the family farm alive. Exploration Acres has operated its award-winning corn maze for almost 20 years. Breakthroughs, discoveries, and DIY tips sent every weekday. The adage refers to a farmer's goal for their crops if they hope to make the October harvest. And while most Midwesterners are familiar with the axiom, Tim Fitzgerald knows the folksy refrain lost its relevancy decades ago.
- North America > United States > Indiana > Tippecanoe County > Lafayette (0.05)
- North America > United States > Idaho (0.05)
- Food & Agriculture > Agriculture (0.48)
- Energy (0.48)
proposes a new approach (R1), and the idea of error-correction mechanism is intuitive (R1), novel (R2) and smart
Is any special feature operation applied in ETN? & Does a larger K help? The motivation to compute affinity matrices & How to achieve the error diffusion. Please see Figure 1 in submission for example. Performance issues, including increased training burden and running time. Thanks for pointing out the mistake in real-time stylization, which will be corrected in revision.
proposes a new approach (R1), and the idea of error-correction mechanism is intuitive (R1), novel (R2) and smart
Is any special feature operation applied in ETN? & Does a larger K help? The motivation to compute affinity matrices & How to achieve the error diffusion. Please see Figure 1 in submission for example. Performance issues, including increased training burden and running time. Thanks for pointing out the mistake in real-time stylization, which will be corrected in revision.
e3: Learning to Explore Enables Extrapolation of Test-Time Compute for LLMs
Setlur, Amrith, Yang, Matthew Y. R., Snell, Charlie, Greer, Jeremy, Wu, Ian, Smith, Virginia, Simchowitz, Max, Kumar, Aviral
Test-time scaling offers a promising path to improve LLM reasoning by utilizing more compute at inference time; however, the true promise of this paradigm lies in extrapolation (i.e., improvement in performance on hard problems as LLMs keep "thinking" for longer, beyond the maximum token budget they were trained on). Surprisingly, we find that most existing reasoning models do not extrapolate well. We show that one way to enable extrapolation is by training the LLM to perform in-context exploration: training the LLM to effectively spend its test time budget by chaining operations (such as generation, verification, refinement, etc.), or testing multiple hypotheses before it commits to an answer. To enable in-context exploration, we identify three key ingredients as part of our recipe e3: (1) chaining skills that the base LLM has asymmetric competence in, e.g., chaining verification (easy) with generation (hard), as a way to implement in-context search; (2) leveraging "negative" gradients from incorrect traces to amplify exploration during RL, resulting in longer search traces that chains additional asymmetries; and (3) coupling task difficulty with training token budget during training via a specifically-designed curriculum to structure in-context exploration. Our recipe e3 produces the best known 1.7B model according to AIME'25 and HMMT'25 scores, and extrapolates to 2x the training token budget. Our e3-1.7B model not only attains high pass@1 scores, but also improves pass@k over the base model.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > Virginia (0.04)
Review for NeurIPS paper: Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning
Weaknesses: My main concern with the submission is that the evaluation scenario OSAKA seems too specific and designed primarily for a set of algorithms in between Meta- & Continual-Learning while failing to make a broader argument for other approaches to Continual Learning. While certain aspects of OSAKA are certainly desirable (OOD tasks, Unknown task changes, Online Evaluation) there is a strong assumption made in allowing for Pre-training that may not be adequate in certain CL settings, limiting the generality of OSAKA. Furthermore, it is unclear how aspects such as controllable non-stationarity would be implemented in Reinforcement Learning. Furthermore, I personally feel that if task-revisiting is to be implemented, new OOD tasks should be designed in a way that explicitly re-uses skills that can be learned on a previous problem in a novel setting, instead of merely re-visiting the problem without modification. The problem with this assumption in general is that Catastrophic Forgetting is significantly reduced through an implicit form of replay provided by the environment, making it difficult to tell to which extent catastrophic forgetting is actually a problem of these algorithms.
Breaking Down the Hierarchy: A New Approach to Leukemia Classification
Hamdi, Ibraheem, El-Gendy, Hosam, Sharshar, Ahmed, Saeed, Mohamed, Ridzuan, Muhammad, Hashmi, Shahrukh K., Syed, Naveed, Mirza, Imran, Hussain, Shakir, Abdalla, Amira Mahmoud, Yaqub, Mohammad
The complexities inherent to leukemia, multifaceted cancer affecting white blood cells, pose considerable diagnostic and treatment challenges, primarily due to reliance on laborious morphological analyses and expert judgment that are susceptible to errors. Addressing these challenges, this study presents a refined, comprehensive strategy leveraging advanced deep-learning techniques for the classification of leukemia subtypes. We commence by developing a hierarchical label taxonomy, paving the way for differentiating between various subtypes of leukemia. The research further introduces a novel hierarchical approach inspired by clinical procedures capable of accurately classifying diverse types of leukemia alongside reactive and healthy cells. An integral part of this study involves a meticulous examination of the performance of Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) as classifiers. The proposed method exhibits an impressive success rate, achieving approximately 90\% accuracy across all leukemia subtypes, as substantiated by our experimental results. A visual representation of the experimental findings is provided to enhance the model's explainability and aid in understanding the classification process.
- Asia > Middle East > UAE > Abu Dhabi Emirate > Abu Dhabi (0.15)
- Europe > Middle East (0.04)
- Africa > Middle East (0.04)
- Health & Medicine > Therapeutic Area > Oncology > Leukemia (1.00)
- Health & Medicine > Therapeutic Area > Hematology (1.00)
Review for NeurIPS paper: Bootstrap Your Own Latent - A New Approach to Self-Supervised Learning
Weaknesses: -As mentioned in the paper, the proposed method has a trivial solution, that both models output 0's. To me, the method is too simple to be true. I tried to reimplement it, but no success. It is highly recommend to opensource the code for reproduceable research. How can you learn detection with frozen representation? Please use the standard settings, e.g. as in MoCo.
Review for NeurIPS paper: Domain Adaptation with Conditional Distribution Matching and Generalized Label Shift
This paper proposes a new approach to unsupervised domain adaptation (UDA) under label shift. The idea is a generalized label shift (GLS) assumption where conditional invariance is placed in representation rather than input space. The main contributions include 1) generalizing the information-theoretic lower bound of error to multiple classes; 2) devising generalization bounds in the target domain based on the balanced error rate and conditional error gap; 3) deriving necessary and sufficient conditions for GLS; 4) efficient importance reweighting algorithm for target/source label distributions using the integral probability metric. Overall, all reviewers including myself find the GLS framework interesting, providing an important new approach to UDA that can be flexibility embedded in existing methods. The theoretical foundation is also solid.