mop
Boundary Decomposition for Nadir Objective Vector Estimation
The nadir objective vector plays a key role in solving multi-objective optimization problems (MOPs), where it is often used to normalize the objective space and guide the search. The current methods for estimating the nadir objective vector perform effectively only on specific MOPs. This paper reveals the limitations of these methods: exact methods can only work on discrete MOPs, while heuristic methods cannot deal with the MOP with a complicated feasible objective region. To fill this gap, we propose a general and rigorous method, namely boundary decomposition for nadir objective vector estimation (BDNE).
Representation Calibration and Uncertainty Guidance for Class-Incremental Learning based on Vision Language Model
Tan, Jiantao, Ma, Peixian, Yu, Tong, Zhang, Wentao, Wang, Ruixuan
Abstract--Class-incremental learning requires a learning system to continually learn knowledge of new classes and meanwhile try to preserve previously learned knowledge of old classes. As current state-of-the-art methods based on Vision-Language Models (VLMs) still suffer from the issue of differentiating classes across learning tasks. Here a novel VLM-based continual learning framework for image classification is proposed. In this framework, task-specific adapters are added to the pre-trained and frozen image encoder to learn new knowledge, and a novel cross-task representation calibration strategy based on a mixture of light-weight projectors is used to help better separate all learned classes in a unified feature space, alleviating class confusion across tasks. In addition, a novel inference strategy guided by prediction uncertainty is developed to more accurately select the most appropriate image feature for class prediction. Extensive experiments on multiple datasets under various settings demonstrate the superior performance of our method compared to existing ones.
Mosaic Pruning: A Hierarchical Framework for Generalizable Pruning of Mixture-of-Experts Models
Hu, Wentao, Zhao, Mingkuan, Song, Shuangyong, Zhu, Xiaoyan, Lai, Xin, Wang, Jiayin
Sparse Mixture-of-Experts (SMoE) architectures have enabled a new frontier in scaling Large Language Models (LLMs), offering superior performance by activating only a fraction of their total parameters during inference. However, their practical deployment is severely hampered by substantial static memory overhead, as all experts must be loaded into memory. Existing post-training pruning methods, while reducing model size, often derive their pruning criteria from a single, general-purpose corpus. This leads to a critical limitation: a catastrophic performance degradation when the pruned model is applied to other domains, necessitating a costly re-pruning for each new domain. To address this generalization gap, we introduce Mosaic Pruning (MoP). The core idea of MoP is to construct a functionally comprehensive set of experts through a structured ``cluster-then-select" process. This process leverages a similarity metric that captures expert performance across different task domains to functionally cluster the experts, and subsequently selects the most representative expert from each cluster based on our proposed Activation Variability Score. Unlike methods that optimize for a single corpus, our proposed Mosaic Pruning ensures that the pruned model retains a functionally complementary set of experts, much like the tiles of a mosaic that together form a complete picture of the original model's capabilities, enabling it to handle diverse downstream tasks.Extensive experiments on various MoE models demonstrate the superiority of our approach. MoP significantly outperforms prior work, achieving a 7.24\% gain on general tasks and 8.92\% on specialized tasks like math reasoning and code generation.
Prime Day is over. If you forgot a vacuum robot, grab this one for 68% off
Prime Day came and went in a blur. It isn't until the post-haze deliveries start showing up that you realize you spent far too much on silly trinkets and toiletries instead of investment purchases. A hot dog toaster or a cute nightlight can't take chores off your plate, but this 3-in-1 robotic vacuum and mop can. And thankfully, there's still time to grab it for 68 percent off. The ECOVACS DEEBOT T20 Omni is your secret weapon for a squeaky clean home--with minimal intervention required from you.
Boundary Decomposition for Nadir Objective Vector Estimation
The nadir objective vector plays a key role in solving multi-objective optimization problems (MOPs), where it is often used to normalize the objective space and guide the search. The current methods for estimating the nadir objective vector perform effectively only on specific MOPs. This paper reveals the limitations of these methods: exact methods can only work on discrete MOPs, while heuristic methods cannot deal with the MOP with a complicated feasible objective region. To fill this gap, we propose a general and rigorous method, namely boundary decomposition for nadir objective vector estimation (BDNE). By utilizing bilevel optimization, boundary subproblems are optimized and adjusted alternately, thereby refining their optimal solutions to align with the nadir objective vector.
eufy launches the world's first robot vacuum with a portable deep cleaner (plus other powerful model)
Are you ready to completely overhaul your floor cleaning routines? The eufy E28 and E25 have just landed, and we bet you're going to want one of these robotic cleaners delivered to your home sooner rather than later. Under pre-sale now, be sure to order and save your spot in line with these powerful units. The E25 and E28 are two brand-new models unveiled by the company. Both feature eufy's award-winning HydroJet mopping technology and deliver a jaw-dropping 20,000Pa suction power for a deep clean.
On the Limits of Language Generation: Trade-Offs Between Hallucination and Mode Collapse
Kalavasis, Alkis, Mehrotra, Anay, Velegkas, Grigoris
Specifying all desirable properties of a language model is challenging, but certain requirements seem essential. Given samples from an unknown language, the trained model should produce valid strings not seen in training and be expressive enough to capture the language's full richness. Otherwise, outputting invalid strings constitutes "hallucination," and failing to capture the full range leads to "mode collapse." We ask if a language model can meet both requirements. We investigate this within a statistical language generation setting building on Gold and Angluin. Here, the model receives random samples from a distribution over an unknown language K, which belongs to a possibly infinite collection of languages. The goal is to generate unseen strings from K. We say the model generates from K with consistency and breadth if, as training size increases, its output converges to all unseen strings in K. Kleinberg and Mullainathan [KM24] asked if consistency and breadth in language generation are possible. We answer this negatively: for a large class of language models, including next-token prediction models, this is impossible for most collections of candidate languages. This contrasts with [KM24]'s result, showing consistent generation without breadth is possible for any countable collection of languages. Our finding highlights that generation with breadth fundamentally differs from generation without breadth. As a byproduct, we establish near-tight bounds on the number of samples needed for generation with or without breadth. Finally, our results offer hope: consistent generation with breadth is achievable for any countable collection of languages when negative examples (strings outside K) are available alongside positive ones. This suggests that post-training feedback, which encodes negative examples, can be crucial in reducing hallucinations while limiting mode collapse.
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- North America > United States > New York > New York County > New York City (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- (13 more...)
I switched to a robot mop: The highs, lows, and ewws
Our household only recently discovered the glorious, hands-free reality of a AI-enabled, self-emptying robot vacuum, one that scooted expertly around our apartment, busting dust for weeks on end with little input from us. But even with our trusty Roomba busting dust left and right, we knew something was missing: The bot was only sweeping, not mopping. We'd long dismissed the idea of a robot mop--after all, the earliest ones did little more than drag a damp cloth across the floor. But robot mopping tech has advanced at a furious pace, with the latest robot vacuum-and-mop combos boasting mop heads that apply downward pressure and lift themselves to avoid carpets, while base stations can clean the mop pads themselves--with hot water, no less. Could a vacuum-and-mop robot really measure up to a standard stick mop?
Roomba robot vacuums are up to 620 off right now
Roomba robot vacuums are up to 620 off right now via Wellbots. These are some real-deal discounts, so if you've been saving up for a robovac, it might be time to take the plunge. The best deal of the bunch is for the iRobot Roomba Combo j9, which is down to 780. The standard asking price is a whopping 1,400, so this is a savings of 620 on the hybrid vacuum/mop. All you have to do is jot down "ENGAD620" in the checkout field.
One Prompt is not Enough: Automated Construction of a Mixture-of-Expert Prompts
Wang, Ruochen, An, Sohyun, Cheng, Minhao, Zhou, Tianyi, Hwang, Sung Ju, Hsieh, Cho-Jui
Large Language Models (LLMs) exhibit strong generalization capabilities to novel tasks when prompted with language instructions and in-context demos. Since this ability sensitively depends on the quality of prompts, various methods have been explored to automate the instruction design. While these methods demonstrated promising results, they also restricted the searched prompt to one instruction. Such simplification significantly limits their capacity, as a single demo-free instruction might not be able to cover the entire complex problem space of the targeted task. To alleviate this issue, we adopt the Mixture-of-Expert paradigm and divide the problem space into a set of sub-regions; Each sub-region is governed by a specialized expert, equipped with both an instruction and a set of demos. A two-phase process is developed to construct the specialized expert for each region: (1) demo assignment: Inspired by the theoretical connection between in-context learning and kernel regression, we group demos into experts based on their semantic similarity; (2) instruction assignment: A region-based joint search of an instruction per expert complements the demos assigned to it, yielding a synergistic effect. The resulting method, codenamed Mixture-of-Prompts (MoP), achieves an average win rate of 81% against prior arts across several major benchmarks.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- Europe > Austria > Vienna (0.14)
- Asia > Middle East > Jordan (0.04)
- (2 more...)
- Media > Film (0.67)
- Leisure & Entertainment (0.67)
- Information Technology > Artificial Intelligence > Natural Language > Large Language Model (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Optimization (0.93)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.47)