Goto

Collaborating Authors

 Communications: Instructional Materials


Forgetting, Ignorance or Myopia: Revisiting Key Challenges in Online Continual Learning

Neural Information Processing Systems

Online continual learning (OCL) requires the models to learn from constant, endless streams of data. While significant efforts have been made in this field, most were focused on mitigating the catastrophic forgetting issue to achieve better classification ability, at the cost of a much heavier training workload. They overlooked that in real-world scenarios, e.g., in high-speed data stream environments, data do not pause to accommodate slow models. In this paper, we emphasize that model throughput-defined as the maximum number of training samples that a model can process within a unit of time - is equally important. It directly limits how much data a model can utilize and presents a challenging dilemma for current methods. With this understanding, we revisit key challenges in OCL from both empirical and theoretical perspectives, highlighting two critical issues beyond the well-documented catastrophic forgetting: (i) Model's ignorance: the single-pass nature of OCL challenges models to learn effective features within constrained training time and storage capacity, leading to a trade-off between effective learning and model throughput; (ii) Model's myopia: the local learning nature of OCL on the current task leads the model to adopt overly simplified, task-specific features and excessively sparse classifier, resulting in the gap between the optimal solution for the current task and the global objective. To tackle these issues, we propose the Non-sparse Classifier Evolution framework (NsCE) to facilitate effective global discriminative feature learning with minimal time cost. NsCE integrates non-sparse maximum separation regularization and targeted experience replay techniques with the help of pre-trained models, enabling rapid acquisition of new globally discriminative features. Extensive experiments demonstrate the substantial improvements of our framework in performance, throughput and real-world practicality.


Benchmarking Multimodal Agents for Open-Ended Tasks in Real Computer Environments

Neural Information Processing Systems

Autonomous agents that accomplish complex computer tasks with minimal human interventions can significantly enhance accessibility and productivity of humancomputer interactions. Existing benchmarks either lack interactive environments or are limited to specific applications/domains, failing to reflect the diversity and complexity of real-world computer use and limiting agent scalability.


A FineWeb Datasheet Dataset Details Purpose of the dataset

Neural Information Processing Systems

We released FineWeb to make large language model training more accessible to the machine learning community at large. The dataset was curated by Hugging Face. The dataset was funded by Hugging Face. The dataset is released under the Open Data Commons Attribution License (ODC-By) v1.0 license. The use of this dataset is also subject to Common-Crawl's Terms of Use.


Appendix 19 B Ethics Statement

Neural Information Processing Systems

A toric fibration is a surjective flat map f: X Y with connected fibres where (a) X is a toric variety (b) Y is a normal algebraic variety (c) dim(Y) < dim(X).


DenseFusion-1M: Merging Vision Experts for Comprehensive Multimodal Perception

Neural Information Processing Systems

Existing Multimodal Large Language Models (MLLMs) increasingly emphasize complex understanding of various visual elements, including multiple objects, text information, and spatial relations.


Joint Learning of Label and Environment Causal Independence for Graph Out-of-Distribution Generalization

Neural Information Processing Systems

We tackle the problem of graph out-of-distribution (OOD) generalization. Existing graph OOD algorithms either rely on restricted assumptions or fail to exploit environment information in training data. In this work, we propose to simultaneously incorporate label and environment causal independence (LECI) to fully make use of label and environment information, thereby addressing the challenges faced by prior methods on identifying causal and invariant subgraphs. We further develop an adversarial training strategy to jointly optimize these two properties for causal subgraph discovery with theoretical guarantees. Extensive experiments and analysis show that LECI significantly outperforms prior methods on both synthetic and real-world datasets, establishing LECI as a practical and effective solution for graph OOD generalization. Our code is available at https://github.com/divelab/LECI.




You can try Microsoft's free AI skills training for two more weeks, and I recommend you do

ZDNet

I know you've heard of gamification, but have you ever heard of festification? That's what Microsoft did last month and is continuing until May 28, with the Microsoft AI Skills Fest. It's a little odd, but it also looks like it might be a heck of a lot of fun. And you still three full weeks to participate. Microsoft's AI Skills Fest offers courses that are open for all skill levels.


New Google Labs experiments help you learn new languages in 'bite-sized' lessons

ZDNet

My wife and I like to travel to other countries, but we always face a familiar obstacle -- how to learn the language well enough to converse with people. We've tried taking language lessons, yet we invariably run into situations where we can't find the right words to express ourselves. Now, Google has launched a trio of translation tools that could help overcome this obstacle. Also: Want a quick daily podcast based on your interests? Try Google's latest AI experiment Launched on Tuesday as Google Labs experiments, the "Little Language Lessons" are designed to assist you in specific situations, especially when you're traveling in a foreign country.