osaka
Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning
Continual learning agents experience a stream of (related) tasks. The main challenge is that the agent must not forget previous tasks and also adapt to novel tasks in the stream. We are interested in the intersection of two recent continual-learning scenarios. In meta-continual learning, the model is pre-trained using meta-learning to minimize catastrophic forgetting of previous tasks. In continual-meta learning, the aim is to train agents for faster remembering of previous tasks through adaptation.
A A unifying framework Data Distribution Model for Fast Weights Slow Weights Updates Evaluation Supervised Learning S, Q C f
For readability, we omit OSAKA pre-training. Replay-based methods store representative samples from the past, either in their original form (e.g., rehearsal Most prior-based methods rely on task boundaries. Since non-stationary data distributions breaks the i.i.d assumption for The update is computed from a parametric combination of the gradient of the current and previous task. Despite that, meta-continual learning is actively researched [61, 6]. Bayesian change-point detection scheme to identify whether a task has changed.
- Asia > Japan > Honshū > Kansai > Osaka Prefecture > Osaka (0.45)
- North America > Canada > Quebec > Montreal (0.14)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Instructional Material (0.46)
- Research Report > New Finding (0.46)
- Education (0.93)
- Health & Medicine (0.68)
- Information Technology > Data Science > Data Mining (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Personal Assistant Systems (0.68)
- Asia > Japan > Honshū > Kansai > Osaka Prefecture > Osaka (0.45)
- North America > Canada > Quebec > Montreal (0.14)
- Europe > Germany > North Rhine-Westphalia > Upper Bavaria > Munich (0.04)
- Instructional Material (0.46)
- Research Report > New Finding (0.46)
- Education (0.93)
- Health & Medicine (0.68)
- Information Technology > Data Science > Data Mining (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Undirected Networks > Markov Models (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.68)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Personal Assistant Systems (0.68)
Review for NeurIPS paper: Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning
Weaknesses: My main concern with the submission is that the evaluation scenario OSAKA seems too specific and designed primarily for a set of algorithms in between Meta- & Continual-Learning while failing to make a broader argument for other approaches to Continual Learning. While certain aspects of OSAKA are certainly desirable (OOD tasks, Unknown task changes, Online Evaluation) there is a strong assumption made in allowing for Pre-training that may not be adequate in certain CL settings, limiting the generality of OSAKA. Furthermore, it is unclear how aspects such as controllable non-stationarity would be implemented in Reinforcement Learning. Furthermore, I personally feel that if task-revisiting is to be implemented, new OOD tasks should be designed in a way that explicitly re-uses skills that can be learned on a previous problem in a novel setting, instead of merely re-visiting the problem without modification. The problem with this assumption in general is that Catastrophic Forgetting is significantly reduced through an implicit form of replay provided by the environment, making it difficult to tell to which extent catastrophic forgetting is actually a problem of these algorithms.
An LLM Benchmark for Addressee Recognition in Multi-modal Multi-party Dialogue
Inoue, Koji, Lala, Divesh, Elmers, Mikey, Ochi, Keiko, Kawahara, Tatsuya
Handling multi-party dialogues represents a significant step for advancing spoken dialogue systems, necessitating the development of tasks specific to multi-party interactions. To address this challenge, we are constructing a multi-modal multi-party dialogue corpus of triadic (three-participant) discussions. This paper focuses on the task of addressee recognition, identifying who is being addressed to take the next turn, a critical component unique to multi-party dialogue systems. A subset of the corpus was annotated with addressee information, revealing that explicit addressees are indicated in approximately 20% of conversational turns. To evaluate the task's complexity, we benchmarked the performance of a large language model (GPT-4o) on addressee recognition. The results showed that GPT-4o achieved an accuracy only marginally above chance, underscoring the challenges of addressee recognition in multi-party dialogue. These findings highlight the need for further research to enhance the capabilities of large language models in understanding and navigating the intricacies of multi-party conversational dynamics.
- Asia > Japan > Honshū > Kansai > Osaka Prefecture > Osaka (0.06)
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.05)
- Asia > Japan > Honshū > Kansai > Kyoto Prefecture > Kyoto (0.05)
- (3 more...)
Online Fast Adaptation and Knowledge Accumulation (OSAKA): a New Approach to Continual Learning
Continual learning agents experience a stream of (related) tasks. The main challenge is that the agent must not forget previous tasks and also adapt to novel tasks in the stream. We are interested in the intersection of two recent continual-learning scenarios. In meta-continual learning, the model is pre-trained using meta-learning to minimize catastrophic forgetting of previous tasks. In continual-meta learning, the aim is to train agents for faster remembering of previous tasks through adaptation.