EEG-GPT: Exploring Capabilities of Large Language Models for EEG Classification and Interpretation
Kim, Jonathan W., Alaa, Ahmed, Bernardo, Danilo
–arXiv.org Artificial Intelligence
Large language models (LLMs) such as ChatGPT have garnered substantial attention in the media and among the machine learning (ML) community. LLMs represent a pivotal paradigm shift in artificial intelligence (AI), consisting of transformer architectures substantially larger in scale compared to their predecessors, such as Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks [1], and leverage internet-scale text corpora, thus excelling not only on text completion tasks but demonstrating emergent capabilities in rudimentary language reasoning [2, 3]. LLMs display several features conducive to the small data regime present in most EEG datasets, where the largest datasets typically have on the order of only thousands of EEGs. Primarily, LLMs have the capability to perform fewand even zero-shot learning [4]. Recent research has investigated how LLMs can perform few-shot learning in domains ranging from cancer drug synergy prediction to cardiac signal analysis [5, 6]. Other work has demonstrated the ability of LLMs to outperform experts in annotating political Twitter messages with zero-shot learning [7]. Additionally, previous work has shown that transformer architectures are capable of utilizing in-context learning for zero-shot tasks - in other words, utilizing information provided in the prompt in order to yield better performance on various tasks [8].
arXiv.org Artificial Intelligence
Feb-3-2024
- Country:
- Europe > Poland
- Greater Poland Province > Poznań (0.04)
- North America > United States
- California
- Alameda County > Berkeley (0.04)
- San Francisco County > San Francisco (0.14)
- Illinois > Champaign County
- Champaign (0.04)
- California
- Europe > Poland
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine > Therapeutic Area > Neurology (0.47)
- Technology: