Goto

Collaborating Authors

 gimlet



GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning

Neural Information Processing Systems

Molecule property prediction has gained significant attention in recent years. The main bottleneck is the label insufficiency caused by expensive lab experiments. In order to alleviate this issue and to better leverage textual knowledge for tasks, this study investigates the feasibility of employing natural language instructions to accomplish molecule-related tasks in a zero-shot setting. We discover that existing molecule-text models perform poorly in this setting due to inadequate treatment of instructions and limited capacity for graphs. To overcome these issues, we propose GIMLET, which unifies language models for both graph and text data. By adopting generalized position embedding, our model is extended to encode both graph structures and instruction text without additional graph encoding modules.



GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning

Neural Information Processing Systems

Molecule property prediction has gained significant attention in recent years. The main bottleneck is the label insufficiency caused by expensive lab experiments. In order to alleviate this issue and to better leverage textual knowledge for tasks, this study investigates the feasibility of employing natural language instructions to accomplish molecule-related tasks in a zero-shot setting. We discover that existing molecule-text models perform poorly in this setting due to inadequate treatment of instructions and limited capacity for graphs. To overcome these issues, we propose GIMLET, which unifies language models for both graph and text data.


MolecularGPT: Open Large Language Model (LLM) for Few-Shot Molecular Property Prediction

Liu, Yuyan, Ding, Sirui, Zhou, Sheng, Fan, Wenqi, Tan, Qiaoyu

arXiv.org Artificial Intelligence

Molecular property prediction (MPP) is a fundamental and crucial task in drug discovery. However, prior methods are limited by the requirement for a large number of labeled molecules and their restricted ability to generalize for unseen and new tasks, both of which are essential for real-world applications. To address these challenges, we present MolecularGPT for few-shot MPP. From a perspective on instruction tuning, we fine-tune large language models (LLMs) based on curated molecular instructions spanning over 1000 property prediction tasks. This enables building a versatile and specialized LLM that can be adapted to novel MPP tasks without any fine-tuning through zero- and few-shot in-context learning (ICL). MolecularGPT exhibits competitive in-context reasoning capabilities across 10 downstream evaluation datasets, setting new benchmarks for few-shot molecular prediction tasks. More importantly, with just two-shot examples, MolecularGPT can outperform standard supervised graph neural network methods on 4 out of 7 datasets. It also excels state-of-the-art LLM baselines by up to 16.6% increase on classification accuracy and decrease of 199.17 on regression metrics (e.g., RMSE) under zero-shot. This study demonstrates the potential of LLMs as effective few-shot molecular property predictors. The code is available at https://github.com/NYUSHCS/MolecularGPT.


The Lifelike Illusions of A.I.

The New Yorker

In January, 1999, the Washington Post reported that the National Security Agency had issued a memo on its intranet with the subject "Furby Alert." According to the Post, the memo decreed that employees were prohibited from bringing to work any recording devices, including "toys, such as'Furbys,' with built-in recorders that repeat the audio with synthesized sound." That holiday season, the Furby, an animatronic toy resembling a small owl, had been a retail sensation; nearly two million were sold by year's end. They were now banned from N.S.A. headquarters. A worry, according to one source for the Post, was that the toy might "start talking classified." Tiger Electronics, the makers of the Furby, was perplexed.


GIMLET: A Unified Graph-Text Model for Instruction-Based Molecule Zero-Shot Learning

Zhao, Haiteng, Liu, Shengchao, Ma, Chang, Xu, Hannan, Fu, Jie, Deng, Zhi-Hong, Kong, Lingpeng, Liu, Qi

arXiv.org Artificial Intelligence

Molecule property prediction has gained significant attention in recent years. The main bottleneck is the label insufficiency caused by expensive lab experiments. In order to alleviate this issue and to better leverage textual knowledge for tasks, this study investigates the feasibility of employing natural language instructions to accomplish molecule-related tasks in a zero-shot setting. We discover that existing molecule-text models perform poorly in this setting due to inadequate treatment of instructions and limited capacity for graphs. To overcome these issues, we propose GIMLET, which unifies language models for both graph and text data. By adopting generalized position embedding, our model is extended to encode both graph structures and instruction text without additional graph encoding modules. GIMLET also decouples encoding of the graph from tasks instructions in the attention mechanism, enhancing the generalization of graph features across novel tasks. We construct a dataset consisting of more than two thousand molecule tasks with corresponding instructions derived from task descriptions. We pretrain GIMLET on the molecule tasks along with instructions, enabling the model to transfer effectively to a broad range of tasks. Experimental results demonstrate that GIMLET significantly outperforms molecule-text baselines in instruction-based zero-shot learning, even achieving closed results to supervised GNN models on tasks such as toxcast and muv.