Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 95% Match Research Paper Molecular scientists,AI researchers in chemistry,Drug discovery professionals 1 week ago

Omni-Mol: Multitask Molecular Model for Any-to-any Modalities

large-language-models › multimodal-llms
📄 Abstract

Abstract: In the molecular domain, numerous studies have explored the use of multimodal large language models (LLMs) to construct a general-purpose, multi-task molecular model. However, these efforts are still far from achieving a truly universal molecular model. We identify three key challenges in this endeavor: (1) Existing molecular task datasets are typically small in scale and lack comprehensive domain coverage. (2) Tasks from different molecular subfields are difficult to effectively learn jointly through LLMs due to significant distributional shifts and competition among tasks, which introduces instability in the learning process. (3) Both inter-task and intra-task molecular representations demand different intrinsic dimensions in the language space, making it challenging to balance between redundancy and insufficiency in language model representations. To address these challenges, we innovatively categorize existing small-molecule tasks into four types: Mol2Mol, Mol2Text, Mol2Num, and Text2Mol. We then collect a dataset encompassing over 16 tasks with more than 1.4 million samples, making it the largest molecular instruction-tuning dataset to date. Leveraging the extensive pretraining of LLMs on existing chemical literature, we propose a novel multimodal LLM framework, named Omni-Mol, which unifies all small-molecule tasks and supports both molecular generation and understanding. The core of Omni-Mol is our proposed MoGE, which dynamically adapts to the intrinsic rank of different tasks. This mixture-of-experts architecture enhances the model's ability to handle diverse tasks and modalities effectively. Our model achieves unified instruction tuning across 16 tasks and attains state-of-the-art performance on 13 of them. Extensive experiments further demonstrate the scalability and versatility of Omni-Mol.
Authors (6)
Chengxin Hu
Hao Li
Yihe Yuan
Zezheng Song
Chenyang Zhao
Haixin Wang
Submitted
February 3, 2025
arXiv Category
cs.LG
arXiv PDF

Key Contributions

This paper introduces Omni-Mol, a multitask molecular model designed for any-to-any modalities. It addresses key challenges in molecular LLMs, including small dataset scales, domain coverage, task competition due to distributional shifts, and the need for balanced inter-task and intra-task representations in language spaces. The proposed model aims to achieve a truly universal molecular model.

Business Value

Enables more efficient and comprehensive molecular modeling, potentially accelerating drug discovery and materials science research by providing a unified platform for diverse molecular tasks.