Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ml 85% Match Research Paper Machine Learning Researchers,Scientific Computing Experts,AI Engineers,Physicists,Mathematicians 4 weeks ago

Decoding Partial Differential Equations: Cross-Modal Adaptation of Decoder-only Models to PDEs

large-language-models › model-architecture
📄 Abstract

Abstract: Large language models have shown great success on natural language tasks in recent years, but they have also shown great promise when adapted to new modalities, e.g., for scientific machine learning tasks. Even though decoder-only models are more popular within NLP and scale exceedingly well at generating natural language, most proposed approaches for cross-modal adaptation focus on encoder-only models, raising the question of how model architecture affects these approaches. In this paper, we therefore perform a series of ablation studies to answer this question, systematically comparing encoder-only and decoder-only models on cross-modal adaptation for time-dependent simulation tasks based on partial differential equations (PDEs). We find that decoder-only models are far worse than encoder-only models, when existing approaches are applied unmodified. In contrast to several other domains, scaling decoder-only models also does not help. To harness the potential of decoder-only models in this context, we introduce two novel approaches, Parallel Flipping and Sequence Doubling, attempting to mimic bidirectionality in autoregressive models. Both our methods improve overall performance using decoder-only models for all tasks and all cross-model adaptation methods, closing the gap to encoder-only model performance. We hope that our findings broaden the spectrum of models used on cross-modal adaptation tasks to further scientific ML.

Key Contributions

This paper investigates the effectiveness of decoder-only models for scientific machine learning tasks like solving PDEs, comparing them systematically against encoder-only models. It finds that decoder-only models perform significantly worse and do not benefit from scaling in this domain, challenging assumptions from NLP.

Business Value

Provides crucial insights for developing effective AI models for scientific simulations, guiding researchers and engineers on which architectures are best suited for specific tasks, potentially accelerating scientific discovery.