Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_cv 90% Match Research Paper AI researchers,Machine learning engineers,Scientists using microscopy,Data scientists,Foundation model developers 20 hours ago

Adapting General-Purpose Foundation Models for X-ray Ptychography in Low-Data Regimes

large-language-models › multimodal-llms
📄 Abstract

Abstract: The automation of workflows in advanced microscopy is a key goal where foundation models like Language Models (LLMs) and Vision-Language Models (VLMs) show great potential. However, adapting these general-purpose models for specialized scientific tasks is critical, and the optimal domain adaptation strategy is often unclear. To address this, we introduce PtychoBench, a new multi-modal, multi-task benchmark for ptychographic analysis. Using this benchmark, we systematically compare two specialization strategies: Supervised Fine-Tuning (SFT) and In-Context Learning (ICL). We evaluate these strategies on a visual artifact detection task with VLMs and a textual parameter recommendation task with LLMs in a data-scarce regime. Our findings reveal that the optimal specialization pathway is task-dependent. For the visual task, SFT and ICL are highly complementary, with a fine-tuned model guided by context-aware examples achieving the highest mean performance (Micro-F1 of 0.728). Conversely, for the textual task, ICL on a large base model is the superior strategy, reaching a peak Micro-F1 of 0.847 and outperforming a powerful "super-expert" SFT model (0-shot Micro-F1 of 0.839). We also confirm the superiority of context-aware prompting and identify a consistent contextual interference phenomenon in fine-tuned models. These results, benchmarked against strong baselines including GPT-4o and a DINOv3-based classifier, offer key observations for AI in science: the optimal specialization path in our benchmark is dependent on the task modality, offering a clear framework for developing more effective science-based agentic systems.

Key Contributions

Introduces PtychoBench, a benchmark for ptychographic analysis, and systematically compares Supervised Fine-Tuning (SFT) and In-Context Learning (ICL) for adapting foundation models (VLMs and LLMs) in low-data regimes. It demonstrates that the optimal adaptation strategy is task-dependent, with SFT and ICL being complementary for visual tasks.

Business Value

Accelerates the adoption of advanced AI models in scientific research by providing clear guidance on domain adaptation strategies. This can lead to faster discoveries and more efficient use of scientific instruments like ptychography microscopes.