Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Current large language models excel at broad, general-purpose tasks, but
consistently underperform when exposed to highly specialized domains that
require deep cultural, linguistic, and subject-matter expertise. In particular,
traditional medical systems such as Ayurveda embody centuries of nuanced
textual and clinical knowledge that mainstream LLMs fail to accurately
interpret or apply. We introduce AyurParam-2.9B, a domain-specialized,
bilingual language model fine-tuned from Param-1-2.9B using an extensive,
expertly curated Ayurveda dataset spanning classical texts and clinical
guidance. AyurParam's dataset incorporates context-aware, reasoning, and
objective-style Q&A in both English and Hindi, with rigorous annotation
protocols for factual precision and instructional clarity. Benchmarked on
BhashaBench-Ayur, AyurParam not only surpasses all open-source
instruction-tuned models in its size class (1.5--3B parameters), but also
demonstrates competitive or superior performance compared to much larger
models. The results from AyurParam highlight the necessity for authentic domain
adaptation and high-quality supervision in delivering reliable, culturally
congruent AI for specialized medical knowledge.
Key Contributions
AyurParam-2.9B is introduced as a state-of-the-art, domain-specialized, bilingual (Hindi-English) language model for Ayurveda. Fine-tuned on an extensive, expertly curated dataset of classical texts and clinical guidance, it significantly outperforms existing models in its size class on Ayurvedic benchmarks, demonstrating superior understanding of nuanced traditional medical knowledge.
Business Value
Enables better access to and understanding of Ayurvedic knowledge, supporting practitioners, researchers, and patients, and potentially integrating traditional medicine into modern healthcare systems.