Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Wearable systems can recognize activities from IMU data but often fail to
explain their underlying causes or contextual significance. To address this
limitation, we introduce two large-scale resources: SensorCap, comprising
35,960 IMU--caption pairs, and OpenSQA, with 199,701 question--answer pairs
designed for causal and explanatory reasoning. OpenSQA includes a curated
tuning split (Tune-OpenSQA) optimized for scientific accuracy, narrative
clarity, and diagnostic insight. Leveraging these datasets, we develop LLaSA
(Large Language and Sensor Assistant), a family of compact sensor-aware
language models (7B and 13B) that generate interpretable, context-rich
responses to open-ended questions grounded in raw IMU data. LLaSA outperforms
commercial LLMs, including GPT-3.5 and GPT-4o-mini, on benchmark and real-world
tasks, demonstrating the effectiveness of domain supervision and model
alignment for sensor reasoning. Our code repository and datasets can be found
at https://github.com/BASHLab/LLaSA.