Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
Introduces 'Speak & Spell', an LLM-driven data augmentation method for improving Dialogue State Tracking (DST) robustness against Automatic Speech Recognition (ASR) errors. The method uses LLMs to control the placement of phonetically similar errors on keywords via prompts, generating sufficient error patterns to significantly improve DST accuracy in noisy ASR environments.
Enhances the reliability and user experience of voice-based applications (e.g., virtual assistants, call center bots) by making them more resilient to speech recognition inaccuracies.