Redirecting to original paper in 30 seconds...

Click below to go immediately or wait for automatic redirect

arxiv_ai 95% Match Research Paper LLM Researchers,AI Educators,Machine Learning Engineers 2 weeks ago

Code Execution as Grounded Supervision for LLM Reasoning

large-language-models › reasoning
📄 Abstract

Abstract: Training large language models (LLMs) with chain-of-thought (CoT) supervision has proven effective for enhancing their reasoning abilities. However, obtaining reliable and accurate reasoning supervision remains a significant challenge. We propose a scalable method for generating a high-quality CoT supervision dataset by leveraging the determinism of program execution. Unlike existing reasoning dataset generation methods that rely on costly human annotations or error-prone LLM-generated CoT, our approach extracts verifiable, step-by-step reasoning traces from code execution and transforms them into a natural language CoT reasoning. Experiments on reasoning benchmarks across various domains show that our method effectively equips LLMs with transferable reasoning abilities across diverse tasks. Furthermore, the ablation studies validate that our method produces highly accurate reasoning data and reduces overall token length during inference by reducing meaningless repetition and overthinking.
Authors (3)
Dongwon Jung
Wenxuan Zhou
Muhao Chen
Submitted
June 12, 2025
arXiv Category
cs.CL
arXiv PDF

Key Contributions

Proposes a scalable method to generate high-quality Chain-of-Thought (CoT) supervision data by leveraging the determinism of program execution. This approach extracts verifiable, step-by-step reasoning traces from code and converts them into natural language CoT, leading to LLMs with improved and transferable reasoning abilities.

Business Value

Enables more efficient and effective training of LLMs for complex reasoning tasks, potentially leading to more capable and reliable AI systems in areas requiring logical deduction.