Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Large Language Models (LLMs), constrained by limited context windows, often
face significant performance degradation when reasoning over long contexts. To
address this, Retrieval-Augmented Generation (RAG) retrieves and reasons over
chunks but frequently sacrifices logical coherence due to its reliance on
similarity-based rankings. Similarly, divide-and-conquer frameworks (DCF) split
documents into small chunks for independent reasoning and aggregation. While
effective for local reasoning, DCF struggles to capture long-range dependencies
and risks inducing conflicts by processing chunks in isolation. To overcome
these limitations, we propose ToM, a novel Tree-oriented MapReduce framework
for long-context reasoning. ToM leverages the inherent hierarchical structure
of long documents (e.g., main headings and subheadings) by constructing a
DocTree through hierarchical semantic parsing and performing bottom-up
aggregation. Using a Tree MapReduce approach, ToM enables recursive reasoning:
in the Map step, rationales are generated at child nodes; in the Reduce step,
these rationales are aggregated across sibling nodes to resolve conflicts or
reach consensus at parent nodes. Experimental results on 70B+ LLMs show that
ToM significantly outperforms existing divide-and-conquer frameworks and
retrieval-augmented generation methods, achieving better logical coherence and
long-context reasoning. Our code is available at
https://github.com/gjn12-31/ToM .
Authors (8)
Jiani Guo
Zuchao Li
Jie Wu
Qianren Wang
Yun Li
Lefei Zhang
+2 more
Submitted
November 1, 2025
Key Contributions
Introduces ToM, a novel Tree-oriented MapReduce framework for long-context reasoning in LLMs. By leveraging hierarchical document structure through semantic parsing and bottom-up aggregation, ToM overcomes limitations of RAG and standard divide-and-conquer methods, enabling recursive processing for better long-range dependency capture and logical coherence.
Business Value
Enables LLMs to process and reason over much larger documents (e.g., legal contracts, research papers, books) more effectively, unlocking new applications in legal tech, research analysis, and enterprise knowledge management.