View a PDF of the paper titled LEDOM: Reverse Language Model, by Xunjian Yin and 8 other authors
View PDF
HTML (experimental)
Abstract:Autoregressive language models are trained exclusively left-to-right. We explore the complementary factorization, training right-to-left at scale, and ask what reasoning patterns emerge when a model conditions on future context to predict the past. We train LEDOM, an open-source purely reverse autoregressive language model (2B/7B parameters, 435B tokens), and find it develops capabilities distinct from forward models, including abductive inference, question synthesis, and natural resolution of the reversal curse. We then explore one application of the reverse model: combining forward likelihood $P(y \mid x)$ with reverse posterior $P(x \mid y)$ through noisy channel duality. We propose Reverse Reward, which reranks forward outputs using reverse posterior estimates, and prove that bidirectional scoring penalizes hallucinated reasoning chains whose backward reconstruction degrades. Reverse Reward yields gains of up to 6.6\% on AIME 2024 and 15\% on AMC 2023 across multiple strong baselines. We release all models, code, and data here.
Submission history
From: Xunjian Yin [view email]
[v1]
Wed, 2 Jul 2025 03:52:00 UTC (310 KB)
[v2]
Wed, 7 Jan 2026 22:42:53 UTC (302 KB)
[v3]
Tue, 3 Mar 2026 05:49:08 UTC (305 KB)


![[2507.01335] LEDOM: Reverse Language Model Measuring Intelligence Efficiency of Local AI](https://skytik.cc/wp-content/uploads/2025/11/Measuring-Intelligence-Efficiency-of-Local-AI-768x448.png)