Abstract
Parallel test-time scaling is enabled for latent reasoning models using uncertainty-inspired sampling strategies and a Latent Reward Model for effective trajectory selection.
Parallel test-time scaling (TTS) is a pivotal approach for enhancing large language models (LLMs), typically by sampling multiple token-based chains-of-thought in parallel and aggregating outcomes through voting or search. Recent advances in latent reasoning, where intermediate reasoning unfolds in continuous vector spaces, offer a more efficient alternative to explicit Chain-of-Thought, yet whether such latent models can similarly benefit from parallel TTS remains open, mainly due to the absence of sampling mechanisms in continuous space, and the lack of probabilistic signals for advanced trajectory aggregation. \ This work enables parallel TTS for latent reasoning models by addressing the above issues. For sampling, we introduce two uncertainty-inspired stochastic strategies: Monte Carlo Dropout and Additive Gaussian Noise. For aggregation, we design a Latent Reward Model (LatentRM) trained with step-wise contrastive objective to score and guide latent reasoning. Extensive experiments and visualization analyses show that both sampling strategies scale effectively with compute and exhibit distinct exploration dynamics, while LatentRM enables effective trajectory selection. Together, our explorations open a new direction for scalable inference in continuous spaces. Code released at https://github.com/YRYangang/LatentTTS.
Community
The paper presents new parallel test-time scaling methods for latent reasoning models by introducing stochastic sampling techniques and a contrastive latent reward model, enabling more effective and scalable reasoning in continuous vector space.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Implicit Reasoning in Large Language Models: A Comprehensive Survey (2025)
- SIM-CoT: Supervised Implicit Chain-of-Thought (2025)
- LaDiR: Latent Diffusion Enhances LLMs for Text Reasoning (2025)
- LTA-thinker: Latent Thought-Augmented Training Framework for Large Language Models on Complex Reasoning (2025)
- Learning to Ponder: Adaptive Reasoning in Latent Space (2025)
- KaVa: Latent Reasoning via Compressed KV-Cache Distillation (2025)
- Thinking on the Fly: Test-Time Reasoning Enhancement via Latent Thought Policy Optimization (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper