Power-SMC: Low-Latency Sequence-Level Power Sampling for Training-Free LLM Reasoning
arXiv:2602.10273v1 Announce Type: new Abstract: Many recent reasoning gains in large language models can be explained as distribution sharpening: biasing generation toward high-likelihood trajectories already supported by the pretrained model, rather than modifying its weights. A natural formalization is the sequence-level power distribution $pi_alpha(ymid x)propto p_theta(ymid x)^alpha$ ($alpha>1$), which concentrates mass on whole sequences instead of adjusting token-level temperature. Prior work shows that Metropolis–Hastings (MH) sampling from this distribution recovers strong reasoning performance, but at order-of-magnitude inference slowdowns. […]