Current large language models generate tokens by iteratively updating high-dimensional representations. When solving math problems with chain-of-thought prompting, the sequence of reasoning steps can be viewed as successive states forming a trajectory through the model's representation space. We characterize this trajectory and find it is highly structured: each reasoning step occupies a distinct, linearly separable region that becomes progressively more delineated at deeper layers. This organization is already present in base models — reasoning training primarily reshapes when convergence occurs rather than introducing new representational structure.
Building on this, we show that correct and incorrect solutions follow similar early-step paths but diverge systematically at late steps, yielding actionable mid-reasoning correctness signals. We further introduce trajectory-based steering, an inference-time intervention framework that enables both error correction and reasoning length control without retraining.