Temporal Straightening for Latent Planning

Ying Wang1, Oumayma Bounou1, Gaoyue Zhou1, Randall Balestriero2, Tim G. J. Rudner3, Yann LeCun1*, and Mengye Ren1*
1New York University 2Brown University 3University of Toronto
* Equal advising.

Abstract

Learning good representations is essential for latent planning with world models. While pretrained visual encoders produce strong semantic visual features, they are not tailored to planning and contain information irrelevant -- or even detrimental -- to planning. Inspired by the perceptual straightening hypothesis in human visual processing, we introduce temporal straightening to improve representation learning for latent planning. Using a curvature regularizer that encourages locally straightened latent trajectories, we jointly learn an encoder and a predictor. We show that reducing curvature this way makes the Euclidean distance in latent space a better proxy for the geodesic distance and improves the conditioning of the planning objective. We demonstrate empirically that temporal straightening makes gradient-based planning more stable and yields significantly higher success rates across a suite of goal-reaching tasks.

Temporal straightening teaser

Method

Inspired by the perceptual straightening hypothesis in human vision, which posits that visual systems transform complex videos into straighter internal representations, we introduce a simple approach to straighten latent trajectories for planning. Concretely, we jointly learn an encoder and a predictor of a world model, while imposing regularization on the curvature of latent trajectories during training. The training objective is:

Lpred = || ẑt+1 - sg(zt+1) ||22

Lcurv = 1 - C, where C = cos(zt+1 - zt, zt+2 - zt+1)

Ltotal = Lpred + λLcurv

Here, sg denotes stop-gradient and λ controls the strength of the straightening.

Training and planning architecture

How Good Is the Embedding Space?

We inspect the learned embedding space by measuring latent trajectory curvatures, PCA projections of latent trajectories, and latent Euclidean distances to understand the impact of straightening.

  1. implicit straightening can happen when training the encoder using the predictor loss alone;
  2. adding straightening regularization further decreases curvature of the resulting embeddings;
  3. straightening encourages the latent Euclidean distance to better align with the geodesic distance;
  4. near-perfect reconstruction can be attained with a very low feature dimensionality.
Latent curvature and success rate
Latent Curvature and Open-Loop GD Success Rate for Different Encoders. Higher cosine similarity indicates lower curvature.

We visualize the Euclidean distance between the embedding of a target state (denoted by the star) and all other states in the maze. Blue indicates smaller distance, and red indicates larger distance.

We also visualize the learned trajectory representations using PCA. While latent trajectories are highly curved in the pretrained embedding space, they become significantly smoother after straightening, and Euclidean distance becomes a more faithful proxy for geodesic progress toward the goal.

Planning

We performgradient-based planning using our models on four environments: Wall, PointMaze-UMaze, PointMaze-Medium, and PushT. We report both open-loop planning and closed-loop MPC. Open-loop planning optimizes a length-H action sequence using the terminal embedding distance to the target, while MPC executes the first action and replans at every step. Across environments, temporal straightening substantially improves planning performance.

Main planning results table

Closed-loop MPC replans at every step. The success-rate curves below show that straightening reaches high MPC success quickly, especially on Wall and UMaze.

Wall MPC results
Wall.
UMaze MPC results
PointMaze-UMaze.
Medium MPC results
PointMaze-Medium.
PushT MPC results
PushT.

Below are examples of open-loop planning across the four environments.

BibTeX

@misc{wang2026temporalstraighteninglatentplanning,
      title={Temporal Straightening for Latent Planning}, 
      author={Ying Wang and Oumayma Bounou and Gaoyue Zhou and Randall Balestriero and Tim G. J. Rudner and Yann LeCun and Mengye Ren},
      year={2026},
      eprint={2603.12231},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      url={https://arxiv.org/abs/2603.12231}
}