Schrödinger's Navigator: Imagining an Ensemble of Futures for Zero-Shot Object Navigation

1Fudan University, 2Shanghai Jiao Tong University, 3Shanghai University of International Business and Economics, 4Shanghai Innovation Institute

Schrödinger's Navigator: Imagining an Ensemble of Futures for Zero-Shot Object Navigation

We present qualitative examples in both simulated and real-world environments in the attached video.

Simulation. We select several challenging cases characterized by severe visual occlusions.

Real world. We show three indoor scenes: office, classroom, and common room. In the classroom, we further demonstrate robustness under two types of dynamic conditions:

  • Moving objects, such as a chair in motion.
  • Sudden occlusion, caused by a chair abruptly entering the view.
Teaser Image

Trajectory-conditioned 3D imagination helps robots see beyond occlusions.

Real-world zero-shot object navigation often fails when the target object (e.g., a cat) is hidden behind occlusions and surrounded by unknown or potentially hazardous space. Conventional navigation systems typically perceive only the immediate occluder and cannot infer what exists beyond it.

Schrödinger’s Navigator addresses this challenge by modeling unobserved regions as multiple plausible futures. It samples several trajectories around the occluding structure and uses a trajectory-conditioned 3DGS imagination model to predict expected observations along each path. This enables the robot to anticipate the post-occlusion scene and select safer, less-occluded routes that increase the likelihood of locating the target.

Abstract

Zero-shot object navigation (ZSON) requires a robot to locate a target object in a previously unseen environment without relying on pre-built maps or task-specific training. However, existing ZSON methods often struggle in realistic and cluttered environments, particularly when the scene contains heavy occlusions, unknown risks, or dynamically moving target objects. To address these challenges, we propose Schrödinger’s Navigator, a navigation framework inspired by Schrödinger’s thought experiment on uncertainty. The framework treats unobserved space as a set of plausible future worlds and reasons over them before acting. Conditioned on egocentric visual inputs and three candidate trajectories, a trajectory-conditioned 3D world model imagines future observations along each path. This enables the agent to see beyond occlusions and anticipate risks in unseen regions without requiring extra detours or dense global mapping. The imagined 3D observations are fused into the navigation map and used to update a value map. These updates guide the policy toward trajectories that avoid occlusions, reduce exposure to uncertain space, and better track moving targets. Experiments on a Go2 quadruped robot across three challenging scenarios, including severe static occlusions, unknown risks, and dynamically moving targets, show that Schrödinger’s Navigator consistently outperforms strong ZSON baselines in self-localization, object localization, and overall Success Rate in occlusion-heavy environments. These results demonstrate the effectiveness of trajectory-conditioned 3D imagination in enabling robust zero-shot object navigation.

The framework of our Schrödinger's Navigator

Introduction Image

Overview of our Navigator pipeline. Left: The system receives a goal instruction, RGB-D observations, and the robot pose as input. Bottom center: A trajectory sampler deterministically selects three candidate trajectories and conditions a 3D world model. The model predicts future 3DGS observations along these trajectories—left bypass, right bypass, and over-the-top—to infer occluded and unobserved regions. Top right: The predicted cues are fused with current observations to construct and update multi-sourced value maps and enable future-aware reasoning. This process produces a final affordance map used for intermediate waypoint selection. Bottom right: The execution unit follows the selected waypoint and generates control commands to navigate the robot continuously toward the goal.

Results

We report results in both real-world and simulation settings. The real-robot experiments validate robustness under occlusions, unknown risks, and moving targets, while the simulator experiments provide controlled comparisons across diverse environments.

Real-world (Go2)

Real-world experiment results

Performance under three indoor scenes: office, classroom, and common room.

Table: Comparison with baseline method in real-world environments. Results show success counts over ten trials per environment. The last column summarizes performance across all trials.
Scene Office Classroom Common Room All
Search for static objects
InstructNav 7/10 7/10 8/10 22/30
Ours 8/10 8/10 7/10 23/30
Search for dynamic objects
InstructNav 3/10 4/10 3/10 10/30
Ours 5/10 5/10 6/10 16/30
Sudden Obstacles
InstructNav 4/10 3/10 5/10 12/30
Ours 6/10 7/10 6/10 19/30

Simulation (HM3D)

Table: Quantitative Comparison on Simulation Results. Cell background colors indicate the best, second best, or third best.
Method Training Free HM3D
SR↑ SPL↑ DTG↓
ZSON0.2550.126--
PixNav0.3790.205--
SPNet0.3120.101--
SGM0.6020.308--
ESC0.3920.223--
VLFM0.5250.304--
VoroNav0.4200.260--
L3MVN0.5040.2314.43
TriHelper0.5650.2533.87
GAMap0.5310.260--
InstructNav0.5100.1872.89
InstructNav*0.4530.1863.38
CogNav0.7250.262--
ApexNav0.7620.380--
Ours 0.609 0.237 2.23