Beyond 3D VQAs: Injecting 3D Spatial Priors into
Vision-Language Models for Enhanced Geometric Reasoning

1FAIR at Meta 2UC Berkeley 3HKU
CVPR 2026 Main Track
GASP teaser
TL;DR
  • Why it matters: VLMs' internal visual representations have near-zero geometric consistency. They can't reliably tell if the same object appears in two different views. Existing fixes (VQA fine-tuning, 3D encoders) either overfit or add rigid modules.
  • Our approach: Instead of memorizing QA pairs, GASP teaches the LLM fundamental geometry through point correspondence + depth consistency supervision at every transformer layer. The training head is discarded at inference with zero overhead.
  • The payoff: Internal correspondence jumps from <5% to 70%+. Downstream: +18.2% on spatial reasoning, +29.0% on object counting, all without any 3D VQA data.

Abstract

Vision-Language Models (VLMs) often struggle with robust 3D spatial reasoning. Prevailing methods that rely on fine-tuning with 3D visual question-answering (VQA) datasets may overfit dataset-specific biases, while integrating specialized 3D visual encoders is often inflexible and cumbersome. In this paper, we argue that genuine spatial understanding should emerge from learning fundamental geometric priors, not only from high-level VQA supervision. We propose GASP (Geometric-Aware Spatial Priors), a framework that injects these priors directly into the LLM's transformer layers. GASP employs a small correspondence head, applied as a deep supervision signal across all layers, and is trained with a dual objective leveraging ground-truth geometry from large-scale video scenes: a contrastive loss on ground-truth point correspondences enforces 2D view-invariance, while a depth consistency supervision resolves 3D geometric ambiguities. Our analysis first provides a diagnostic showing that standard VLMs' internal correspondence matching accuracy is very low (often below 5%). We then demonstrate that our training substantially improves this behavior, boosting peak layer-wise correspondence to over 70% and maintaining over 85% temporal robustness while baselines remain below 5%. These internal improvements translate to significant gains on downstream spatial benchmarks, including +18.2% on All-Angles Bench and +29.0% on VSI-Bench, all without training on any 3D VQA data.

Key Findings


VLMs Are Geometrically Blind

VLMs cannot track objects across views. Q-K correspondence is below 5%. GASP boosts this to 70%+ across all layers.

Geometry > Memorization

Same data, VQA format: performance drops. GASP's geometric losses: +18.2% camera pose, +29.0% counting, +15.0% multi-view.

Train Geometry, Infer Standard

Head discarded after training. Geometric priors baked into attention weights. Zero overhead at inference.

Method

GASP method overview

GASP Framework. We augment the VLM by attaching a lightweight correspondence head to every LLM transformer layer. The head is initialized via SVD decomposition of the pre-trained query projection weights. During training, it receives a dual geometric supervision signal:

  • Contrastive Correspondence Loss: An InfoNCE loss on ground-truth point correspondences from large-scale video scenes (DL3DV) enforces 2D view-invariance across frames.
  • Depth Consistency Loss: A soft-argmax depth prediction using the correspondence distribution acts as a discriminative geometric regularizer, forcing the model to distinguish visually similar objects at different depths.

At inference, the correspondence head is discarded. The geometric priors are permanently embedded in the LLM's learned attention weights, enabling robust spatial reasoning without auxiliary inputs or additional parameters.

Training Data Examples

Training correspondence examples. For each scene, the top row shows RGB frames with point tracks (red dots) projected from the query frame; the bottom row shows corresponding depth maps used for depth consistency supervision.

Correspondence Analysis

Correspondence analysis

Visual correspondence diagnostic on LLaVA-NeXT-Video-7B (top) and Qwen2.5-VL-7B (bottom). We analyze three complementary metrics across all transformer layers:

  • (a, d) Layer-wise PCK: Baseline VLMs exhibit near-zero correspondence accuracy across all layers. GASP dramatically improves this, with peak accuracy exceeding 70%.
  • (b, e) Confidence-Accuracy Correlation: Baselines show negative correlation (high confidence predicts incorrect matches), revealing positional bias. GASP achieves strong positive correlation, indicating calibrated geometric understanding.
  • (c, f) Temporal Robustness: GASP maintains over 85% of its matching performance at long temporal distances, while baseline accuracy collapses to under 5%.

Correspondence Visualization

Visual correspondence visualization

Visual correspondence visualization. Our GASP's visual correspondence on both VLM backbones (LLaVA-NeXT-Video and Qwen2.5-VL). (a) Patch-wise correspondence with optical flow encoding movement. (b) Attention heatmap demonstrating learned correspondence for a query point.

Quantitative Results

Comparison with state-of-the-art VLMs on spatial reasoning benchmarks. We evaluate on All-Angles Bench, VSI-Bench, and BLINK. GASP consistently improves spatial reasoning across both VLM backbones.

Methods All-Angles Bench VSI-Bench BLINK
Cam. PoseManip.Rel. Dir. Obj. CountRouteRel. Dir.App. Order Spa. Rela.Rel. DepthMulti-View
GPT-4o27.341.440.946.231.541.328.576.964.560.2
Gemini-1.5-Pro25.040.329.856.236.046.334.667.150.041.3
InternVL2.5-8B31.843.734.116.928.841.134.789.577.444.4
Qwen2.5-VL-72B34.145.048.314.328.427.631.488.881.553.4
LLaVA-Onevision-72B20.547.733.843.532.539.944.678.378.253.4
VG-LLM16.530.026.967.932.440.759.284.377.250.8
AoTD32.437.626.723.528.841.423.361.549.245.1
VLM-3R22.735.930.970.245.480.540.148.347.650.1
Baseline (SFT)22.739.924.723.524.732.411.553.144.442.1
+ DL3DV VQA19.838.128.221.425.131.89.254.544.042.5
+ GASP (Full)40.943.529.852.532.541.222.047.648.457.1
Δ Improvement↑18.2↑3.6↑5.1↑29.0↑7.8↑8.8↑10.5↓5.5↑4.0↑15.0
Baseline (SFT)34.141.336.933.826.834.326.580.278.941.5
+ DL3DV VQA31.541.536.233.227.134.325.381.078.142.0
+ GASP (Full)52.840.137.241.630.440.635.088.880.753.4
Δ Improvement↑18.7↓1.2↑0.3↑7.8↑3.6↑6.3↑8.5↑8.6↑1.8↑11.9

Generalization Analysis

Generalization heatmap

The overfitting problem of 3D-VQA fine-tuning. We show the performance change of specialized spatial VLMs relative to their base models across five benchmarks. While VQA fine-tuning yields large gains on specific datasets (e.g., VSI-Bench), it consistently degrades performance on out-of-distribution benchmarks like MMSI-Bench and SpaceVista.

This stark pattern confirms that standard VQA-based training memorizes dataset-specific biases rather than acquiring genuine spatial understanding. GASP avoids this by learning fundamental geometric priors that generalize across domains.

BibTeX

@inproceedings{yeh2026gasp,
  title={Beyond 3D VQAs: Injecting 3D Spatial Priors into Vision-Language Models for Enhanced Geometric Reasoning},
  author={Yeh, Chun-Hsiao and Qian, Shengyi and Wang, Manchen and Ma, Yi and Tighe, Joseph and Xiao, Fanyi},
  booktitle={IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
  year={2026}
}