GeomPrompt: Geometric Prompt Learning for RGB-D Semantic Segmentation Under Missing and Degraded Depth

CVPR 2026 URVIS Workshop
Krishna Jaganathan, Patricio Vela
Georgia Institute of Technology
GeomPrompt teaser figure
GeomPrompt synthesizes a geometric prompt from an RGB image for downstream semantic segmentation on a frozen RGB-D segmenter in the case of missing depth. GeomPrompt-Recovery extends this by allowing for corrections on existing degraded depth inputs.

Abstract

Multimodal perception systems for robotics and embodied AI often assume reliable RGB-D sensing, but in practice, depth is frequently missing, noisy, or corrupted. We thus present GeomPrompt, a lightweight cross-modal adaptation module that synthesizes a task-driven geometric prompt from RGB alone for the fourth channel of a frozen RGB-D semantic segmentation model, without depth supervision. We further introduce GeomPrompt-Recovery, an adaptation module that compensates for degraded depth by predicting the fourth channel correction relevant for the frozen segmenter. Both modules are trained solely with downstream segmentation supervision, enabling recovery of the geometric prior useful for segmentation, rather than estimating depth signals. On SUN RGB-D, GeomPrompt improves over RGB-only inference by +6.1 mIoU on DFormer and +3.0 mIoU on GeminiFusion, while remaining competitive with strong monocular depth estimators. For degraded depth, GeomPrompt-Recovery consistently improves robustness, yielding gains up to +3.6 mIoU under severe depth corruptions. GeomPrompt is also substantially more efficient than monocular depth baselines, reaching 7.8 ms latency versus 38.3 ms and 71.9 ms. These results suggest that task-driven geometric prompting is an efficient mechanism for cross-modal compensation under missing and degraded depth inputs in RGB-D perception.

Method Overview

GeomPrompt architecture

(a) Overview of the GeomPrompt architecture.

GeomPrompt-Recovery architecture

(b) Depth handling in GeomPrompt-Recovery.

Model architecture diagrams. (a) Overview of the GeomPrompt architecture. (b) Depth handling in GeomPrompt-Recovery.

Missing Depth (GeomPrompt)

Given RGB input \(x\), the module predicts a task-driven prompt \(p^*\) and feeds \(S(x, p^*)\) into a frozen RGB-D segmenter.

Degraded Depth (GeomPrompt-Recovery)

Given RGB and corrupted depth \(\tilde{d}\), the module predicts a bounded residual correction and produces a recovered prompt used by the same frozen segmenter.

Quantitative Results

Method DFormer mIoU DFormer PA GeminiFusion mIoU GeminiFusion PA
GT Depth (Upper Bound)51.283.452.782.8
RGB-only41.778.343.478.8
Depth Anything 244.080.847.781.4
Depth Anything 2 [Hypersim]47.581.844.579.6
Metric3Dv246.681.846.680.6
GeomPrompt47.881.646.480.3

GeomPrompt improves over RGB-only by +6.1 mIoU on DFormer and +3.0 mIoU on GeminiFusion.

Qualitative Results

Qualitative segmentation comparisons
Qualitative comparison of segmentation outputs and geometric representations.
RGB depth prompt qualitative grid
Visual comparison of our synthesized geometric prompt against the original RGB input and ground truth depth.
GeomPrompt-Recovery degradation qualitative grid
GeomPrompt-Recovery under simulated depth failures.

Recovery Under Degraded Depth

Recovery evaluation under depth degradation
Degraded depth vs. our recovered prompt across severities and degradation types.
Degradation Mean mIoU Gain High-Severity Gain
Quantization+1.4+2.3
Dropout+2.0+1.5
Noise+2.5+3.6

Recovered gains across different degradation types.

Efficiency

ModelLatency (ms)GFLOPsParams (M)
Depth Anything 238.3280.797.5
Metric3Dv271.9315.237.5
GeomPrompt7.844.023.4

Efficiency comparison across methods. Lower is better.

Ablations and Controls

Training ablations mIoU plot
mIoU of training ablations vs. our baseline.
MethodDFormerGeminiFusion
RGB-only41.743.4
Luminance Gray25.443.2
Laplacian Edges40.342.4
Scharr39.843.0
Canny38.643.5
GeomPrompt47.846.4

Naive baselines and references, separated by segmenter. Best results in each segmenter group are in bold.

Conclusion

In this work, we presented GeomPrompt and GeomPrompt-Recovery as lightweight cross-modal adaptation modules for RGB-D perception under missing and degraded depth. Instead of reconstructing metric depth, the method learns task-driven geometric prompts from downstream supervision alone, enabling a frozen multimodal segmenter to remain effective when one sensing stream is unavailable or corrupted. GeomPrompt improves over RGB-only inference while remaining competitive with monocular depth estimators, and GeomPrompt-Recovery improves robustness under simulated sensor failures. These results suggest that task-driven cross-modal compensation can be a practical strategy for robust multimodal segmentation in embodied systems, especially when real-world deployment requires efficiency and graceful degradation under unreliable sensing. Future work could study whether this prompting view extends to other multimodal perception tasks useful for embodied AI, such as mapping, navigation, or manipulation.