Avada11's picture
Improve model card: Add pipeline tag, paper link, code link, and usage (#1)
e6aa46d verified
---
license: cc-by-nc-4.0
pipeline_tag: depth-estimation
---
# Manipulation as in Simulation: Enabling Accurate Geometry Perception in Robots
This repository contains the Camera Depth Models (CDMs) from the paper [Manipulation as in Simulation: Enabling Accurate Geometry Perception in Robots](https://huggingface.co/papers/2509.02530).
CDMs are proposed as a simple plugin for daily-use depth cameras, taking RGB images and raw depth signals as input to output denoised, accurate metric depth. This enables policies trained purely in simulation to transfer directly to real robots, effectively bridging the sim-to-real gap for manipulation tasks.
* **Project page:** https://manipulation-as-in-simulation.github.io/
* **Code repository:** https://github.com/ByteDance-Seed/manip-as-in-sim-suite
## Usage
To run depth inference on RGB-D camera data, follow the example from the [GitHub repository's CDM section](https://github.com/ByteDance-Seed/manip-as-in-sim-suite/tree/main/cdm):
```bash
cd cdm
python infer.py \
--encoder vitl \
--model-path /path/to/model.pth \
--rgb-image /path/to/rgb.jpg \
--depth-image /path/to/depth.png \
--output result.png
```