Improve model card: Add pipeline tag, paper link, code link, and usage (#1)
Browse files- Improve model card: Add pipeline tag, paper link, code link, and usage (9d34cfc1fcf0bc7179b56266e99537d3d0441d11)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
@@ -1,9 +1,27 @@
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
|
|
3 |
---
|
4 |
|
5 |
-
|
6 |
|
7 |
-
|
8 |
|
9 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: cc-by-nc-4.0
|
3 |
+
pipeline_tag: depth-estimation
|
4 |
---
|
5 |
|
6 |
+
# Manipulation as in Simulation: Enabling Accurate Geometry Perception in Robots
|
7 |
|
8 |
+
This repository contains the Camera Depth Models (CDMs) from the paper [Manipulation as in Simulation: Enabling Accurate Geometry Perception in Robots](https://huggingface.co/papers/2509.02530).
|
9 |
|
10 |
+
CDMs are proposed as a simple plugin for daily-use depth cameras, taking RGB images and raw depth signals as input to output denoised, accurate metric depth. This enables policies trained purely in simulation to transfer directly to real robots, effectively bridging the sim-to-real gap for manipulation tasks.
|
11 |
+
|
12 |
+
* **Project page:** https://manipulation-as-in-simulation.github.io/
|
13 |
+
* **Code repository:** https://github.com/ByteDance-Seed/manip-as-in-sim-suite
|
14 |
+
|
15 |
+
## Usage
|
16 |
+
|
17 |
+
To run depth inference on RGB-D camera data, follow the example from the [GitHub repository's CDM section](https://github.com/ByteDance-Seed/manip-as-in-sim-suite/tree/main/cdm):
|
18 |
+
|
19 |
+
```bash
|
20 |
+
cd cdm
|
21 |
+
python infer.py \
|
22 |
+
--encoder vitl \
|
23 |
+
--model-path /path/to/model.pth \
|
24 |
+
--rgb-image /path/to/rgb.jpg \
|
25 |
+
--depth-image /path/to/depth.png \
|
26 |
+
--output result.png
|
27 |
+
```
|