Add image-to-text pipeline tag and link to code
Browse filesThis PR improves the model card by setting the appropriate `pipeline_tag` as well as adding a link to the code.
README.md
CHANGED
@@ -1,22 +1,23 @@
|
|
1 |
---
|
|
|
2 |
license: apache-2.0
|
|
|
3 |
tags:
|
4 |
- vision
|
5 |
widget:
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
example_title: Bee
|
10 |
-
library_name: transformers
|
11 |
-
pipeline_tag: zero-shot-image-classification
|
12 |
---
|
13 |
|
14 |
# SigLIP 2 Large
|
15 |
|
16 |
-
[SigLIP 2](https://
|
17 |
[SigLIP](https://huggingface.co/papers/2303.15343) with prior, independently developed techniques
|
18 |
into a unified recipe, for improved semantic understanding, localization, and dense features.
|
19 |
|
|
|
|
|
20 |
## Intended uses
|
21 |
|
22 |
You can use the raw model for tasks like zero-shot image classification and
|
@@ -83,9 +84,7 @@ The model was trained on up to 2048 TPU-v5e chips.
|
|
83 |
|
84 |
## Evaluation results
|
85 |
|
86 |
-
Evaluation of SigLIP 2 is shown below (taken from the paper).
|
87 |
-
|
88 |
-

|
89 |
|
90 |
### BibTeX entry and citation info
|
91 |
|
@@ -99,4 +98,4 @@ Evaluation of SigLIP 2 is shown below (taken from the paper).
|
|
99 |
primaryClass={cs.CV},
|
100 |
url={https://arxiv.org/abs/2502.14786},
|
101 |
}
|
102 |
-
```
|
|
|
1 |
---
|
2 |
+
library_name: transformers
|
3 |
license: apache-2.0
|
4 |
+
pipeline_tag: image-to-text
|
5 |
tags:
|
6 |
- vision
|
7 |
widget:
|
8 |
+
- src: https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg
|
9 |
+
candidate_labels: bee in the sky, bee on the flower
|
10 |
+
example_title: Bee
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
# SigLIP 2 Large
|
14 |
|
15 |
+
[SigLIP 2](https://hf.co/papers/2502.14786) extends the pretraining objective of
|
16 |
[SigLIP](https://huggingface.co/papers/2303.15343) with prior, independently developed techniques
|
17 |
into a unified recipe, for improved semantic understanding, localization, and dense features.
|
18 |
|
19 |
+
Code: https://github.com/google-research/big_vision
|
20 |
+
|
21 |
## Intended uses
|
22 |
|
23 |
You can use the raw model for tasks like zero-shot image classification and
|
|
|
84 |
|
85 |
## Evaluation results
|
86 |
|
87 |
+
Evaluation of SigLIP 2 is shown below (taken from the paper).\n\n
|
|
|
|
|
88 |
|
89 |
### BibTeX entry and citation info
|
90 |
|
|
|
98 |
primaryClass={cs.CV},
|
99 |
url={https://arxiv.org/abs/2502.14786},
|
100 |
}
|
101 |
+
```
|