loic-dagnas-sinequa commited on
Commit
716f57c
·
verified ·
1 Parent(s): 6820cbc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +94 -95
README.md CHANGED
@@ -1,95 +1,94 @@
1
- ---
2
- pipeline_tag: sentence-similarity
3
- tags:
4
- - feature-extraction
5
- - sentence-similarity
6
- language:
7
- - en
8
- ---
9
-
10
- # Model Card for `vectorizer.vanilla`
11
-
12
- This model is a vectorizer developed by Sinequa. It produces an embedding vector given a passage or a query. The passage vectors are stored in our vector index and the query vector is used at query time to look up relevant passages in the index.
13
-
14
- Model name: `vectorizer.vanilla`
15
-
16
- ## Supported Languages
17
-
18
- The model was trained and tested in the following languages:
19
-
20
- - English
21
-
22
- ## Scores
23
-
24
- | Metric | Value |
25
- |:-----------------------|------:|
26
- | Relevance (Recall@100) | 0.639 |
27
-
28
- Note that the relevance score is computed as an average over 14 retrieval datasets (see
29
- [details below](#evaluation-metrics)).
30
-
31
- ## Inference Times
32
-
33
- | GPU | Quantization type | Batch size 1 | Batch size 32 |
34
- |:------------------------------------------|:------------------|---------------:|---------------:|
35
- | NVIDIA A10 | FP16 | 1 ms | 5 ms |
36
- | NVIDIA A10 | FP32 | 2 ms | 20 ms |
37
- | NVIDIA T4 | FP16 | 1 ms | 14 ms |
38
- | NVIDIA T4 | FP32 | 2 ms | 53 ms |
39
- | NVIDIA L4 | FP16 | 1 ms | 5 ms |
40
- | NVIDIA L4 | FP32 | 3 ms | 25 ms |
41
-
42
- ## Gpu Memory usage
43
-
44
- | Quantization type | Memory |
45
- |:-------------------------------------------------|-----------:|
46
- | FP16 | 300 MiB |
47
- | FP32 | 500 MiB |
48
-
49
- Note that GPU memory usage only includes how much GPU memory the actual model consumes on an NVIDIA T4 GPU with a batch
50
- size of 32. It does not include the fix amount of memory that is consumed by the ONNX Runtime upon initialization which
51
- can be around 0.5 to 1 GiB depending on the used GPU.
52
-
53
- ## Requirements
54
-
55
- - Minimal Sinequa version: 11.10.0
56
- - Minimal Sinequa version for using FP16 models and GPUs with CUDA compute capability of 8.9+ (like NVIDIA L4): 11.11.0
57
- - [Cuda compute capability](https://developer.nvidia.com/cuda-gpus): above 5.0 (above 6.0 for FP16 use)
58
-
59
- ## Model Details
60
-
61
- ### Overview
62
-
63
- - Number of parameters: 23 million
64
- - Base language model: [English MiniLM-L6-H384](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased)
65
- - Insensitive to casing and accents
66
- - Output dimensions: 256 (reduced with an additional dense layer)
67
- - Training procedure: Query-passage-negative triplets for datasets that have mined hard negative data, Query-passage pairs for the rest. Number of negatives is augmented with in-batch negative strategy.
68
-
69
- ### Training Data
70
-
71
- The model have been trained using all datasets that are cited in the [all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) model.
72
-
73
- ### Evaluation Metrics
74
-
75
- To determine the relevance score, we averaged the results that we obtained when evaluating on the datasets of the
76
- [BEIR benchmark](https://github.com/beir-cellar/beir). Note that all these datasets are in English.
77
-
78
- | Dataset | Recall@100 |
79
- |:------------------|-----------:|
80
- | Average | 0.639 |
81
- | | |
82
- | Arguana | 0.969 |
83
- | CLIMATE-FEVER | 0.509 |
84
- | DBPedia Entity | 0.409 |
85
- | FEVER | 0.839 |
86
- | FiQA-2018 | 0.702 |
87
- | HotpotQA | 0.609 |
88
- | MS MARCO | 0.849 |
89
- | NFCorpus | 0.315 |
90
- | NQ | 0.786 |
91
- | Quora | 0.995 |
92
- | SCIDOCS | 0.497 |
93
- | SciFact | 0.911 |
94
- | TREC-COVID | 0.129 |
95
- | Webis-Touche-2020 | 0.427 |
 
1
+ ---
2
+ pipeline_tag: sentence-similarity
3
+ tags:
4
+ - feature-extraction
5
+ - sentence-similarity
6
+ language:
7
+ - en
8
+ ---
9
+
10
+ # Model Card for `vectorizer.vanilla`
11
+
12
+ This model is a vectorizer developed by Sinequa. It produces an embedding vector given a passage or a query. The passage vectors are stored in our vector index and the query vector is used at query time to look up relevant passages in the index.
13
+
14
+ Model name: `vectorizer.vanilla`
15
+
16
+ ## Supported Languages
17
+
18
+ The model was trained and tested in the following languages:
19
+
20
+ - English
21
+
22
+ ## Scores
23
+
24
+ | Metric | Value |
25
+ |:-----------------------|------:|
26
+ | Relevance (Recall@100) | 0.639 |
27
+
28
+ Note that the relevance score is computed as an average over 14 retrieval datasets (see
29
+ [details below](#evaluation-metrics)).
30
+
31
+ ## Inference Times
32
+
33
+ | GPU | Quantization type | Batch size 1 | Batch size 32 |
34
+ |:------------------------------------------|:------------------|---------------:|---------------:|
35
+ | NVIDIA A10 | FP16 | 1 ms | 5 ms |
36
+ | NVIDIA A10 | FP32 | 2 ms | 20 ms |
37
+ | NVIDIA T4 | FP16 | 1 ms | 14 ms |
38
+ | NVIDIA T4 | FP32 | 2 ms | 53 ms |
39
+ | NVIDIA L4 | FP16 | 1 ms | 5 ms |
40
+ | NVIDIA L4 | FP32 | 3 ms | 25 ms |
41
+
42
+ ## Gpu Memory usage
43
+
44
+ | Quantization type | Memory |
45
+ |:-------------------------------------------------|-----------:|
46
+ | FP16 | 300 MiB |
47
+ | FP32 | 500 MiB |
48
+
49
+ Note that GPU memory usage only includes how much GPU memory the actual model consumes on an NVIDIA T4 GPU with a batch
50
+ size of 32. It does not include the fix amount of memory that is consumed by the ONNX Runtime upon initialization which
51
+ can be around 0.5 to 1 GiB depending on the used GPU.
52
+
53
+ ## Requirements
54
+
55
+ - Minimal Sinequa version: 11.10.0
56
+ - [Cuda compute capability](https://developer.nvidia.com/cuda-gpus): above 5.0 (above 6.0 for FP16 use)
57
+
58
+ ## Model Details
59
+
60
+ ### Overview
61
+
62
+ - Number of parameters: 23 million
63
+ - Base language model: [English MiniLM-L6-H384](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased)
64
+ - Insensitive to casing and accents
65
+ - Output dimensions: 256 (reduced with an additional dense layer)
66
+ - Training procedure: Query-passage-negative triplets for datasets that have mined hard negative data, Query-passage pairs for the rest. Number of negatives is augmented with in-batch negative strategy.
67
+
68
+ ### Training Data
69
+
70
+ The model have been trained using all datasets that are cited in the [all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) model.
71
+
72
+ ### Evaluation Metrics
73
+
74
+ To determine the relevance score, we averaged the results that we obtained when evaluating on the datasets of the
75
+ [BEIR benchmark](https://github.com/beir-cellar/beir). Note that all these datasets are in English.
76
+
77
+ | Dataset | Recall@100 |
78
+ |:------------------|-----------:|
79
+ | Average | 0.639 |
80
+ | | |
81
+ | Arguana | 0.969 |
82
+ | CLIMATE-FEVER | 0.509 |
83
+ | DBPedia Entity | 0.409 |
84
+ | FEVER | 0.839 |
85
+ | FiQA-2018 | 0.702 |
86
+ | HotpotQA | 0.609 |
87
+ | MS MARCO | 0.849 |
88
+ | NFCorpus | 0.315 |
89
+ | NQ | 0.786 |
90
+ | Quora | 0.995 |
91
+ | SCIDOCS | 0.497 |
92
+ | SciFact | 0.911 |
93
+ | TREC-COVID | 0.129 |
94
+ | Webis-Touche-2020 | 0.427 |