IlyasMoutawwakil HF staff commited on
Commit
a71ee4d
·
verified ·
1 Parent(s): e07f364

Upload cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_training_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "model": "openai-community/gpt2",
@@ -81,7 +81,7 @@
81
  "machine": "x86_64",
82
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
- "python_version": "3.10.14",
85
  "gpu": [
86
  "NVIDIA A10G"
87
  ],
@@ -107,7 +107,7 @@
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
- "max_ram": 1106.993152,
111
  "max_global_vram": 3563.585536,
112
  "max_process_vram": 0.0,
113
  "max_reserved": 2915.04128,
@@ -116,24 +116,24 @@
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
- "total": 0.7857572059631348,
120
- "mean": 0.15715144119262697,
121
- "stdev": 0.2165399637889725,
122
- "p50": 0.04862771224975586,
123
- "p90": 0.37436704101562507,
124
- "p95": 0.4822972595214843,
125
- "p99": 0.5686414343261719,
126
  "values": [
127
- 0.5902274780273438,
128
- 0.04862771224975586,
129
- 0.04781465530395508,
130
- 0.048510974884033206,
131
- 0.050576385498046876
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
- "value": 63.63288764079859
137
  },
138
  "energy": null,
139
  "efficiency": null
@@ -141,7 +141,7 @@
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
- "max_ram": 1106.993152,
145
  "max_global_vram": 3563.585536,
146
  "max_process_vram": 0.0,
147
  "max_reserved": 2915.04128,
@@ -150,21 +150,21 @@
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
- "total": 0.6388551902770996,
154
- "mean": 0.3194275951385498,
155
- "stdev": 0.270799882888794,
156
- "p50": 0.3194275951385498,
157
- "p90": 0.536067501449585,
158
- "p95": 0.5631474897384643,
159
- "p99": 0.5848114803695679,
160
  "values": [
161
- 0.5902274780273438,
162
- 0.04862771224975586
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
- "value": 12.522399632583479
168
  },
169
  "energy": null,
170
  "efficiency": null
@@ -172,7 +172,7 @@
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
- "max_ram": 1106.993152,
176
  "max_global_vram": 3563.585536,
177
  "max_process_vram": 0.0,
178
  "max_reserved": 2915.04128,
@@ -181,22 +181,22 @@
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
- "total": 0.14690201568603517,
185
- "mean": 0.04896733856201172,
186
- "stdev": 0.0011727430156611848,
187
- "p50": 0.048510974884033206,
188
- "p90": 0.05016330337524414,
189
- "p95": 0.05036984443664551,
190
- "p99": 0.0505350772857666,
191
  "values": [
192
- 0.04781465530395508,
193
- 0.048510974884033206,
194
- 0.050576385498046876
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 122.53065361928265
200
  },
201
  "energy": null,
202
  "efficiency": null
 
3
  "name": "cuda_training_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "model": "openai-community/gpt2",
 
81
  "machine": "x86_64",
82
  "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
+ "python_version": "3.10.12",
85
  "gpu": [
86
  "NVIDIA A10G"
87
  ],
 
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
+ "max_ram": 1126.117376,
111
  "max_global_vram": 3563.585536,
112
  "max_process_vram": 0.0,
113
  "max_reserved": 2915.04128,
 
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
+ "total": 0.8457041625976562,
120
+ "mean": 0.16914083251953124,
121
+ "stdev": 0.2386646202265978,
122
+ "p50": 0.04981145477294922,
123
+ "p90": 0.4080410461425782,
124
+ "p95": 0.5272553253173827,
125
+ "p99": 0.6226267486572266,
126
  "values": [
127
+ 0.6464696044921875,
128
+ 0.05039820861816406,
129
+ 0.04981145477294922,
130
+ 0.04964659118652344,
131
+ 0.04937830352783203
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
+ "value": 59.12232931007519
137
  },
138
  "energy": null,
139
  "efficiency": null
 
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
+ "max_ram": 1126.117376,
145
  "max_global_vram": 3563.585536,
146
  "max_process_vram": 0.0,
147
  "max_reserved": 2915.04128,
 
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
+ "total": 0.6968678131103515,
154
+ "mean": 0.34843390655517575,
155
+ "stdev": 0.29803569793701173,
156
+ "p50": 0.34843390655517575,
157
+ "p90": 0.5868624649047851,
158
+ "p95": 0.6166660346984862,
159
+ "p99": 0.6405088905334473,
160
  "values": [
161
+ 0.6464696044921875,
162
+ 0.05039820861816406
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
+ "value": 11.479939020706603
168
  },
169
  "energy": null,
170
  "efficiency": null
 
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
+ "max_ram": 1126.117376,
176
  "max_global_vram": 3563.585536,
177
  "max_process_vram": 0.0,
178
  "max_reserved": 2915.04128,
 
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
+ "total": 0.1488363494873047,
185
+ "mean": 0.04961211649576824,
186
+ "stdev": 0.00017850560871420263,
187
+ "p50": 0.04964659118652344,
188
+ "p90": 0.04977848205566406,
189
+ "p95": 0.04979496841430664,
190
+ "p99": 0.0498081575012207,
191
  "values": [
192
+ 0.04981145477294922,
193
+ 0.04964659118652344,
194
+ 0.04937830352783203
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
+ "value": 120.9381986457236
200
  },
201
  "energy": null,
202
  "efficiency": null