IlyasMoutawwakil HF staff commited on
Commit
928e692
·
verified ·
1 Parent(s): fab4ac4

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 757.424128,
108
  "max_global_vram": 1218.9696,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 589.299712,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.372923828125,
117
- "mean": 7.372923828125,
118
  "stdev": 0.0,
119
- "p50": 7.372923828125,
120
- "p90": 7.372923828125,
121
- "p95": 7.372923828125,
122
- "p99": 7.372923828125,
123
  "values": [
124
- 7.372923828125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 9.494746263887541e-07,
131
- "ram": 5.04684826709593e-07,
132
- "gpu": 1.7002791379994858e-06,
133
- "total": 3.154438591097833e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 924.676096,
141
  "max_global_vram": 1229.45536,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 589.299712,
@@ -145,213 +145,211 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 182,
149
- "total": 1.0035744962692261,
150
- "mean": 0.0055141455838968465,
151
- "stdev": 0.00018017208046143998,
152
- "p50": 0.005557760000228881,
153
- "p90": 0.005708083057403565,
154
- "p95": 0.005764198422431946,
155
- "p99": 0.005893867349624634,
156
  "values": [
157
- 0.005832704067230224,
158
- 0.006045695781707764,
159
- 0.00586137580871582,
160
- 0.005765120029449463,
161
- 0.005836800098419189,
162
- 0.005776383876800537,
163
- 0.005781536102294922,
164
- 0.005701632022857666,
165
- 0.005787648200988769,
166
- 0.005673984050750733,
167
- 0.00561356782913208,
168
- 0.00568012809753418,
169
- 0.005610527992248535,
170
- 0.005676032066345215,
171
- 0.005725183963775635,
172
- 0.005656576156616211,
173
- 0.005542912006378174,
174
- 0.005504000186920166,
175
- 0.005544960021972656,
176
- 0.005603328227996827,
177
  0.0057118721008300784,
178
- 0.005629951953887939,
179
- 0.005673984050750733,
180
- 0.005730303764343261,
181
- 0.005627903938293457,
182
- 0.005675007820129394,
183
- 0.005612544059753418,
184
- 0.005536767959594726,
185
- 0.005537792205810547,
186
- 0.005588992118835449,
187
- 0.005686272144317627,
188
- 0.005675007820129394,
189
  0.005686272144317627,
190
- 0.005708799839019775,
191
- 0.005659647941589355,
192
- 0.0056852478981018065,
193
- 0.005599232196807862,
194
- 0.0056780800819396975,
195
- 0.0056566400527954105,
196
- 0.005628928184509278,
197
- 0.005695487976074219,
198
- 0.005657599925994873,
199
- 0.00555622386932373,
200
- 0.00550707197189331,
201
- 0.005476352214813233,
202
- 0.005549056053161621,
203
- 0.005545983791351319,
204
- 0.005618688106536865,
205
- 0.00561356782913208,
206
- 0.005619711875915527,
207
- 0.0056217598915100095,
208
- 0.005604351997375488,
209
- 0.005697535991668701,
210
- 0.005746687889099121,
211
- 0.005667840003967285,
212
- 0.005740543842315674,
213
- 0.005667840003967285,
214
- 0.005699584007263184,
215
- 0.005615615844726562,
216
- 0.0055552000999450684,
217
- 0.00557260799407959,
218
- 0.005640192031860352,
219
- 0.0056555519104003905,
220
- 0.005610496044158936,
221
- 0.0056483840942382815,
222
- 0.005654528141021729,
223
- 0.005637119770050049,
224
- 0.005619711875915527,
225
- 0.005486591815948487,
226
- 0.005578752040863037,
227
- 0.005636096000671387,
228
- 0.005559296131134033,
229
- 0.00567193603515625,
230
- 0.005677055835723877,
231
- 0.00561356782913208,
232
- 0.005561344146728516,
233
- 0.005570559978485107,
234
- 0.0056780800819396975,
235
  0.005610496044158936,
236
- 0.005653503894805908,
237
- 0.005603328227996827,
 
 
 
 
 
 
 
 
 
 
238
  0.0055848960876464845,
 
 
 
239
  0.005605375766754151,
240
- 0.005654528141021729,
241
- 0.005695487976074219,
242
- 0.005695487976074219,
243
- 0.005635072231292724,
244
- 0.005715968132019043,
245
- 0.005676032066345215,
246
- 0.005699584007263184,
247
- 0.005547008037567139,
248
- 0.005657599925994873,
249
- 0.006032383918762207,
250
- 0.005519360065460205,
251
- 0.005472256183624268,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
  0.005610496044158936,
253
- 0.005582848072052002,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
254
  0.0055203838348388675,
255
- 0.005525599956512451,
256
- 0.005587967872619629,
257
- 0.00530841588973999,
258
- 0.005362688064575195,
259
- 0.00557260799407959,
260
- 0.005486591815948487,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
261
  0.0052930560111999515,
262
- 0.005275648117065429,
263
- 0.005261312007904053,
264
- 0.005281792163848877,
265
- 0.005354496002197265,
266
- 0.005470208168029785,
267
- 0.005468160152435303,
268
- 0.005537792205810547,
269
- 0.005472256183624268,
270
  0.005585919857025146,
271
- 0.005611519813537597,
272
- 0.005738495826721191,
273
- 0.005617663860321045,
274
- 0.005345280170440674,
275
- 0.005343232154846191,
276
- 0.005329919815063477,
277
- 0.005287936210632324,
278
- 0.00531660795211792,
279
- 0.0053340158462524415,
280
- 0.005299200057983398,
281
- 0.005342207908630371,
282
- 0.005312511920928955,
283
- 0.005276671886444092,
284
- 0.005346303939819336,
285
- 0.0052971520423889164,
286
- 0.005277696132659912,
287
- 0.005329919815063477,
288
- 0.00530944013595581,
289
- 0.005282815933227539,
290
- 0.005349376201629639,
291
- 0.0052971520423889164,
292
- 0.005276671886444092,
293
- 0.005320703983306885,
294
- 0.005320703983306885,
295
- 0.005259263992309571,
296
- 0.00526643180847168,
297
- 0.005318655967712403,
298
- 0.005251071929931641,
299
- 0.005255167961120606,
300
- 0.005336063861846924,
 
 
 
 
 
 
 
301
  0.005245952129364013,
302
- 0.0052674560546875,
303
- 0.005312511920928955,
304
- 0.0052715520858764645,
305
- 0.0052715520858764645,
306
- 0.005288959980010987,
307
- 0.0052715520858764645,
308
- 0.005281792163848877,
309
- 0.00530944013595581,
310
- 0.005387263774871826,
311
- 0.005442560195922852,
312
- 0.005723135948181152,
313
- 0.005831679821014404,
314
- 0.005435391902923584,
315
- 0.00531763219833374,
316
- 0.0053043198585510255,
317
- 0.005337088108062744,
318
- 0.00557260799407959,
319
- 0.00531660795211792,
320
- 0.005346303939819336,
321
- 0.00531660795211792,
322
- 0.0053002238273620605,
323
- 0.005314559936523438,
324
- 0.005288959980010987,
325
- 0.00526643180847168,
326
- 0.005305344104766845,
327
- 0.005298175811767578,
328
- 0.00531763219833374,
329
- 0.005361663818359375,
330
- 0.005281792163848877,
331
- 0.0053002238273620605,
332
  0.005319680213928223,
333
- 0.005279744148254394,
334
- 0.005282815933227539,
335
- 0.005313536167144775,
336
- 0.0053012480735778805,
337
- 0.005283840179443359,
338
- 0.005302271842956543
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
339
  ]
340
  },
341
  "throughput": {
342
  "unit": "samples/s",
343
- "value": 181.35175881469928
344
  },
345
  "energy": {
346
  "unit": "kWh",
347
- "cpu": 6.220120351585025e-08,
348
- "ram": 3.400685456814235e-08,
349
- "gpu": 1.2783896207329774e-07,
350
- "total": 2.2404702015729035e-07
351
  },
352
  "efficiency": {
353
  "unit": "samples/kWh",
354
- "value": 4463348.806415538
355
  }
356
  }
357
  }
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 783.101952,
108
  "max_global_vram": 1218.9696,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 589.299712,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.43168505859375,
117
+ "mean": 7.43168505859375,
118
  "stdev": 0.0,
119
+ "p50": 7.43168505859375,
120
+ "p90": 7.43168505859375,
121
+ "p95": 7.43168505859375,
122
+ "p99": 7.43168505859375,
123
  "values": [
124
+ 7.43168505859375
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 9.290390680554989e-07,
131
+ "ram": 4.935459182248625e-07,
132
+ "gpu": 1.6130568459998308e-06,
133
+ "total": 3.0356418322801925e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1038.700544,
141
  "max_global_vram": 1229.45536,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 589.299712,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 180,
149
+ "total": 1.0010449900627136,
150
+ "mean": 0.0055613610559039645,
151
+ "stdev": 0.0002906952603483861,
152
+ "p50": 0.005551615953445435,
153
+ "p90": 0.005910220766067505,
154
+ "p95": 0.005975142359733581,
155
+ "p99": 0.006510613136291504,
156
  "values": [
157
+ 0.006139840126037598,
158
+ 0.005922815799713135,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
159
  0.0057118721008300784,
 
 
 
 
 
 
 
 
 
 
 
160
  0.005686272144317627,
161
+ 0.005682176113128662,
162
+ 0.00566374397277832,
163
+ 0.005642240047454834,
164
+ 0.005576704025268555,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
165
  0.005610496044158936,
166
+ 0.005616640090942383,
167
+ 0.005582848072052002,
168
+ 0.005577727794647217,
169
+ 0.0055552000999450684,
170
+ 0.005601280212402344,
171
+ 0.005689343929290771,
172
+ 0.005623807907104492,
173
+ 0.005499904155731201,
174
+ 0.005433343887329102,
175
+ 0.005501952171325684,
176
+ 0.005585984230041504,
177
+ 0.005538815975189209,
178
  0.0055848960876464845,
179
+ 0.005649407863616943,
180
+ 0.005643263816833496,
181
+ 0.005740543842315674,
182
  0.005605375766754151,
183
+ 0.005593088150024414,
184
+ 0.005505023956298828,
185
+ 0.005761023998260498,
186
+ 0.006044672012329101,
187
+ 0.005916672229766846,
188
+ 0.005909503936767578,
189
+ 0.006073344230651856,
190
+ 0.005975039958953857,
191
+ 0.0059269118309021,
192
+ 0.005894144058227539,
193
+ 0.005891071796417236,
194
+ 0.005971968173980713,
195
+ 0.006058015823364258,
196
+ 0.00592796802520752,
197
+ 0.005896192073822021,
198
+ 0.005826560020446778,
199
+ 0.005734399795532226,
200
+ 0.005471231937408447,
201
+ 0.00552243185043335,
202
+ 0.00551526403427124,
203
+ 0.005565440177917481,
204
+ 0.0055244798660278325,
205
+ 0.00553984022140503,
206
+ 0.005689343929290771,
207
+ 0.005574656009674072,
208
+ 0.005628928184509278,
209
+ 0.005624832153320313,
210
  0.005610496044158936,
211
+ 0.005661695957183838,
212
+ 0.006564864158630371,
213
+ 0.00649619197845459,
214
+ 0.006985727787017822,
215
+ 0.005846015930175781,
216
+ 0.005964799880981446,
217
+ 0.005937151908874512,
218
+ 0.005955584049224853,
219
+ 0.005844992160797119,
220
+ 0.005583871841430664,
221
+ 0.005716991901397705,
222
+ 0.005525504112243652,
223
+ 0.0054876160621643065,
224
+ 0.005472256183624268,
225
+ 0.005543935775756836,
226
+ 0.005508096218109131,
227
+ 0.005519360065460205,
228
+ 0.005643263816833496,
229
+ 0.005554175853729248,
230
+ 0.005447679996490478,
231
  0.0055203838348388675,
232
+ 0.0056555519104003905,
233
+ 0.005573631763458252,
234
+ 0.0054917120933532714,
235
+ 0.0054876160621643065,
236
+ 0.00552345609664917,
237
+ 0.00551526403427124,
238
+ 0.00553984022140503,
239
+ 0.0056258559226989744,
240
+ 0.005649407863616943,
241
+ 0.005801983833312988,
242
+ 0.005637119770050049,
243
+ 0.005649407863616943,
244
+ 0.00561356782913208,
245
+ 0.005908480167388916,
246
+ 0.005833727836608887,
247
+ 0.005866496086120605,
248
+ 0.005858304023742676,
249
+ 0.0062044157981872555,
250
+ 0.0058419198989868165,
251
+ 0.005837823867797852,
252
+ 0.005763072013854981,
253
+ 0.005746687889099121,
254
+ 0.005777408123016357,
255
+ 0.005563392162322998,
256
+ 0.00562278413772583,
257
+ 0.005784575939178467,
258
+ 0.005492735862731934,
259
+ 0.005481472015380859,
260
+ 0.005287936210632324,
261
+ 0.005419007778167725,
262
+ 0.005242879867553711,
263
+ 0.005239808082580567,
264
  0.0052930560111999515,
265
+ 0.005695487976074219,
266
+ 0.005815296173095703,
267
+ 0.005790719985961914,
268
+ 0.0058091521263122555,
269
+ 0.00586956787109375,
270
+ 0.005865471839904785,
271
+ 0.0059770879745483394,
272
+ 0.005642240047454834,
273
  0.005585919857025146,
274
+ 0.005542912006378174,
275
+ 0.005541888236999512,
276
+ 0.0055511040687561035,
277
+ 0.005526527881622314,
278
+ 0.005527552127838135,
279
+ 0.005679103851318359,
280
+ 0.005511168003082275,
281
+ 0.0055285758972167965,
282
+ 0.00555622386932373,
283
+ 0.005439487934112549,
284
+ 0.005464064121246338,
285
+ 0.005479423999786377,
286
+ 0.005426176071166992,
287
+ 0.005552127838134766,
288
+ 0.005494783878326416,
289
+ 0.005179391860961914,
290
+ 0.0051660799980163576,
291
+ 0.005167103767395019,
292
+ 0.005158912181854248,
293
+ 0.00520908784866333,
294
+ 0.005192704200744629,
295
+ 0.0051701760292053225,
296
+ 0.0051476478576660155,
297
+ 0.0051660799980163576,
298
+ 0.005138432025909424,
299
+ 0.005152768135070801,
300
+ 0.005172224044799805,
301
+ 0.005183487892150879,
302
+ 0.005185535907745361,
303
+ 0.005173247814178467,
304
+ 0.005303296089172363,
305
+ 0.005248000144958496,
306
+ 0.005214208126068115,
307
+ 0.005399551868438721,
308
+ 0.0052295680046081545,
309
+ 0.005246975898742676,
310
+ 0.005248000144958496,
311
  0.005245952129364013,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
312
  0.005319680213928223,
313
+ 0.005249023914337158,
314
+ 0.00520908784866333,
315
+ 0.005224448204040527,
316
+ 0.005238783836364746,
317
+ 0.0052367358207702636,
318
+ 0.005350399971008301,
319
+ 0.005228544235229492,
320
+ 0.0051998720169067385,
321
+ 0.005232639789581299,
322
+ 0.00521727991104126,
323
+ 0.005244927883148193,
324
+ 0.0052367358207702636,
325
+ 0.005223423957824707,
326
+ 0.005274623870849609,
327
+ 0.005214208126068115,
328
+ 0.005249023914337158,
329
+ 0.005264383792877197,
330
+ 0.0052367358207702636,
331
+ 0.005224448204040527,
332
+ 0.005231616020202637,
333
+ 0.005234687805175781,
334
+ 0.005234687805175781,
335
+ 0.005224448204040527,
336
+ 0.005224448204040527
337
  ]
338
  },
339
  "throughput": {
340
  "unit": "samples/s",
341
+ "value": 179.81209814428354
342
  },
343
  "energy": {
344
  "unit": "kWh",
345
+ "cpu": 6.272360744782444e-08,
346
+ "ram": 3.424517297276035e-08,
347
+ "gpu": 1.2708270013756815e-07,
348
+ "total": 2.2405148055815296e-07
349
  },
350
  "efficiency": {
351
  "unit": "samples/kWh",
352
+ "value": 4463259.950386483
353
  }
354
  }
355
  }