IlyasMoutawwakil HF staff commited on
Commit
37a6d34
·
verified ·
1 Parent(s): 1ff023b

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.6.0.dev20240917+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 805.441536,
108
  "max_global_vram": 1226.309632,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 589.299712,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 8.5140986328125,
117
- "mean": 8.5140986328125,
118
  "stdev": 0.0,
119
- "p50": 8.5140986328125,
120
- "p90": 8.5140986328125,
121
- "p95": 8.5140986328125,
122
- "p99": 8.5140986328125,
123
  "values": [
124
- 8.5140986328125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 9.640726444584086e-07,
131
- "ram": 5.024944429325085e-07,
132
- "gpu": 1.7833347600026972e-06,
133
- "total": 3.249901847393614e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1057.243136,
141
  "max_global_vram": 1236.795392,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 589.299712,
@@ -145,209 +145,212 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 178,
149
- "total": 1.0019687366485595,
150
- "mean": 0.005629037846340223,
151
- "stdev": 0.000190092244698167,
152
- "p50": 0.005594111919403076,
153
- "p90": 0.005733068752288819,
154
- "p95": 0.005921689510345461,
155
- "p99": 0.006492856445312498,
156
  "values": [
157
- 0.006271999835968017,
158
- 0.006441984176635742,
159
- 0.006318079948425293,
160
- 0.0062679038047790524,
161
- 0.006179840087890625,
162
- 0.006232063770294189,
163
- 0.006170623779296875,
164
- 0.006663167953491211,
165
- 0.005761023998260498,
166
- 0.005482495784759522,
167
- 0.005470208168029785,
168
- 0.005450751781463623,
169
- 0.005467135906219483,
170
- 0.005430272102355957,
171
- 0.005484543800354004,
172
- 0.005440512180328369,
173
- 0.00587775993347168,
174
- 0.005652480125427246,
175
- 0.005629951953887939,
176
- 0.00586956787109375,
177
- 0.005600255966186523,
178
- 0.005595136165618897,
179
- 0.005591040134429932,
180
- 0.005601280212402344,
181
  0.00563097620010376,
182
- 0.005586944103240967,
183
- 0.005633024215698243,
184
- 0.005602303981781006,
 
 
 
 
 
 
 
185
  0.005612544059753418,
186
- 0.005629951953887939,
187
- 0.005598207950592041,
188
- 0.005652480125427246,
189
- 0.005631008148193359,
190
- 0.0056442880630493165,
191
- 0.005666816234588623,
192
- 0.005616640090942383,
193
- 0.0056442880630493165,
194
- 0.005561344146728516,
195
- 0.005647359848022461,
196
- 0.005602303981781006,
197
- 0.005587967872619629,
198
  0.005601280212402344,
 
 
199
  0.0055920639038085935,
200
- 0.005608448028564453,
201
- 0.0056780800819396975,
202
- 0.005791744232177734,
203
- 0.005662720203399658,
 
 
204
  0.005605375766754151,
205
  0.005598207950592041,
206
- 0.0055920639038085935,
207
- 0.005594111919403076,
208
- 0.005658624172210694,
209
- 0.005578752040863037,
210
- 0.005598207950592041,
211
- 0.005658624172210694,
212
- 0.005603328227996827,
213
- 0.005609471797943116,
214
- 0.005574656009674072,
215
- 0.005568511962890625,
216
- 0.005608448028564453,
217
- 0.005594111919403076,
218
- 0.005604351997375488,
219
- 0.00572108793258667,
220
- 0.005617663860321045,
221
- 0.005576704025268555,
222
- 0.005542912006378174,
223
- 0.0055920639038085935,
224
- 0.005582848072052002,
225
- 0.0056217598915100095,
226
- 0.00555622386932373,
227
- 0.005578752040863037,
228
- 0.005787648200988769,
229
- 0.00561356782913208,
230
- 0.005627903938293457,
231
- 0.005647359848022461,
232
- 0.005573631763458252,
233
- 0.005608448028564453,
234
- 0.005593088150024414,
235
- 0.00561356782913208,
236
- 0.005595136165618897,
237
- 0.005606400012969971,
238
- 0.005609471797943116,
239
- 0.005633024215698243,
240
- 0.005631999969482422,
241
- 0.005626880168914795,
242
- 0.0056217598915100095,
243
- 0.0056145920753479005,
244
- 0.005571616172790527,
245
- 0.005615615844726562,
246
- 0.0055808000564575196,
247
- 0.005591040134429932,
248
- 0.005600255966186523,
249
- 0.005570559978485107,
250
- 0.005649407863616943,
251
- 0.005796864032745362,
252
- 0.00561356782913208,
253
- 0.005643263816833496,
254
- 0.005686272144317627,
255
- 0.0056145920753479005,
256
- 0.005511168003082275,
257
- 0.00552345609664917,
258
- 0.005506048202514649,
259
- 0.005541888236999512,
260
- 0.00553984022140503,
261
- 0.005533696174621582,
262
- 0.005516287803649903,
263
- 0.005541888236999512,
264
- 0.005497856140136719,
265
- 0.005594111919403076,
266
- 0.005559296131134033,
267
- 0.005676032066345215,
268
- 0.005848063945770264,
269
- 0.005683199882507324,
270
- 0.005627903938293457,
271
- 0.0055511040687561035,
272
- 0.005505023956298828,
273
- 0.005530623912811279,
274
- 0.00547430419921875,
275
- 0.005534719944000244,
276
- 0.005480447769165039,
277
- 0.0055101442337036136,
278
- 0.00553984022140503,
279
- 0.005511168003082275,
280
- 0.00566476821899414,
281
- 0.0055511040687561035,
282
  0.00551526403427124,
283
- 0.005545983791351319,
284
- 0.005529600143432617,
285
- 0.0055848960876464845,
286
- 0.005669888019561768,
287
- 0.005598207950592041,
288
- 0.005543935775756836,
289
- 0.005527552127838135,
290
- 0.005532671928405761,
291
- 0.0055623679161071774,
292
- 0.005519360065460205,
293
- 0.005565440177917481,
294
- 0.005500927925109863,
295
- 0.005524511814117432,
296
- 0.0055285758972167965,
297
- 0.005508096218109131,
298
  0.0055285758972167965,
299
- 0.005513216018676758,
300
- 0.005763072013854981,
301
- 0.005638144016265869,
302
- 0.005499904155731201,
303
- 0.005492735862731934,
304
- 0.005548031806945801,
305
- 0.005530623912811279,
306
- 0.005765120029449463,
307
- 0.005516287803649903,
308
- 0.005519360065460205,
309
- 0.005582848072052002,
310
  0.0055214080810546875,
311
- 0.005702655792236328,
312
- 0.005627903938293457,
313
- 0.005653503894805908,
314
- 0.005665791988372803,
315
- 0.005633024215698243,
 
316
  0.005600255966186523,
317
- 0.005534719944000244,
318
- 0.005547008037567139,
 
 
319
  0.00556441593170166,
 
 
 
 
 
 
 
 
 
 
320
  0.005547008037567139,
321
- 0.005545983791351319,
322
- 0.0055214080810546875,
323
- 0.005513216018676758,
324
- 0.005529600143432617,
325
- 0.005493760108947754,
326
- 0.005493760108947754,
327
  0.005516287803649903,
328
- 0.005488639831542969,
329
- 0.005536767959594726,
330
- 0.005543935775756836,
331
- 0.006716415882110595,
332
- 0.005701632022857666,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
333
  0.005594111919403076,
334
- 0.0055582718849182125
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
335
  ]
336
  },
337
  "throughput": {
338
  "unit": "samples/s",
339
- "value": 177.65025343543576
340
  },
341
  "energy": {
342
  "unit": "kWh",
343
- "cpu": 6.661654104201233e-08,
344
- "ram": 3.631792577680325e-08,
345
- "gpu": 1.3294703373187728e-07,
346
- "total": 2.3588150055069284e-07
347
  },
348
  "efficiency": {
349
  "unit": "samples/kWh",
350
- "value": 4239416.815924028
351
  }
352
  }
353
  }
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 792.457216,
108
  "max_global_vram": 1226.309632,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 589.299712,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.98373388671875,
117
+ "mean": 7.98373388671875,
118
  "stdev": 0.0,
119
+ "p50": 7.98373388671875,
120
+ "p90": 7.98373388671875,
121
+ "p95": 7.98373388671875,
122
+ "p99": 7.98373388671875,
123
  "values": [
124
+ 7.98373388671875
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 2.3917015840276515e-06,
131
+ "ram": 1.291773976574912e-06,
132
+ "gpu": 3.511947253999944e-06,
133
+ "total": 7.195422814602508e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1048.00256,
141
  "max_global_vram": 1236.795392,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 589.299712,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 181,
149
+ "total": 1.0017149133682248,
150
+ "mean": 0.005534336537945995,
151
+ "stdev": 0.00032811593046687886,
152
+ "p50": 0.0055316481590271,
153
+ "p90": 0.005672959804534912,
154
+ "p95": 0.005851136207580566,
155
+ "p99": 0.007023206233978258,
156
  "values": [
157
+ 0.007910399913787843,
158
+ 0.007987199783325195,
159
+ 0.006801407814025879,
160
+ 0.005938176155090332,
161
+ 0.0058716158866882326,
162
+ 0.005896192073822021,
163
+ 0.005825535774230957,
164
+ 0.00588595199584961,
165
+ 0.005666816234588623,
166
+ 0.005658656120300293,
167
+ 0.005708799839019775,
168
+ 0.005638144016265869,
169
+ 0.005658624172210694,
170
+ 0.005573631763458252,
171
+ 0.005559296131134033,
 
 
 
 
 
 
 
 
 
172
  0.00563097620010376,
173
+ 0.005492735862731934,
174
+ 0.005459968090057373,
175
+ 0.005437439918518067,
176
+ 0.005467135906219483,
177
+ 0.00557260799407959,
178
+ 0.005654528141021729,
179
+ 0.005634047985076904,
180
+ 0.005620736122131348,
181
+ 0.005620736122131348,
182
+ 0.005676032066345215,
183
  0.005612544059753418,
184
+ 0.005633024215698243,
185
+ 0.005553152084350586,
186
+ 0.00552243185043335,
187
+ 0.0055797758102417,
188
+ 0.005582848072052002,
189
+ 0.005646336078643799,
 
 
 
 
 
 
190
  0.005601280212402344,
191
+ 0.00559007978439331,
192
+ 0.0056514558792114256,
193
  0.0055920639038085935,
194
+ 0.005637119770050049,
195
+ 0.005569536209106446,
196
+ 0.005543968200683594,
197
+ 0.005618688106536865,
198
+ 0.005606400012969971,
199
+ 0.005602303981781006,
200
  0.005605375766754151,
201
  0.005598207950592041,
202
+ 0.0055316481590271,
203
+ 0.005444608211517334,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
204
  0.00551526403427124,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
205
  0.0055285758972167965,
206
+ 0.005532671928405761,
207
+ 0.005611519813537597,
208
+ 0.00552345609664917,
209
+ 0.00552345609664917,
210
+ 0.0055552000999450684,
 
 
 
 
 
 
211
  0.0055214080810546875,
212
+ 0.006051839828491211,
213
+ 0.005758975982666016,
214
+ 0.005851136207580566,
215
+ 0.005604351997375488,
216
+ 0.005588992118835449,
217
+ 0.005654528141021729,
218
  0.005600255966186523,
219
+ 0.00563097620010376,
220
+ 0.005544960021972656,
221
+ 0.005493760108947754,
222
+ 0.005529600143432617,
223
  0.00556441593170166,
224
+ 0.006354944229125976,
225
+ 0.005596159934997558,
226
+ 0.005653503894805908,
227
+ 0.005573631763458252,
228
+ 0.005602303981781006,
229
+ 0.005635072231292724,
230
+ 0.005495808124542236,
231
+ 0.005431295871734619,
232
+ 0.005465087890625,
233
+ 0.00552243185043335,
234
  0.005547008037567139,
 
 
 
 
 
 
235
  0.005516287803649903,
236
+ 0.005606400012969971,
237
+ 0.005552127838134766,
238
+ 0.005435391902923584,
239
+ 0.005548031806945801,
240
+ 0.005571584224700928,
241
+ 0.00550707197189331,
242
+ 0.0055552000999450684,
243
+ 0.0055316481590271,
244
+ 0.0055920639038085935,
245
+ 0.005567488193511963,
246
+ 0.0055316481590271,
247
+ 0.005672959804534912,
248
+ 0.005632063865661621,
249
+ 0.00566374397277832,
250
+ 0.005640192031860352,
251
+ 0.005585919857025146,
252
+ 0.00567193603515625,
253
+ 0.005604351997375488,
254
+ 0.005607423782348633,
255
+ 0.005586944103240967,
256
+ 0.0055797758102417,
257
+ 0.005703680038452149,
258
+ 0.00551423978805542,
259
+ 0.005440512180328369,
260
+ 0.005560319900512695,
261
+ 0.005538815975189209,
262
+ 0.005483520030975342,
263
+ 0.005398528099060058,
264
+ 0.005398528099060058,
265
+ 0.005372928142547607,
266
+ 0.005292031764984131,
267
+ 0.005438464164733887,
268
+ 0.005489664077758789,
269
+ 0.005346303939819336,
270
+ 0.0052367358207702636,
271
+ 0.005277696132659912,
272
+ 0.005227519989013672,
273
+ 0.0052367358207702636,
274
+ 0.00530841588973999,
275
+ 0.005261312007904053,
276
  0.005594111919403076,
277
+ 0.005596159934997558,
278
+ 0.005595136165618897,
279
+ 0.005708799839019775,
280
+ 0.005754879951477051,
281
+ 0.005786623954772949,
282
+ 0.005607423782348633,
283
+ 0.005526527881622314,
284
+ 0.005553152084350586,
285
+ 0.005552127838134766,
286
+ 0.005438464164733887,
287
+ 0.00547430419921875,
288
+ 0.005400576114654541,
289
+ 0.005452799797058106,
290
+ 0.005414912223815918,
291
+ 0.00537497615814209,
292
+ 0.005476352214813233,
293
+ 0.005405695915222168,
294
+ 0.00535756778717041,
295
+ 0.00542310380935669,
296
+ 0.005375999927520752,
297
+ 0.00537497615814209,
298
+ 0.005421055793762207,
299
+ 0.005379072189331055,
300
+ 0.005375999927520752,
301
+ 0.005417984008789062,
302
+ 0.005406720161437988,
303
+ 0.00537395191192627,
304
+ 0.005633024215698243,
305
+ 0.005252096176147461,
306
+ 0.005234687805175781,
307
+ 0.005248000144958496,
308
+ 0.005219327926635742,
309
+ 0.005252096176147461,
310
+ 0.005272575855255127,
311
+ 0.005234687805175781,
312
+ 0.0052367358207702636,
313
+ 0.005291007995605469,
314
+ 0.0052408318519592285,
315
+ 0.005215231895446777,
316
+ 0.005250048160552978,
317
+ 0.005237760066986084,
318
+ 0.005228544235229492,
319
+ 0.005250048160552978,
320
+ 0.0052705278396606445,
321
+ 0.005243904113769531,
322
+ 0.005228544235229492,
323
+ 0.005268479824066162,
324
+ 0.005228544235229492,
325
+ 0.005228544235229492,
326
+ 0.005275680065155029,
327
+ 0.005224448204040527,
328
+ 0.005228544235229492,
329
+ 0.0052674560546875,
330
+ 0.005201920032501221,
331
+ 0.00521830415725708,
332
+ 0.005253119945526123,
333
+ 0.0052367358207702636,
334
+ 0.005221375942230225,
335
+ 0.0052008957862854,
336
+ 0.005254144191741943,
337
+ 0.005228544235229492
338
  ]
339
  },
340
  "throughput": {
341
  "unit": "samples/s",
342
+ "value": 180.69013207699481
343
  },
344
  "energy": {
345
  "unit": "kWh",
346
+ "cpu": 6.31605106826256e-08,
347
+ "ram": 3.448539041444924e-08,
348
+ "gpu": 1.3397321592553254e-07,
349
+ "total": 2.3161911702260742e-07
350
  },
351
  "efficiency": {
352
  "unit": "samples/kWh",
353
+ "value": 4317432.916827819
354
  }
355
  }
356
  }