IlyasMoutawwakil HF staff commited on
Commit
cb54947
·
verified ·
1 Parent(s): ab4b835

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1024.741376,
115
- "max_global_vram": 875.823104,
116
- "max_process_vram": 44651.073536,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.63286083984375,
124
- "mean": 7.63286083984375,
125
  "stdev": 0.0,
126
- "p50": 7.63286083984375,
127
- "p90": 7.63286083984375,
128
- "p95": 7.63286083984375,
129
- "p99": 7.63286083984375,
130
  "values": [
131
- 7.63286083984375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,169 +138,178 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1155.575808,
142
- "max_global_vram": 942.518272,
143
- "max_process_vram": 204460.675072,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 141,
150
- "total": 1.004502215385437,
151
- "mean": 0.0071241291871307596,
152
- "stdev": 0.0004448359471596756,
153
- "p50": 0.007152297019958496,
154
- "p90": 0.007479176044464111,
155
- "p95": 0.00768125581741333,
156
- "p99": 0.00889325122833252,
157
  "values": [
158
- 0.0077199749946594235,
159
- 0.007685255050659179,
160
- 0.007798854827880859,
161
- 0.007341416835784912,
162
- 0.0074494161605834965,
163
- 0.007484295845031738,
164
- 0.00768125581741333,
165
- 0.007479176044464111,
166
- 0.007806213855743408,
167
- 0.007375977039337158,
168
- 0.0075087761878967286,
169
- 0.007199497222900391,
170
- 0.007167656898498535,
171
- 0.007080777168273926,
172
- 0.007613894939422608,
173
- 0.0073322968482971195,
174
- 0.009403169631958008,
175
- 0.008724291801452637,
176
- 0.006786698818206787,
177
- 0.006913416862487793,
178
- 0.007005097866058349,
179
- 0.007025897026062012,
180
- 0.006978218078613281,
181
- 0.007134377002716064,
182
- 0.006931337833404541,
183
- 0.0069084582328796384,
184
- 0.006986536979675293,
185
- 0.006847978115081787,
186
- 0.006731658935546875,
187
- 0.006804458141326904,
188
- 0.006831337928771972,
189
- 0.007094377040863037,
190
- 0.007051018238067627,
191
- 0.00679597806930542,
192
- 0.006839657783508301,
193
- 0.007008617877960205,
194
- 0.00675485897064209,
195
- 0.0068924579620361325,
196
- 0.006873417854309082,
197
- 0.006726858139038086,
198
- 0.006679819107055664,
199
- 0.006723817825317383,
200
- 0.00673101806640625,
201
- 0.006769898891448974,
202
- 0.0066948580741882325,
203
- 0.00672429895401001,
204
- 0.006748777866363525,
205
- 0.006711018085479736,
206
- 0.006711019039154053,
207
- 0.006736777782440186,
208
- 0.006800938129425049,
209
- 0.006674698829650879,
210
- 0.006734538078308106,
211
- 0.006741418838500977,
212
- 0.006688777923583984,
213
- 0.00673085880279541,
214
- 0.006771498203277588,
215
- 0.00683085823059082,
216
- 0.00655229902267456,
217
- 0.006726697921752929,
218
- 0.006559819221496582,
219
- 0.006532138824462891,
220
- 0.006791818141937256,
221
- 0.006570219039916992,
222
- 0.006568939208984375,
223
- 0.006547658920288086,
224
- 0.006563498973846436,
225
- 0.006500298976898193,
226
- 0.006523338794708252,
227
- 0.006573578834533691,
228
- 0.006542698860168457,
229
- 0.006548939228057861,
230
- 0.00654573917388916,
231
- 0.0065302190780639644,
232
- 0.006743819236755371,
233
- 0.006719818115234375,
234
- 0.006747817993164063,
235
- 0.006746058940887451,
236
- 0.0066714978218078615,
237
- 0.006529259204864502,
238
- 0.007035498142242432,
239
- 0.007448935985565185,
240
- 0.00741101598739624,
241
- 0.00740669584274292,
242
- 0.007398856163024902,
243
- 0.007401416778564453,
244
- 0.007386536121368408,
245
- 0.007367815971374512,
246
- 0.007368136882781982,
247
- 0.007373095989227295,
248
- 0.0073265361785888675,
249
- 0.007349417209625244,
250
- 0.007326375961303711,
251
- 0.0073182168006896975,
252
- 0.007346695899963379,
253
- 0.007349256038665771,
254
- 0.007332137107849121,
255
- 0.0073196558952331545,
256
- 0.007327177047729492,
257
- 0.007311975955963135,
258
- 0.007321257114410401,
259
- 0.007330535888671875,
260
- 0.00731693696975708,
261
- 0.007324776172637939,
262
- 0.007362697124481201,
263
- 0.007362055778503418,
264
- 0.007365096092224121,
265
- 0.007337417125701904,
266
- 0.007318376064300537,
267
- 0.007343496799468994,
268
- 0.0073412561416625974,
269
- 0.007443016052246094,
270
- 0.007274537086486816,
271
- 0.0073718161582946775,
272
- 0.007348617076873779,
273
- 0.007307655811309814,
274
- 0.007429416179656982,
275
- 0.007340617179870605,
276
- 0.007382376194000244,
277
- 0.007372935771942138,
278
- 0.0073647770881652835,
279
- 0.007325736045837403,
280
- 0.007370697021484375,
281
- 0.007395815849304199,
282
- 0.007383495807647705,
283
- 0.007345096111297607,
284
- 0.007324616909027099,
285
- 0.007305576801300049,
286
- 0.007309415817260742,
287
- 0.009005890846252442,
288
- 0.007651014804840088,
289
- 0.007639976024627686,
290
- 0.007601095199584961,
291
- 0.0071002979278564455,
292
- 0.006999016761779785,
293
- 0.006980298042297363,
294
- 0.007152297019958496,
295
- 0.007260137081146241,
296
- 0.007037257194519043,
297
- 0.006940137863159179,
298
- 0.006793257236480713
 
 
 
 
 
 
 
 
 
299
  ]
300
  },
301
  "throughput": {
302
  "unit": "samples/s",
303
- "value": 140.36803288273182
304
  },
305
  "energy": null,
306
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1384.79616,
115
+ "max_global_vram": 11.20256,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.71164208984375,
124
+ "mean": 7.71164208984375,
125
  "stdev": 0.0,
126
+ "p50": 7.71164208984375,
127
+ "p90": 7.71164208984375,
128
+ "p95": 7.71164208984375,
129
+ "p99": 7.71164208984375,
130
  "values": [
131
+ 7.71164208984375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1484.038144,
142
+ "max_global_vram": 11.264,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 150,
150
+ "total": 1.0043431382179258,
151
+ "mean": 0.006695620921452841,
152
+ "stdev": 0.001085625500033163,
153
+ "p50": 0.006451494932174682,
154
+ "p90": 0.0068565820693969725,
155
+ "p95": 0.007041421055793762,
156
+ "p99": 0.01252131572723383,
157
  "values": [
158
+ 0.007047973155975342,
159
+ 0.00703341293334961,
160
+ 0.0070025320053100585,
161
+ 0.006885892868041992,
162
+ 0.006869254112243653,
163
+ 0.00689773416519165,
164
+ 0.006820134162902832,
165
+ 0.00678605318069458,
166
+ 0.0067977337837219235,
167
+ 0.0069692530632019046,
168
+ 0.006730372905731201,
169
+ 0.006716934204101563,
170
+ 0.006666533946990967,
171
+ 0.006640775203704834,
172
+ 0.006665733814239502,
173
+ 0.006659334182739258,
174
+ 0.006585413932800293,
175
+ 0.006589415073394775,
176
+ 0.006602055072784424,
177
+ 0.006569893836975098,
178
+ 0.00658429479598999,
179
+ 0.006663494110107422,
180
+ 0.006590854167938232,
181
+ 0.006609893798828125,
182
+ 0.006572134971618652,
183
+ 0.006550054073333741,
184
+ 0.006561254978179932,
185
+ 0.006567173957824707,
186
+ 0.006570535182952881,
187
+ 0.009346842765808105,
188
+ 0.006748134136199951,
189
+ 0.006403335094451904,
190
+ 0.006395814895629882,
191
+ 0.006500134944915772,
192
+ 0.006439495086669922,
193
+ 0.006610214233398437,
194
+ 0.006697894096374512,
195
+ 0.006402854919433594,
196
+ 0.006452294826507568,
197
+ 0.006412294864654541,
198
+ 0.006373735904693603,
199
+ 0.006417415142059326,
200
+ 0.006621253967285156,
201
+ 0.006372135162353515,
202
+ 0.006374855041503907,
203
+ 0.006372775077819824,
204
+ 0.006399335861206054,
205
+ 0.006631653785705566,
206
+ 0.0064140548706054686,
207
+ 0.006388454914093017,
208
+ 0.006360935211181641,
209
+ 0.00655581521987915,
210
+ 0.00642973518371582,
211
+ 0.006388615131378174,
212
+ 0.006602374076843262,
213
+ 0.006434374809265136,
214
+ 0.006406374931335449,
215
+ 0.006410695075988769,
216
+ 0.00640061616897583,
217
+ 0.006629414081573486,
218
+ 0.006408774852752685,
219
+ 0.0063761348724365235,
220
+ 0.006454695224761963,
221
+ 0.006454695224761963,
222
+ 0.00660845422744751,
223
+ 0.006432775020599365,
224
+ 0.00642109489440918,
225
+ 0.006432775020599365,
226
+ 0.006450695037841797,
227
+ 0.0064404549598693845,
228
+ 0.006446215152740478,
229
+ 0.006367975234985351,
230
+ 0.006418535232543945,
231
+ 0.0064441351890563965,
232
+ 0.0064276552200317385,
233
+ 0.0064233350753784176,
234
+ 0.0063975749015808105,
235
+ 0.006440774917602539,
236
+ 0.006432935237884522,
237
+ 0.006417415142059326,
238
+ 0.015571299552917481,
239
+ 0.006397415161132813,
240
+ 0.0064134149551391606,
241
+ 0.006438694953918457,
242
+ 0.0064564552307128905,
243
+ 0.006459014892578125,
244
+ 0.00645981502532959,
245
+ 0.0064140548706054686,
246
+ 0.006423494815826416,
247
+ 0.006665254116058349,
248
+ 0.006750372886657715,
249
+ 0.006855174064636231,
250
+ 0.006814692974090576,
251
+ 0.006503174781799317,
252
+ 0.006452294826507568,
253
+ 0.006429894924163819,
254
+ 0.0064042949676513675,
255
+ 0.006629414081573486,
256
+ 0.006402375221252442,
257
+ 0.006399815082550049,
258
+ 0.006415016174316406,
259
+ 0.006389894962310791,
260
+ 0.006692934036254883,
261
+ 0.006415335178375245,
262
+ 0.006444455146789551,
263
+ 0.0064282951354980465,
264
+ 0.006400294780731201,
265
+ 0.0063927750587463375,
266
+ 0.006432455062866211,
267
+ 0.006442695140838623,
268
+ 0.006416775226593017,
269
+ 0.006386374950408936,
270
+ 0.006438374996185303,
271
+ 0.006399014949798584,
272
+ 0.006376935005187988,
273
+ 0.006411334991455078,
274
+ 0.006392615795135498,
275
+ 0.006412934780120849,
276
+ 0.006424135208129883,
277
+ 0.006417415142059326,
278
+ 0.006428774833679199,
279
+ 0.006398055076599121,
280
+ 0.006429574966430664,
281
+ 0.0064281349182128905,
282
+ 0.006588453769683838,
283
+ 0.006729574203491211,
284
+ 0.006640294075012207,
285
+ 0.006628133773803711,
286
+ 0.006422854900360107,
287
+ 0.006398215770721436,
288
+ 0.006415174961090088,
289
+ 0.0064116549491882325,
290
+ 0.006403174877166748,
291
+ 0.01574617862701416,
292
+ 0.006422854900360107,
293
+ 0.006463014125823975,
294
+ 0.006459175109863281,
295
+ 0.0072943720817565914,
296
+ 0.007261092185974121,
297
+ 0.007268771171569825,
298
+ 0.007325891971588135,
299
+ 0.006927813053131103,
300
+ 0.00680125379562378,
301
+ 0.006604613780975342,
302
+ 0.006809412956237793,
303
+ 0.006827013969421387,
304
+ 0.006634054183959961,
305
+ 0.006853413105010986,
306
+ 0.006851653099060059,
307
+ 0.006702854156494141
308
  ]
309
  },
310
  "throughput": {
311
  "unit": "samples/s",
312
+ "value": 149.3513464592939
313
  },
314
  "energy": null,
315
  "efficiency": null