IlyasMoutawwakil HF staff commited on
Commit
09301a4
·
verified ·
1 Parent(s): 53240e0

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1380.610048,
108
- "max_global_vram": 0.0,
109
- "max_process_vram": 0.0,
110
  "max_reserved": 589.299712,
111
  "max_allocated": 533.571072
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.803873046875,
117
- "mean": 7.803873046875,
118
  "stdev": 0.0,
119
- "p50": 7.803873046875,
120
- "p90": 7.803873046875,
121
- "p95": 7.803873046875,
122
- "p99": 7.803873046875,
123
  "values": [
124
- 7.803873046875
125
  ]
126
  },
127
  "throughput": null,
@@ -131,172 +131,169 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1479.753728,
135
- "max_global_vram": 0.0,
136
- "max_process_vram": 0.0,
137
  "max_reserved": 589.299712,
138
  "max_allocated": 439.456768
139
  },
140
  "latency": {
141
  "unit": "s",
142
- "count": 144,
143
- "total": 1.002315646648407,
144
- "mean": 0.006960525323947271,
145
- "stdev": 0.003916376624115386,
146
- "p50": 0.006557100057601928,
147
- "p90": 0.00699897050857544,
148
- "p95": 0.007232826066017151,
149
- "p99": 0.007638065500259399,
150
  "values": [
151
- 0.0076327757835388185,
152
- 0.007301097869873047,
153
- 0.05367535781860352,
154
- 0.007525257110595703,
155
- 0.006696619033813476,
156
- 0.0065849399566650394,
157
- 0.006598380088806153,
158
- 0.006553100109100342,
159
- 0.0065238199234008785,
160
- 0.006938698768615722,
161
- 0.006508140087127685,
162
- 0.006449739933013916,
163
- 0.006442379951477051,
164
- 0.006431019783020019,
165
- 0.006460461139678955,
166
- 0.006454219818115234,
167
- 0.006432139873504638,
168
- 0.006469900131225586,
169
- 0.006435181140899658,
170
- 0.00641565990447998,
171
- 0.00644782018661499,
172
- 0.006407979965209961,
173
- 0.00645534086227417,
174
- 0.0076420559883117675,
175
- 0.007114218235015869,
176
- 0.006395980834960937,
177
- 0.006369740009307861,
178
- 0.006371981143951416,
179
- 0.006364299774169922,
180
- 0.00635374116897583,
181
- 0.006351019859313965,
182
- 0.006408780097961426,
183
- 0.006368300914764404,
184
- 0.006354219913482666,
185
- 0.006385900974273682,
186
- 0.006404300212860107,
187
- 0.0063748607635498045,
188
- 0.006586538791656494,
189
- 0.006503819942474365,
190
- 0.006570219993591309,
191
- 0.006877418994903565,
192
- 0.006514860153198242,
193
- 0.006800619125366211,
194
- 0.006645420074462891,
195
- 0.006575500011444092,
196
- 0.006726539134979248,
197
- 0.006643340110778808,
198
- 0.006559500217437744,
199
- 0.006666698932647705,
200
- 0.0065041399002075196,
201
- 0.0069111790657043455,
202
- 0.006611499786376953,
203
- 0.006746058940887451,
204
- 0.006644619941711426,
205
- 0.006681579113006592,
206
- 0.006913739204406738,
207
- 0.0065342202186584476,
208
- 0.006733579158782959,
209
- 0.006833579063415527,
210
- 0.006707339763641357,
211
- 0.006572939872741699,
212
- 0.0066364588737487795,
213
- 0.006554699897766113,
214
- 0.006625420093536377,
215
- 0.006532780170440674,
216
- 0.006780619144439697,
217
- 0.006599979877471924,
218
- 0.006690218925476075,
219
- 0.006934379100799561,
220
- 0.006616779804229736,
221
- 0.006762059211730957,
222
- 0.006634860038757324,
223
- 0.006653899192810058,
224
- 0.006362220764160156,
225
- 0.006583020210266113,
226
- 0.006420139789581299,
227
- 0.0064191799163818355,
228
- 0.0069705390930175784,
229
- 0.0069798178672790525,
230
- 0.006851338863372803,
231
- 0.006739659786224365,
232
- 0.007152937889099121,
233
- 0.006869578838348388,
234
- 0.00697885799407959,
235
- 0.006837258815765381,
236
- 0.006884298801422119,
237
- 0.00720317792892456,
238
- 0.0068353390693664555,
239
- 0.00695789909362793,
240
- 0.006963179111480713,
241
- 0.006939817905426025,
242
- 0.007095179080963135,
243
- 0.006919497966766357,
244
- 0.006836618900299072,
245
- 0.007238058090209961,
246
- 0.007042219161987304,
247
- 0.006844618797302246,
248
- 0.006968138217926026,
249
- 0.006910058975219726,
250
- 0.007007178783416748,
251
- 0.006592138767242432,
252
- 0.006602220058441162,
253
- 0.006631499767303467,
254
- 0.006619819164276123,
255
- 0.00640286111831665,
256
- 0.006409739971160889,
257
- 0.006591819763183594,
258
- 0.006435979843139648,
259
- 0.006413101196289062,
260
- 0.006413259983062744,
261
- 0.006407340049743652,
262
- 0.006413900852203369,
263
- 0.0064044599533081056,
264
- 0.006417419910430908,
265
- 0.006428299903869629,
266
- 0.00640286111831665,
267
- 0.006419819831848144,
268
- 0.00644782018661499,
269
- 0.006411820888519287,
270
- 0.006411019802093506,
271
- 0.006395659923553467,
272
- 0.006406061172485352,
273
- 0.0064187002182006835,
274
- 0.006403180122375488,
275
- 0.006404780864715577,
276
- 0.006392139911651612,
277
- 0.0064006209373474125,
278
- 0.006379660129547119,
279
- 0.006387340068817139,
280
- 0.006408781051635742,
281
- 0.006428619861602783,
282
- 0.006401420116424561,
283
- 0.006382061004638672,
284
- 0.006399179935455322,
285
- 0.006402700901031494,
286
- 0.006388460159301758,
287
- 0.00640638017654419,
288
- 0.006396621227264404,
289
- 0.006423980236053467,
290
- 0.006437739849090576,
291
- 0.0064006209373474125,
292
- 0.0070838179588317875,
293
- 0.00724109697341919,
294
- 0.007252458095550537
295
  ]
296
  },
297
  "throughput": {
298
  "unit": "samples/s",
299
- "value": 143.66731725830516
300
  },
301
  "energy": null,
302
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 901.132288,
108
+ "max_global_vram": 875.33568,
109
+ "max_process_vram": 41708.969984,
110
  "max_reserved": 589.299712,
111
  "max_allocated": 533.571072
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.03537841796875,
117
+ "mean": 7.03537841796875,
118
  "stdev": 0.0,
119
+ "p50": 7.03537841796875,
120
+ "p90": 7.03537841796875,
121
+ "p95": 7.03537841796875,
122
+ "p99": 7.03537841796875,
123
  "values": [
124
+ 7.03537841796875
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1020.461056,
135
+ "max_global_vram": 938.876928,
136
+ "max_process_vram": 225659.518976,
137
  "max_reserved": 589.299712,
138
  "max_allocated": 439.456768
139
  },
140
  "latency": {
141
  "unit": "s",
142
+ "count": 141,
143
+ "total": 1.0041217913627625,
144
+ "mean": 0.007121431144416754,
145
+ "stdev": 0.0011166491282776467,
146
+ "p50": 0.006852780818939209,
147
+ "p90": 0.007442700862884522,
148
+ "p95": 0.007674059867858887,
149
+ "p99": 0.007980586624145507,
150
  "values": [
151
+ 0.007191020965576172,
152
+ 0.007138061046600342,
153
+ 0.006918220996856689,
154
+ 0.0069515018463134765,
155
+ 0.0069175810813903804,
156
+ 0.006910702228546142,
157
+ 0.006890861034393311,
158
+ 0.006862542152404785,
159
+ 0.00683118200302124,
160
+ 0.006833100795745849,
161
+ 0.007999339103698731,
162
+ 0.006797421932220459,
163
+ 0.006742221832275391,
164
+ 0.006783021926879883,
165
+ 0.006733582019805908,
166
+ 0.006813741207122802,
167
+ 0.006756301879882813,
168
+ 0.006830222129821777,
169
+ 0.006819342136383057,
170
+ 0.006814702033996582,
171
+ 0.00678398084640503,
172
+ 0.006881581783294678,
173
+ 0.00684318208694458,
174
+ 0.006835021018981933,
175
+ 0.00701406192779541,
176
+ 0.007157740116119385,
177
+ 0.007442700862884522,
178
+ 0.007359499931335449,
179
+ 0.007644138813018799,
180
+ 0.007247500896453858,
181
+ 0.0070017409324646,
182
+ 0.00703934097290039,
183
+ 0.006983820915222168,
184
+ 0.006976461887359619,
185
+ 0.00722845983505249,
186
+ 0.007196620941162109,
187
+ 0.0073204607963562015,
188
+ 0.00775101900100708,
189
+ 0.007619338989257813,
190
+ 0.007391180992126465,
191
+ 0.01985578727722168,
192
+ 0.007952457904815673,
193
+ 0.0073419008255004884,
194
+ 0.007256619930267334,
195
+ 0.007158700942993164,
196
+ 0.007532780170440674,
197
+ 0.007236461162567139,
198
+ 0.007319180965423584,
199
+ 0.007202540874481201,
200
+ 0.007200300216674805,
201
+ 0.007272621154785156,
202
+ 0.007375979900360107,
203
+ 0.007323821067810058,
204
+ 0.007418540000915528,
205
+ 0.007282380104064941,
206
+ 0.007212141036987305,
207
+ 0.007292140007019043,
208
+ 0.006812942028045654,
209
+ 0.006819662094116211,
210
+ 0.006819342136383057,
211
+ 0.006852780818939209,
212
+ 0.006817421913146973,
213
+ 0.006777101993560791,
214
+ 0.006791821956634522,
215
+ 0.00678206205368042,
216
+ 0.006949581146240234,
217
+ 0.006774221897125244,
218
+ 0.00680446195602417,
219
+ 0.006983981132507324,
220
+ 0.006775822162628174,
221
+ 0.006782382011413574,
222
+ 0.006816300868988037,
223
+ 0.006791182041168213,
224
+ 0.006800302028656006,
225
+ 0.006797582149505615,
226
+ 0.006821742057800293,
227
+ 0.006799341201782227,
228
+ 0.006812141895294189,
229
+ 0.006821102142333985,
230
+ 0.0067809419631958,
231
+ 0.006790542125701905,
232
+ 0.006759982109069824,
233
+ 0.006792780876159668,
234
+ 0.00679022216796875,
235
+ 0.006796462059020996,
236
+ 0.006783502101898193,
237
+ 0.0067900619506835935,
238
+ 0.006789902210235596,
239
+ 0.006789900779724121,
240
+ 0.006767821788787842,
241
+ 0.006751182079315185,
242
+ 0.0068011021614074706,
243
+ 0.006767501831054687,
244
+ 0.006794541835784912,
245
+ 0.006802540779113769,
246
+ 0.006775181770324707,
247
+ 0.007112781047821045,
248
+ 0.007682378768920898,
249
+ 0.007674059867858887,
250
+ 0.007680939197540283,
251
+ 0.0072111811637878415,
252
+ 0.007096780776977539,
253
+ 0.007605259895324707,
254
+ 0.007649259090423584,
255
+ 0.007691500186920166,
256
+ 0.007257579803466797,
257
+ 0.007050220966339111,
258
+ 0.0073939008712768555,
259
+ 0.007338860034942627,
260
+ 0.007150061130523681,
261
+ 0.007265419960021973,
262
+ 0.007410700798034668,
263
+ 0.0072527799606323246,
264
+ 0.007356780052185059,
265
+ 0.007362541198730469,
266
+ 0.0076438188552856445,
267
+ 0.007125741004943848,
268
+ 0.007120300769805908,
269
+ 0.0070807809829711915,
270
+ 0.007051180839538574,
271
+ 0.006980941772460937,
272
+ 0.0068367810249328614,
273
+ 0.0068182220458984375,
274
+ 0.0068041419982910155,
275
+ 0.006796942234039306,
276
+ 0.006834381103515625,
277
+ 0.006826062202453614,
278
+ 0.00680718183517456,
279
+ 0.006782221794128418,
280
+ 0.006787181854248047,
281
+ 0.006787500858306885,
282
+ 0.00680846118927002,
283
+ 0.006801901817321777,
284
+ 0.00680446195602417,
285
+ 0.006803661823272705,
286
+ 0.0068270211219787595,
287
+ 0.0068206219673156735,
288
+ 0.006822381973266602,
289
+ 0.006791341781616211,
290
+ 0.006765582084655762,
291
+ 0.00673566198348999
 
 
 
292
  ]
293
  },
294
  "throughput": {
295
  "unit": "samples/s",
296
+ "value": 140.42121305687357
297
  },
298
  "energy": null,
299
  "efficiency": null