IlyasMoutawwakil HF staff commited on
Commit
17c6734
·
verified ·
1 Parent(s): 89c3301

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1023.762432,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 39808.077824,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.601177734375,
124
- "mean": 7.601177734375,
125
  "stdev": 0.0,
126
- "p50": 7.601177734375,
127
- "p90": 7.601177734375,
128
- "p95": 7.601177734375,
129
- "p99": 7.601177734375,
130
  "values": [
131
- 7.601177734375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,168 +138,179 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1155.047424,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 223504.797696,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 140,
150
- "total": 1.004350363254547,
151
- "mean": 0.007173931166103907,
152
- "stdev": 0.00043545235243979075,
153
- "p50": 0.007355254888534547,
154
- "p90": 0.007545318222045899,
155
- "p95": 0.007592630004882812,
156
- "p99": 0.007652637991905212,
157
  "values": [
158
- 0.007475813865661621,
159
- 0.007361895084381104,
160
- 0.006965577125549316,
161
- 0.0070476560592651364,
162
- 0.007031976222991943,
163
- 0.007060296058654785,
164
- 0.006967016220092773,
165
- 0.007035175800323487,
166
- 0.007229735851287842,
167
- 0.006891496181488037,
168
- 0.007120776176452637,
169
- 0.006821096897125244,
170
- 0.00680109691619873,
171
- 0.006993096828460693,
172
- 0.006769257068634033,
173
- 0.006776937007904053,
174
- 0.006712457180023193,
175
- 0.0067561368942260745,
176
- 0.006787497043609619,
177
- 0.0067476568222045894,
178
- 0.00679725694656372,
179
- 0.006769736766815185,
180
- 0.009892927169799805,
181
- 0.006948135852813721,
182
- 0.006694857120513916,
183
- 0.006692617893218994,
184
- 0.006717096805572509,
185
- 0.007180295944213867,
186
- 0.006696456909179687,
187
- 0.0066926169395446775,
188
- 0.0066881380081176755,
189
- 0.006745256900787353,
190
- 0.006714857101440429,
191
- 0.006697576999664306,
192
- 0.006729258060455322,
193
- 0.006753897190093994,
194
- 0.0066706972122192384,
195
- 0.0067076578140258785,
196
- 0.006674377918243408,
197
- 0.006895176887512207,
198
- 0.006709257125854492,
199
- 0.00669517707824707,
200
- 0.00669501781463623,
201
- 0.006705417156219483,
202
- 0.006737896919250488,
203
- 0.006724297046661377,
204
- 0.006679337978363037,
205
- 0.006716617107391357,
206
- 0.0068081369400024416,
207
- 0.006757896900177002,
208
- 0.006747497081756592,
209
- 0.0066770977973937985,
210
- 0.006703176975250244,
211
- 0.0067422170639038085,
212
- 0.006711658000946045,
213
- 0.00671085786819458,
214
- 0.006744936943054199,
215
- 0.006709736824035644,
216
- 0.006714858055114746,
217
- 0.006684937000274658,
218
- 0.0067250971794128415,
219
- 0.006708776950836182,
220
- 0.006715978145599365,
221
- 0.006759497165679931,
222
- 0.006693256855010986,
223
- 0.00672333812713623,
224
- 0.0067638177871704105,
225
- 0.006752457141876221,
226
- 0.006905255794525146,
227
- 0.006895496845245362,
228
- 0.006755977153778076,
229
- 0.007453414916992187,
230
- 0.007370375156402588,
231
- 0.0074319748878479,
232
- 0.00745325517654419,
233
- 0.007351974964141846,
234
- 0.00759213399887085,
235
- 0.007358534812927246,
236
- 0.007366374969482422,
237
- 0.007424294948577881,
238
- 0.0073498950004577636,
239
- 0.007407495021820068,
240
- 0.007565094947814942,
241
- 0.007518054008483887,
242
- 0.007556934833526611,
243
- 0.007632773876190185,
244
- 0.007522694110870361,
245
- 0.007400294780731201,
246
- 0.007628934860229492,
247
- 0.0075458941459655765,
248
- 0.007518855094909668,
249
- 0.007455813884735108,
250
- 0.0075486149787902836,
251
- 0.007629893779754639,
252
- 0.007533575057983398,
253
- 0.0075394949913024905,
254
- 0.007545254230499268,
255
- 0.007531654834747314,
256
- 0.007485894203186036,
257
- 0.007504455089569092,
258
- 0.00753533411026001,
259
- 0.007542214870452881,
260
- 0.0074618949890136715,
261
- 0.007522853851318359,
262
- 0.007514054775238037,
263
- 0.007495974063873291,
264
- 0.007506855010986328,
265
- 0.007488934993743896,
266
- 0.007518854141235352,
267
- 0.007475494861602783,
268
- 0.0075439739227294925,
269
- 0.007528295040130615,
270
- 0.007529253959655762,
271
- 0.007493895053863525,
272
- 0.007483335018157959,
273
- 0.00765981388092041,
274
- 0.007510694026947021,
275
- 0.007492935180664063,
276
- 0.007513734817504883,
277
- 0.007602054119110108,
278
- 0.007577414035797119,
279
- 0.007540935039520263,
280
- 0.007560773849487305,
281
- 0.007641414165496826,
282
- 0.007522375106811524,
283
- 0.0075343742370605465,
284
- 0.007511654853820801,
285
- 0.007489735126495361,
286
- 0.007486215114593506,
287
- 0.007506054878234863,
288
- 0.007486214160919189,
289
- 0.007466854095458985,
290
- 0.007477254867553711,
291
- 0.007497414112091065,
292
- 0.007495814800262451,
293
- 0.007484934806823731,
294
- 0.007508934020996094,
295
- 0.007478694915771484,
296
- 0.00749469518661499,
297
- 0.007457735061645508
 
 
 
 
 
 
 
 
 
 
 
298
  ]
299
  },
300
  "throughput": {
301
  "unit": "samples/s",
302
- "value": 139.39358726006432
303
  },
304
  "energy": null,
305
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1384.98048,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.83302880859375,
124
+ "mean": 7.83302880859375,
125
  "stdev": 0.0,
126
+ "p50": 7.83302880859375,
127
+ "p90": 7.83302880859375,
128
+ "p95": 7.83302880859375,
129
+ "p99": 7.83302880859375,
130
  "values": [
131
+ 7.83302880859375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1485.099008,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 151,
150
+ "total": 1.003475993156433,
151
+ "mean": 0.006645536378519424,
152
+ "stdev": 0.000400622584209901,
153
+ "p50": 0.006509897232055664,
154
+ "p90": 0.006915656089782715,
155
+ "p95": 0.0072556548118591305,
156
+ "p99": 0.00847132968902588,
157
  "values": [
158
+ 0.007212454795837402,
159
+ 0.006962376117706299,
160
+ 0.006968134880065918,
161
+ 0.0067484560012817385,
162
+ 0.006759975910186768,
163
+ 0.006830056190490722,
164
+ 0.006732295989990234,
165
+ 0.006760937213897705,
166
+ 0.006789896011352539,
167
+ 0.006882534980773926,
168
+ 0.006721736907958985,
169
+ 0.006686056137084961,
170
+ 0.006639017105102539,
171
+ 0.006602376937866211,
172
+ 0.006570857048034668,
173
+ 0.006503017902374267,
174
+ 0.006482216835021972,
175
+ 0.006490056991577149,
176
+ 0.006463976860046386,
177
+ 0.0064882969856262206,
178
+ 0.006539337158203125,
179
+ 0.0065028581619262695,
180
+ 0.006552136898040771,
181
+ 0.006538856983184815,
182
+ 0.006527017116546631,
183
+ 0.0065073370933532715,
184
+ 0.006501737117767334,
185
+ 0.006475176811218262,
186
+ 0.006471497058868408,
187
+ 0.006502857208251953,
188
+ 0.006575497150421142,
189
+ 0.006608457088470459,
190
+ 0.006615817070007324,
191
+ 0.009370686531066895,
192
+ 0.006627816200256348,
193
+ 0.006286218166351319,
194
+ 0.006274538040161133,
195
+ 0.0063089380264282225,
196
+ 0.0062922978401184085,
197
+ 0.006436137199401856,
198
+ 0.006502057075500488,
199
+ 0.00627741813659668,
200
+ 0.006281898021697998,
201
+ 0.00627965784072876,
202
+ 0.006316778182983399,
203
+ 0.006300457954406738,
204
+ 0.0062690978050231935,
205
+ 0.006513737201690674,
206
+ 0.006301417827606202,
207
+ 0.006309416770935059,
208
+ 0.0063046178817749024,
209
+ 0.009516607284545899,
210
+ 0.006798215866088867,
211
+ 0.006741576194763184,
212
+ 0.006639176845550537,
213
+ 0.0067038159370422365,
214
+ 0.006709897041320801,
215
+ 0.006652616024017334,
216
+ 0.0068788561820983885,
217
+ 0.007311173915863037,
218
+ 0.006887015819549561,
219
+ 0.007029735088348389,
220
+ 0.006925896167755127,
221
+ 0.006949416160583496,
222
+ 0.006902375221252441,
223
+ 0.007069735050201416,
224
+ 0.006915656089782715,
225
+ 0.00679421615600586,
226
+ 0.006809415817260742,
227
+ 0.006713897228240967,
228
+ 0.006473416805267334,
229
+ 0.006722856044769287,
230
+ 0.006498857021331787,
231
+ 0.006479177951812744,
232
+ 0.006482697010040283,
233
+ 0.006500297069549561,
234
+ 0.00649165678024292,
235
+ 0.0064719772338867184,
236
+ 0.006478697776794434,
237
+ 0.006500136852264404,
238
+ 0.006506377220153808,
239
+ 0.006475976943969727,
240
+ 0.006493416786193848,
241
+ 0.006520456790924072,
242
+ 0.006511977195739746,
243
+ 0.006487658023834228,
244
+ 0.0064975771903991696,
245
+ 0.006501576900482178,
246
+ 0.00671565580368042,
247
+ 0.006803815841674805,
248
+ 0.006901576042175293,
249
+ 0.00649597692489624,
250
+ 0.006527496814727783,
251
+ 0.006499176979064942,
252
+ 0.0065076580047607425,
253
+ 0.006477578163146973,
254
+ 0.00649597692489624,
255
+ 0.006519337177276611,
256
+ 0.006502696990966797,
257
+ 0.006496936798095703,
258
+ 0.006486056804656982,
259
+ 0.006540616989135742,
260
+ 0.0065169382095336915,
261
+ 0.00653853702545166,
262
+ 0.006499337196350098,
263
+ 0.006522377014160156,
264
+ 0.006491177082061768,
265
+ 0.006535656929016113,
266
+ 0.006501896858215332,
267
+ 0.0065446171760559085,
268
+ 0.006478056907653808,
269
+ 0.0064863767623901366,
270
+ 0.006518377780914307,
271
+ 0.006479337215423584,
272
+ 0.0064834971427917484,
273
+ 0.006486856937408447,
274
+ 0.00649981689453125,
275
+ 0.006499176979064942,
276
+ 0.0064794979095458985,
277
+ 0.006505577087402343,
278
+ 0.006530217170715332,
279
+ 0.006498376846313477,
280
+ 0.006526377201080322,
281
+ 0.006509897232055664,
282
+ 0.006519337177276611,
283
+ 0.006502696990966797,
284
+ 0.006494536876678467,
285
+ 0.0064790182113647465,
286
+ 0.006484457015991211,
287
+ 0.006495817184448242,
288
+ 0.006471497058868408,
289
+ 0.006463337898254395,
290
+ 0.006486536979675293,
291
+ 0.006477736949920655,
292
+ 0.006497257232666016,
293
+ 0.006523656845092774,
294
+ 0.006487337112426758,
295
+ 0.006486537933349609,
296
+ 0.006482217788696289,
297
+ 0.006838215827941895,
298
+ 0.007571972846984863,
299
+ 0.007347333908081055,
300
+ 0.00729885482788086,
301
+ 0.007386213779449463,
302
+ 0.007366213798522949,
303
+ 0.006844295978546143,
304
+ 0.0066842961311340335,
305
+ 0.006809094905853272,
306
+ 0.006867496013641358,
307
+ 0.006690216064453125,
308
+ 0.006806696891784668
309
  ]
310
  },
311
  "throughput": {
312
  "unit": "samples/s",
313
+ "value": 150.47694317532162
314
  },
315
  "energy": null,
316
  "efficiency": null