IlyasMoutawwakil HF staff commited on
Commit
cfc5a2a
·
verified ·
1 Parent(s): c5c0d0d

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
- "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1384.210432,
115
- "max_global_vram": 11.075584,
116
- "max_process_vram": 0.0,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.79958740234375,
124
- "mean": 7.79958740234375,
125
  "stdev": 0.0,
126
- "p50": 7.79958740234375,
127
- "p90": 7.79958740234375,
128
- "p95": 7.79958740234375,
129
- "p99": 7.79958740234375,
130
  "values": [
131
- 7.79958740234375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,182 +138,169 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1484.070912,
142
- "max_global_vram": 11.14112,
143
- "max_process_vram": 0.0,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 154,
150
- "total": 1.0044554195404058,
151
- "mean": 0.0065224377892234104,
152
- "stdev": 0.0003558143102639574,
153
- "p50": 0.00636486005783081,
154
- "p90": 0.0070717532634735104,
155
- "p95": 0.007098450016975403,
156
- "p99": 0.007325057129859924,
157
  "values": [
158
- 0.007096938133239746,
159
- 0.0069058980941772465,
160
- 0.007101257801055908,
161
- 0.007315657138824463,
162
- 0.006985898017883301,
163
- 0.0070196580886840824,
164
- 0.006883657932281494,
165
- 0.0070526180267333986,
166
- 0.006935018062591553,
167
- 0.007335657119750976,
168
- 0.006875018119812012,
169
- 0.006672938823699951,
170
- 0.0066460599899291995,
171
- 0.006621259212493897,
172
- 0.006820619106292725,
173
- 0.006646218776702881,
174
- 0.006772618770599365,
175
- 0.006727179050445556,
176
- 0.006886697769165039,
177
- 0.006878539085388184,
178
- 0.00696957778930664,
179
- 0.006915977954864502,
180
- 0.006851499080657959,
181
- 0.006756938934326172,
182
- 0.006566858768463134,
183
- 0.006538060188293457,
184
- 0.007039177894592285,
185
- 0.008282693862915039,
186
- 0.007239177227020264,
187
- 0.007051177978515625,
188
- 0.007080138206481934,
189
- 0.007087818145751953,
190
- 0.007075976848602295,
191
- 0.007059817790985108,
192
- 0.007079658031463623,
193
- 0.007095657825469971,
194
- 0.0070618982315063476,
195
- 0.007077737808227539,
196
- 0.006786539077758789,
197
- 0.0066297388076782225,
198
- 0.0066866989135742185,
199
- 0.006879017829895019,
200
- 0.006540299892425537,
201
- 0.006563659191131592,
202
- 0.006485579967498779,
203
- 0.006430220127105713,
204
- 0.006364140033721924,
205
- 0.006446859836578369,
206
- 0.006434860229492187,
207
- 0.006216460227966308,
208
- 0.006219820976257324,
209
- 0.006415180206298828,
210
- 0.006192299842834472,
211
- 0.006427340030670166,
212
- 0.006223340988159179,
213
- 0.006230380058288574,
214
- 0.0064737401008605954,
215
- 0.006206861019134521,
216
- 0.0062322998046875,
217
- 0.006359499931335449,
218
- 0.006753098964691162,
219
- 0.006859498977661133,
220
- 0.006768458843231201,
221
- 0.0067031788825988765,
222
- 0.006541099071502686,
223
- 0.006512619972229004,
224
- 0.006665099143981934,
225
- 0.006575499057769775,
226
- 0.006365580081939697,
227
- 0.006544939994812012,
228
- 0.006395979881286621,
229
- 0.006248459815979004,
230
- 0.006261741161346436,
231
- 0.0062670202255249025,
232
- 0.006244941234588623,
233
- 0.006283339977264405,
234
- 0.006276619911193848,
235
- 0.006286221027374268,
236
- 0.006261579990386963,
237
- 0.00624238109588623,
238
- 0.006237740039825439,
239
- 0.006226541042327881,
240
- 0.006215980052947998,
241
- 0.006263181209564209,
242
- 0.0062436599731445315,
243
- 0.006235661029815674,
244
- 0.0062603001594543455,
245
- 0.006223821163177491,
246
- 0.006244140148162842,
247
- 0.006250220775604248,
248
- 0.0062587008476257325,
249
- 0.006270540237426758,
250
- 0.0062417411804199215,
251
- 0.0062510199546813965,
252
- 0.0062519807815551755,
253
- 0.006223820209503174,
254
- 0.006219181060791015,
255
- 0.0062406201362609865,
256
- 0.006221580982208252,
257
- 0.0062222199440002445,
258
- 0.006240460872650146,
259
- 0.006406060218811035,
260
- 0.006245739936828613,
261
- 0.006206380844116211,
262
- 0.006228940010070801,
263
- 0.00625822114944458,
264
- 0.006244140148162842,
265
- 0.006229741096496582,
266
- 0.0062322998046875,
267
- 0.0062558212280273435,
268
- 0.006209740161895752,
269
- 0.006218221187591553,
270
- 0.006227980136871338,
271
- 0.006242540836334229,
272
- 0.0062246198654174805,
273
- 0.00623150110244751,
274
- 0.0062339000701904295,
275
- 0.006212460994720459,
276
- 0.006238379955291748,
277
- 0.006221100807189941,
278
- 0.0062369399070739745,
279
- 0.006222860813140869,
280
- 0.00623006010055542,
281
- 0.006205101013183594,
282
- 0.006185901165008545,
283
- 0.006238540172576904,
284
- 0.006207021236419677,
285
- 0.006212940216064453,
286
- 0.006207981109619141,
287
- 0.006228940010070801,
288
- 0.0062019009590148924,
289
- 0.006243821144104004,
290
- 0.006455018997192383,
291
- 0.006233261108398437,
292
- 0.00627262020111084,
293
- 0.006231980800628662,
294
- 0.006244460105895996,
295
- 0.006239500999450684,
296
- 0.0062583799362182614,
297
- 0.006898859024047852,
298
- 0.007105257034301758,
299
- 0.007076457977294922,
300
- 0.007108617782592774,
301
- 0.0066810989379882816,
302
- 0.00666701889038086,
303
- 0.006845418930053711,
304
- 0.00646205997467041,
305
- 0.006590379238128662,
306
- 0.006552618980407715,
307
- 0.006513579845428467,
308
- 0.0062179012298583984,
309
- 0.0062311792373657224,
310
- 0.006713899135589599,
311
- 0.00711149787902832
312
  ]
313
  },
314
  "throughput": {
315
  "unit": "samples/s",
316
- "value": 153.31690884844213
317
  },
318
  "energy": null,
319
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
11
  "model": "google-bert/bert-base-uncased",
12
  "processor": "google-bert/bert-base-uncased",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1024.741376,
115
+ "max_global_vram": 875.823104,
116
+ "max_process_vram": 44651.073536,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.63286083984375,
124
+ "mean": 7.63286083984375,
125
  "stdev": 0.0,
126
+ "p50": 7.63286083984375,
127
+ "p90": 7.63286083984375,
128
+ "p95": 7.63286083984375,
129
+ "p99": 7.63286083984375,
130
  "values": [
131
+ 7.63286083984375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1155.575808,
142
+ "max_global_vram": 942.518272,
143
+ "max_process_vram": 204460.675072,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 141,
150
+ "total": 1.004502215385437,
151
+ "mean": 0.0071241291871307596,
152
+ "stdev": 0.0004448359471596756,
153
+ "p50": 0.007152297019958496,
154
+ "p90": 0.007479176044464111,
155
+ "p95": 0.00768125581741333,
156
+ "p99": 0.00889325122833252,
157
  "values": [
158
+ 0.0077199749946594235,
159
+ 0.007685255050659179,
160
+ 0.007798854827880859,
161
+ 0.007341416835784912,
162
+ 0.0074494161605834965,
163
+ 0.007484295845031738,
164
+ 0.00768125581741333,
165
+ 0.007479176044464111,
166
+ 0.007806213855743408,
167
+ 0.007375977039337158,
168
+ 0.0075087761878967286,
169
+ 0.007199497222900391,
170
+ 0.007167656898498535,
171
+ 0.007080777168273926,
172
+ 0.007613894939422608,
173
+ 0.0073322968482971195,
174
+ 0.009403169631958008,
175
+ 0.008724291801452637,
176
+ 0.006786698818206787,
177
+ 0.006913416862487793,
178
+ 0.007005097866058349,
179
+ 0.007025897026062012,
180
+ 0.006978218078613281,
181
+ 0.007134377002716064,
182
+ 0.006931337833404541,
183
+ 0.0069084582328796384,
184
+ 0.006986536979675293,
185
+ 0.006847978115081787,
186
+ 0.006731658935546875,
187
+ 0.006804458141326904,
188
+ 0.006831337928771972,
189
+ 0.007094377040863037,
190
+ 0.007051018238067627,
191
+ 0.00679597806930542,
192
+ 0.006839657783508301,
193
+ 0.007008617877960205,
194
+ 0.00675485897064209,
195
+ 0.0068924579620361325,
196
+ 0.006873417854309082,
197
+ 0.006726858139038086,
198
+ 0.006679819107055664,
199
+ 0.006723817825317383,
200
+ 0.00673101806640625,
201
+ 0.006769898891448974,
202
+ 0.0066948580741882325,
203
+ 0.00672429895401001,
204
+ 0.006748777866363525,
205
+ 0.006711018085479736,
206
+ 0.006711019039154053,
207
+ 0.006736777782440186,
208
+ 0.006800938129425049,
209
+ 0.006674698829650879,
210
+ 0.006734538078308106,
211
+ 0.006741418838500977,
212
+ 0.006688777923583984,
213
+ 0.00673085880279541,
214
+ 0.006771498203277588,
215
+ 0.00683085823059082,
216
+ 0.00655229902267456,
217
+ 0.006726697921752929,
218
+ 0.006559819221496582,
219
+ 0.006532138824462891,
220
+ 0.006791818141937256,
221
+ 0.006570219039916992,
222
+ 0.006568939208984375,
223
+ 0.006547658920288086,
224
+ 0.006563498973846436,
225
+ 0.006500298976898193,
226
+ 0.006523338794708252,
227
+ 0.006573578834533691,
228
+ 0.006542698860168457,
229
+ 0.006548939228057861,
230
+ 0.00654573917388916,
231
+ 0.0065302190780639644,
232
+ 0.006743819236755371,
233
+ 0.006719818115234375,
234
+ 0.006747817993164063,
235
+ 0.006746058940887451,
236
+ 0.0066714978218078615,
237
+ 0.006529259204864502,
238
+ 0.007035498142242432,
239
+ 0.007448935985565185,
240
+ 0.00741101598739624,
241
+ 0.00740669584274292,
242
+ 0.007398856163024902,
243
+ 0.007401416778564453,
244
+ 0.007386536121368408,
245
+ 0.007367815971374512,
246
+ 0.007368136882781982,
247
+ 0.007373095989227295,
248
+ 0.0073265361785888675,
249
+ 0.007349417209625244,
250
+ 0.007326375961303711,
251
+ 0.0073182168006896975,
252
+ 0.007346695899963379,
253
+ 0.007349256038665771,
254
+ 0.007332137107849121,
255
+ 0.0073196558952331545,
256
+ 0.007327177047729492,
257
+ 0.007311975955963135,
258
+ 0.007321257114410401,
259
+ 0.007330535888671875,
260
+ 0.00731693696975708,
261
+ 0.007324776172637939,
262
+ 0.007362697124481201,
263
+ 0.007362055778503418,
264
+ 0.007365096092224121,
265
+ 0.007337417125701904,
266
+ 0.007318376064300537,
267
+ 0.007343496799468994,
268
+ 0.0073412561416625974,
269
+ 0.007443016052246094,
270
+ 0.007274537086486816,
271
+ 0.0073718161582946775,
272
+ 0.007348617076873779,
273
+ 0.007307655811309814,
274
+ 0.007429416179656982,
275
+ 0.007340617179870605,
276
+ 0.007382376194000244,
277
+ 0.007372935771942138,
278
+ 0.0073647770881652835,
279
+ 0.007325736045837403,
280
+ 0.007370697021484375,
281
+ 0.007395815849304199,
282
+ 0.007383495807647705,
283
+ 0.007345096111297607,
284
+ 0.007324616909027099,
285
+ 0.007305576801300049,
286
+ 0.007309415817260742,
287
+ 0.009005890846252442,
288
+ 0.007651014804840088,
289
+ 0.007639976024627686,
290
+ 0.007601095199584961,
291
+ 0.0071002979278564455,
292
+ 0.006999016761779785,
293
+ 0.006980298042297363,
294
+ 0.007152297019958496,
295
+ 0.007260137081146241,
296
+ 0.007037257194519043,
297
+ 0.006940137863159179,
298
+ 0.006793257236480713
 
 
 
 
 
 
 
 
 
 
 
 
 
299
  ]
300
  },
301
  "throughput": {
302
  "unit": "samples/s",
303
+ "value": 140.36803288273182
304
  },
305
  "energy": null,
306
  "efficiency": null