IlyasMoutawwakil HF staff commited on
Commit
f4a0927
·
verified ·
1 Parent(s): 6ae0414

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -65,7 +65,7 @@
65
  "name": "process",
66
  "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
67
  "device_isolation": true,
68
- "device_isolation_action": "error",
69
  "numactl": false,
70
  "numactl_kwargs": {},
71
  "start_method": "spawn"
@@ -95,7 +95,7 @@
95
  "optimum_benchmark_commit": null,
96
  "transformers_version": "4.44.2",
97
  "transformers_commit": null,
98
- "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
  "diffusers_version": "0.30.2",
101
  "diffusers_commit": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1386.51648,
115
- "max_global_vram": 11.235328,
116
- "max_process_vram": 0.0,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.8885771484375,
124
- "mean": 7.8885771484375,
125
  "stdev": 0.0,
126
- "p50": 7.8885771484375,
127
- "p90": 7.8885771484375,
128
- "p95": 7.8885771484375,
129
- "p99": 7.8885771484375,
130
  "values": [
131
- 7.8885771484375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,177 +138,167 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1485.96736,
142
- "max_global_vram": 11.247616,
143
- "max_process_vram": 0.0,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 149,
150
- "total": 0.9987602725028993,
151
- "mean": 0.006703089077200667,
152
- "stdev": 0.0003145295402813907,
153
- "p50": 0.006644295215606689,
154
- "p90": 0.007036581134796143,
155
- "p95": 0.007236324882507325,
156
- "p99": 0.007276472263336182,
157
  "values": [
158
- 0.006998854160308838,
159
- 0.0069727740287780764,
160
- 0.007001574039459228,
161
- 0.006973574161529541,
162
- 0.006987174034118652,
163
- 0.00693869400024414,
164
- 0.006880613803863525,
165
- 0.0070724539756774904,
166
- 0.006872934818267822,
167
- 0.0071001329421997075,
168
- 0.00682557487487793,
169
- 0.0067758140563964845,
170
- 0.0067767748832702635,
171
- 0.006691975116729736,
172
- 0.006883173942565918,
173
- 0.006639975070953369,
174
- 0.006794054985046386,
175
- 0.006562856197357178,
176
- 0.006897573947906494,
177
- 0.006914534091949463,
178
- 0.007014214038848877,
179
- 0.007023973941802979,
180
- 0.0065559749603271485,
181
- 0.0065209360122680665,
182
- 0.006514855861663818,
183
- 0.006511175155639649,
184
- 0.006467336177825927,
185
- 0.009035486221313477,
186
- 0.007230052947998047,
187
- 0.00664269495010376,
188
- 0.006535816192626953,
189
- 0.0065492548942565915,
190
- 0.0066207761764526364,
191
- 0.0065609350204467775,
192
- 0.006518856048583984,
193
- 0.006599015235900879,
194
- 0.006618696212768555,
195
- 0.006545895099639892,
196
- 0.006614695072174073,
197
- 0.006706375122070313,
198
- 0.006543816089630127,
199
- 0.006672134876251221,
200
- 0.006832934856414795,
201
- 0.006584935188293457,
202
- 0.006652935028076172,
203
- 0.006494376182556152,
204
- 0.006575335025787354,
205
- 0.006523816108703613,
206
- 0.006600454807281494,
207
- 0.006905095100402832,
208
- 0.006514695167541504,
209
- 0.006564775943756104,
210
- 0.006583815097808838,
211
- 0.006554215908050537,
212
- 0.0066361351013183596,
213
- 0.006644774913787841,
214
- 0.006593095779418945,
215
- 0.006607655048370361,
216
- 0.006609254837036132,
217
- 0.006567016124725342,
218
- 0.0065206151008605955,
219
- 0.006661416053771973,
220
- 0.006648775100708008,
221
- 0.006412775993347168,
222
- 0.006429895877838134,
223
- 0.006429736137390137,
224
- 0.006427175998687744,
225
- 0.006397575855255127,
226
- 0.006380616188049316,
227
- 0.006383016109466553,
228
- 0.00640013599395752,
229
- 0.00639821720123291,
230
- 0.006375175952911377,
231
- 0.006375016212463379,
232
- 0.0064218959808349605,
233
- 0.006417096138000488,
234
- 0.006650694847106934,
235
- 0.006448616027832031,
236
- 0.006394855976104737,
237
- 0.006433256149291992,
238
- 0.006401735782623291,
239
- 0.006388776779174805,
240
- 0.006393735885620117,
241
- 0.0063790159225463865,
242
- 0.006383175849914551,
243
- 0.0063503761291503905,
244
- 0.006364137172698975,
245
- 0.0063790159225463865,
246
- 0.006391975879669189,
247
- 0.006374375820159912,
248
- 0.00638045597076416,
249
- 0.006346696853637695,
250
- 0.006383175849914551,
251
- 0.006396776199340821,
252
- 0.0063913359642028805,
253
- 0.006644295215606689,
254
- 0.00635677719116211,
255
- 0.0063698959350585935,
256
- 0.006358535766601563,
257
- 0.006423175811767578,
258
- 0.006399975776672363,
259
- 0.006644454956054688,
260
- 0.006975654125213623,
261
- 0.006655014991760254,
262
- 0.006683656215667725,
263
- 0.0068326139450073245,
264
- 0.0065686159133911135,
265
- 0.007029733180999756,
266
- 0.00676445484161377,
267
- 0.006608934879302979,
268
- 0.006662055969238281,
269
- 0.006823174953460694,
270
- 0.006644295215606689,
271
- 0.006633574962615967,
272
- 0.006744295120239258,
273
- 0.006603975772857666,
274
- 0.006806694030761719,
275
- 0.006797414779663086,
276
- 0.0068687739372253415,
277
- 0.0068287749290466306,
278
- 0.00670653486251831,
279
- 0.006804133892059326,
280
- 0.006781414985656738,
281
- 0.0067910151481628414,
282
- 0.0067977337837219235,
283
- 0.006789255142211914,
284
- 0.00677309513092041,
285
- 0.0068822140693664555,
286
- 0.006914373874664307,
287
- 0.006874855041503906,
288
- 0.006893094062805176,
289
- 0.006842854022979736,
290
- 0.006596615791320801,
291
- 0.006759014129638672,
292
- 0.006862054824829102,
293
- 0.007066853046417236,
294
- 0.0070002942085266116,
295
- 0.006854694843292236,
296
- 0.007284613132476807,
297
- 0.0072513318061828615,
298
- 0.0072609329223632816,
299
- 0.007238372802734375,
300
- 0.007267652988433838,
301
- 0.007266852855682373,
302
- 0.007233253002166748,
303
- 0.0072591729164123535,
304
- 0.006915654182434082,
305
- 0.007078853130340576,
306
- 0.00706397294998169
307
  ]
308
  },
309
  "throughput": {
310
  "unit": "samples/s",
311
- "value": 149.1849486830359
312
  },
313
  "energy": null,
314
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
65
  "name": "process",
66
  "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
67
  "device_isolation": true,
68
+ "device_isolation_action": "warn",
69
  "numactl": false,
70
  "numactl_kwargs": {},
71
  "start_method": "spawn"
 
95
  "optimum_benchmark_commit": null,
96
  "transformers_version": "4.44.2",
97
  "transformers_commit": null,
98
+ "accelerate_version": "0.34.0",
99
  "accelerate_commit": null,
100
  "diffusers_version": "0.30.2",
101
  "diffusers_commit": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1023.737856,
115
+ "max_global_vram": 68702.69952,
116
+ "max_process_vram": 43568.922624,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.5401123046875,
124
+ "mean": 7.5401123046875,
125
  "stdev": 0.0,
126
+ "p50": 7.5401123046875,
127
+ "p90": 7.5401123046875,
128
+ "p95": 7.5401123046875,
129
+ "p99": 7.5401123046875,
130
  "values": [
131
+ 7.5401123046875
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1155.035136,
142
+ "max_global_vram": 68702.69952,
143
+ "max_process_vram": 227941.052416,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 139,
150
+ "total": 1.0039940123558044,
151
+ "mean": 0.007222978506156868,
152
+ "stdev": 0.0003379172634522948,
153
+ "p50": 0.007181899070739746,
154
+ "p90": 0.007660424995422363,
155
+ "p95": 0.007683721017837525,
156
+ "p99": 0.007738981838226319,
157
  "values": [
158
+ 0.007669258117675781,
159
+ 0.007669897079467774,
160
+ 0.007704617977142334,
161
+ 0.007664776802062988,
162
+ 0.0075670180320739746,
163
+ 0.007596618175506592,
164
+ 0.007520617008209229,
165
+ 0.0074090991020202635,
166
+ 0.007344297885894776,
167
+ 0.007235498905181885,
168
+ 0.00714509916305542,
169
+ 0.007137098789215088,
170
+ 0.00723677921295166,
171
+ 0.00712637996673584,
172
+ 0.007197897911071777,
173
+ 0.007127179145812988,
174
+ 0.007120460033416748,
175
+ 0.007182538986206055,
176
+ 0.007181899070739746,
177
+ 0.007010059833526611,
178
+ 0.007105578899383545,
179
+ 0.007204459190368652,
180
+ 0.007154539108276368,
181
+ 0.006964299201965332,
182
+ 0.006890860080718994,
183
+ 0.0069463791847229,
184
+ 0.007080619812011719,
185
+ 0.00691581916809082,
186
+ 0.007036139965057373,
187
+ 0.007087179183959961,
188
+ 0.006953098773956299,
189
+ 0.009672612190246581,
190
+ 0.007158699035644531,
191
+ 0.006941740036010742,
192
+ 0.007252939224243164,
193
+ 0.007101259231567383,
194
+ 0.007257418155670166,
195
+ 0.007106860160827636,
196
+ 0.006873579025268555,
197
+ 0.006887179851531982,
198
+ 0.006915180206298828,
199
+ 0.00690765905380249,
200
+ 0.0069281401634216306,
201
+ 0.006935019969940185,
202
+ 0.00704413890838623,
203
+ 0.0069145398139953615,
204
+ 0.006924458980560303,
205
+ 0.006809899806976319,
206
+ 0.006843180179595947,
207
+ 0.0072028589248657226,
208
+ 0.006865419864654541,
209
+ 0.007497737884521485,
210
+ 0.007324617862701416,
211
+ 0.00726301908493042,
212
+ 0.007659337043762207,
213
+ 0.007230538845062256,
214
+ 0.007419338226318359,
215
+ 0.007324298858642578,
216
+ 0.007360298156738282,
217
+ 0.007587337970733642,
218
+ 0.0076143770217895505,
219
+ 0.007311338901519775,
220
+ 0.007264298915863037,
221
+ 0.0073498978614807125,
222
+ 0.007154378890991211,
223
+ 0.0072967782020568845,
224
+ 0.007301578998565674,
225
+ 0.007119658946990967,
226
+ 0.007189259052276611,
227
+ 0.007132618904113769,
228
+ 0.007245899200439453,
229
+ 0.007270058155059814,
230
+ 0.007166059017181397,
231
+ 0.007137898921966553,
232
+ 0.007114378929138183,
233
+ 0.007336459159851074,
234
+ 0.007478377819061279,
235
+ 0.0071906991004943846,
236
+ 0.0075039777755737305,
237
+ 0.007574057102203369,
238
+ 0.007471337795257569,
239
+ 0.007297739028930664,
240
+ 0.007235339164733887,
241
+ 0.0072617378234863285,
242
+ 0.007303339004516601,
243
+ 0.0072145390510559085,
244
+ 0.007077259063720703,
245
+ 0.0072764577865600585,
246
+ 0.007104619026184082,
247
+ 0.00714125919342041,
248
+ 0.006972620010375977,
249
+ 0.006975819110870361,
250
+ 0.006834060192108154,
251
+ 0.006837739944458008,
252
+ 0.00692110013961792,
253
+ 0.007019818782806397,
254
+ 0.006917419910430908,
255
+ 0.006896938800811768,
256
+ 0.006976940155029297,
257
+ 0.0071594991683959965,
258
+ 0.00698301887512207,
259
+ 0.006925899982452393,
260
+ 0.007123498916625976,
261
+ 0.0069582200050354,
262
+ 0.006974699020385743,
263
+ 0.006746059894561768,
264
+ 0.006775979995727539,
265
+ 0.006753739833831787,
266
+ 0.006679340839385987,
267
+ 0.00671422004699707,
268
+ 0.006764299869537353,
269
+ 0.0067292599678039555,
270
+ 0.007561898231506348,
271
+ 0.007669417858123779,
272
+ 0.007683176994323731,
273
+ 0.00768861722946167,
274
+ 0.00723469877243042,
275
+ 0.0069964599609375,
276
+ 0.007104299068450928,
277
+ 0.007474857807159424,
278
+ 0.007729257106781006,
279
+ 0.007732137203216553,
280
+ 0.007437098026275635,
281
+ 0.007461098194122315,
282
+ 0.007634218215942383,
283
+ 0.0076785368919372555,
284
+ 0.007672458171844482,
285
+ 0.007465258121490479,
286
+ 0.0071172590255737305,
287
+ 0.007136458873748779,
288
+ 0.0070958189964294436,
289
+ 0.007323819160461426,
290
+ 0.007743176937103272,
291
+ 0.0077086172103881835,
292
+ 0.00753341817855835,
293
+ 0.0072505388259887695,
294
+ 0.007193579196929932,
295
+ 0.007321257114410401,
296
+ 0.007175339221954346
 
 
 
 
 
 
 
 
 
 
297
  ]
298
  },
299
  "throughput": {
300
  "unit": "samples/s",
301
+ "value": 138.4470408083867
302
  },
303
  "energy": null,
304
  "efficiency": null