IlyasMoutawwakil HF staff commited on
Commit
c1ba108
·
verified ·
1 Parent(s): e4b770d

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.0.dev20231010+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -103,7 +103,7 @@
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
- "peft_version": "0.12.0",
107
  "peft_commit": null
108
  }
109
  },
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 947.806208,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 47198.834688,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 8.398337890625,
124
- "mean": 8.398337890625,
125
  "stdev": 0.0,
126
- "p50": 8.398337890625,
127
- "p90": 8.398337890625,
128
- "p95": 8.398337890625,
129
- "p99": 8.398337890625,
130
  "values": [
131
- 8.398337890625
132
  ]
133
  },
134
  "throughput": null,
@@ -138,169 +138,172 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1085.382656,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 240939.261952,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 141,
150
- "total": 0.9994821271896365,
151
- "mean": 0.007088525724749192,
152
- "stdev": 0.0003720870383416912,
153
- "p50": 0.0069707350730896,
154
- "p90": 0.007542575836181641,
155
- "p95": 0.007559855937957763,
156
- "p99": 0.007963856506347653,
157
  "values": [
158
- 0.007580496788024902,
159
- 0.007426255226135254,
160
- 0.007385775089263916,
161
- 0.007386735916137695,
162
- 0.007345776081085205,
163
- 0.007220654964447021,
164
- 0.0074168148040771485,
165
- 0.007381455898284912,
166
- 0.007277455806732178,
167
- 0.0072174549102783205,
168
- 0.00854433822631836,
169
- 0.006808653831481933,
170
- 0.006817295074462891,
171
- 0.006813293933868409,
172
- 0.006805775165557862,
173
- 0.006773454189300537,
174
- 0.0067380938529968265,
175
- 0.006756655216217041,
176
- 0.006744174003601075,
177
- 0.006724333763122559,
178
- 0.006773294925689698,
179
- 0.0068067340850830075,
180
- 0.006747694969177246,
181
- 0.006817614078521729,
182
- 0.006928335189819336,
183
- 0.006733933925628662,
184
- 0.006710734844207764,
185
- 0.006734253883361817,
186
- 0.006724813938140869,
187
- 0.006888655185699463,
188
- 0.0068523349761962895,
189
- 0.008205456733703613,
190
- 0.00741153621673584,
191
- 0.0071032152175903324,
192
- 0.007064334869384766,
193
- 0.006998734951019287,
194
- 0.007074254989624024,
195
- 0.007027695178985596,
196
- 0.007525455951690674,
197
- 0.007501455783843994,
198
- 0.007525297164916992,
199
- 0.007501454830169678,
200
- 0.0075256161689758305,
201
- 0.0075236959457397464,
202
- 0.007484655857086181,
203
- 0.007559855937957763,
204
- 0.007440334796905518,
205
- 0.007552815914154053,
206
- 0.007530896186828613,
207
- 0.007528336048126221,
208
- 0.007522096157073974,
209
- 0.007520495891571045,
210
- 0.0075726561546325686,
211
- 0.007540976047515869,
212
- 0.007564815998077392,
213
- 0.007545936107635498,
214
- 0.007489295959472656,
215
- 0.007539055824279785,
216
- 0.007542575836181641,
217
- 0.007516335964202881,
218
- 0.00746449613571167,
219
- 0.006938254833221436,
220
- 0.0069019351005554196,
221
- 0.006878893852233886,
222
- 0.007049934864044189,
223
- 0.006868655204772949,
224
- 0.006844814777374268,
225
- 0.006681774139404297,
226
- 0.006718733787536621,
227
- 0.0069470548629760745,
228
- 0.007132335186004639,
229
- 0.007064334869384766,
230
- 0.00709521484375,
231
- 0.006900495052337647,
232
- 0.00713121509552002,
233
- 0.0069238548278808595,
234
- 0.0071539349555969235,
235
- 0.006864174842834473,
236
- 0.0070395350456237795,
237
- 0.00698977518081665,
238
- 0.006892014980316162,
239
- 0.006887854099273682,
240
- 0.0069966549873352055,
241
- 0.007114254951477051,
242
- 0.0069238548278808595,
243
- 0.006695854187011719,
244
- 0.006671854972839356,
245
- 0.006675374031066894,
246
- 0.006654573917388916,
247
- 0.006658894062042236,
248
- 0.006650253772735596,
249
- 0.00667265510559082,
250
- 0.006614414215087891,
251
- 0.006645133972167969,
252
- 0.006736494064331054,
253
- 0.006665453910827636,
254
  0.006833614826202393,
255
- 0.006670094013214111,
256
- 0.006856494903564453,
257
- 0.006876015186309815,
258
- 0.0066316938400268555,
259
- 0.006710574150085449,
260
- 0.0067976140975952145,
261
- 0.006995695114135742,
262
- 0.006803215026855469,
263
- 0.006844813823699951,
264
- 0.006677134990692139,
265
- 0.006642094135284424,
266
- 0.006647054195404053,
267
- 0.006644974231719971,
268
- 0.006638254165649414,
269
- 0.006665935039520263,
270
- 0.006651214122772217,
271
- 0.0066979341506958005,
272
- 0.006727853775024414,
273
- 0.006825934886932373,
274
- 0.007485936164855957,
275
- 0.007550096035003662,
276
- 0.007572336196899414,
277
- 0.0075108962059021,
278
- 0.007552336215972901,
279
- 0.0076014561653137205,
280
- 0.007558416843414306,
281
- 0.007275215148925781,
282
- 0.007152975082397461,
283
- 0.007444176197052002,
284
- 0.007531216144561767,
285
- 0.007547376155853272,
286
- 0.007521135807037354,
287
- 0.007534416198730469,
288
- 0.00748689603805542,
289
- 0.00753601598739624,
290
- 0.00750881576538086,
291
- 0.007334896087646484,
292
- 0.006935375213623047,
293
- 0.0069707350730896,
294
- 0.00682305383682251,
295
- 0.006924174785614014,
296
- 0.006804654121398926,
297
- 0.006801774978637695,
298
- 0.0066118531227111815
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
299
  ]
300
  },
301
  "throughput": {
302
  "unit": "samples/s",
303
- "value": 141.0730579009618
304
  },
305
  "energy": null,
306
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
+ "peft_version": null,
107
  "peft_commit": null
108
  }
109
  },
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1023.700992,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 41995.005952,
117
  "max_reserved": 589.299712,
118
  "max_allocated": 533.571072
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.893673828125,
124
+ "mean": 7.893673828125,
125
  "stdev": 0.0,
126
+ "p50": 7.893673828125,
127
+ "p90": 7.893673828125,
128
+ "p95": 7.893673828125,
129
+ "p99": 7.893673828125,
130
  "values": [
131
+ 7.893673828125
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1155.01056,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 216879.77984,
144
  "max_reserved": 589.299712,
145
  "max_allocated": 439.456768
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 144,
150
+ "total": 1.0046126728057863,
151
+ "mean": 0.006976476894484625,
152
+ "stdev": 0.00037703788937979154,
153
+ "p50": 0.00683593487739563,
154
+ "p90": 0.007531345081329348,
155
+ "p95": 0.007593857097625733,
156
+ "p99": 0.007656290588378906,
157
  "values": [
158
+ 0.007316176891326904,
159
+ 0.007202256202697754,
160
+ 0.007275536060333252,
161
+ 0.007153776168823242,
162
+ 0.007083055973052979,
163
+ 0.007108176231384277,
164
+ 0.007019375801086426,
165
+ 0.007064815998077393,
166
+ 0.0072115368843078615,
167
+ 0.0069563350677490235,
168
+ 0.0069958558082580564,
169
+ 0.0069169759750366214,
170
+ 0.006953135967254638,
171
+ 0.0069086551666259766,
172
+ 0.006887856006622314,
173
+ 0.006907855987548828,
174
+ 0.006860815048217774,
175
+ 0.00683841609954834,
176
+ 0.006872654914855957,
177
+ 0.006903215885162354,
178
+ 0.006917935848236084,
179
+ 0.006861295223236084,
180
+ 0.006861936092376709,
181
+ 0.0068145751953125,
182
+ 0.006830736160278321,
183
+ 0.006844015121459961,
184
+ 0.006860496044158936,
185
+ 0.006860654830932617,
186
+ 0.0068532958030700685,
187
+ 0.006969614982604981,
188
+ 0.006871055126190186,
189
+ 0.006851376056671143,
190
+ 0.0068382549285888675,
191
+ 0.006854896068572998,
192
+ 0.006878735065460205,
193
+ 0.006845615863800049,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194
  0.006833614826202393,
195
+ 0.006809775829315186,
196
+ 0.006803695201873779,
197
+ 0.010084182739257812,
198
+ 0.0070192160606384274,
199
+ 0.006704814910888672,
200
+ 0.00679201602935791,
201
+ 0.00672289514541626,
202
+ 0.0067556948661804195,
203
+ 0.006726095199584961,
204
+ 0.006729935169219971,
205
+ 0.00674529504776001,
206
+ 0.006762734889984131,
207
+ 0.006777935981750488,
208
+ 0.006756494998931885,
209
+ 0.006779055118560791,
210
+ 0.006750095844268799,
211
+ 0.006971694946289062,
212
+ 0.0067652959823608395,
213
+ 0.006746735095977783,
214
+ 0.0069641761779785155,
215
+ 0.006991695880889892,
216
+ 0.006754254817962647,
217
+ 0.006754895210266113,
218
+ 0.006799536228179932,
219
+ 0.006782575130462647,
220
+ 0.007304176807403564,
221
+ 0.007293935775756836,
222
+ 0.007304176807403564,
223
+ 0.007239535808563232,
224
+ 0.007193295955657959,
225
+ 0.007151377201080322,
226
+ 0.006904494762420654,
227
+ 0.006793136119842529,
228
+ 0.0067784152030944825,
229
+ 0.006789454936981201,
230
+ 0.0069921760559082035,
231
+ 0.006796976089477539,
232
+ 0.006777774810791015,
233
+ 0.006746574878692627,
234
+ 0.006743215084075928,
235
+ 0.006716015815734863,
236
+ 0.00682929515838623,
237
+ 0.0067897758483886714,
238
+ 0.0068192148208618165,
239
+ 0.006791055202484131,
240
+ 0.006802415847778321,
241
+ 0.00680257511138916,
242
+ 0.006760176181793213,
243
+ 0.006746254920959473,
244
+ 0.0067814550399780274,
245
+ 0.006764494895935059,
246
+ 0.006754096031188965,
247
+ 0.006761294841766357,
248
+ 0.006789134979248047,
249
+ 0.006733136177062988,
250
+ 0.00673057508468628,
251
+ 0.006738414764404297,
252
+ 0.0067817759513854985,
253
+ 0.0067177748680114745,
254
+ 0.006741934776306152,
255
+ 0.006746416091918945,
256
+ 0.006738094806671142,
257
+ 0.006732174873352051,
258
+ 0.006751054763793945,
259
+ 0.006783696174621582,
260
+ 0.006763054847717285,
261
+ 0.006771535873413086,
262
+ 0.006765134811401367,
263
+ 0.006784655094146728,
264
+ 0.006738255023956299,
265
+ 0.0067683358192443845,
266
+ 0.006737935066223145,
267
+ 0.006746894836425781,
268
+ 0.007261456966400147,
269
+ 0.007670257091522217,
270
+ 0.007570736885070801,
271
+ 0.0076022582054138185,
272
+ 0.007564657211303711,
273
+ 0.007637776851654053,
274
+ 0.007632017135620117,
275
+ 0.007406897068023681,
276
+ 0.007026895999908447,
277
+ 0.007293937206268311,
278
+ 0.007580976963043213,
279
+ 0.007580817222595215,
280
+ 0.007594097137451172,
281
+ 0.007587856769561767,
282
+ 0.007592496871948243,
283
+ 0.007618897914886474,
284
+ 0.007558897018432618,
285
+ 0.007635056972503662,
286
+ 0.007467057228088379,
287
+ 0.007259376049041748,
288
+ 0.0073841772079467774,
289
+ 0.007161136150360107,
290
+ 0.007034095764160156,
291
+ 0.006762415885925293,
292
+ 0.007004815101623535,
293
+ 0.006772335052490234,
294
+ 0.006844015121459961,
295
+ 0.006697775840759278,
296
+ 0.006792815208435059,
297
+ 0.006760175228118896,
298
+ 0.0067321758270263675,
299
+ 0.006702895164489746,
300
+ 0.006763054847717285,
301
+ 0.0067361750602722165
302
  ]
303
  },
304
  "throughput": {
305
  "unit": "samples/s",
306
+ "value": 143.33882490036873
307
  },
308
  "energy": null,
309
  "efficiency": null