IlyasMoutawwakil HF staff commited on
Commit
c4dc56a
·
verified ·
1 Parent(s): cec80fc

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -104,7 +104,7 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 720.334848,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
@@ -113,31 +113,31 @@
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.2133818359375,
117
- "mean": 7.2133818359375,
118
  "stdev": 0.0,
119
- "p50": 7.2133818359375,
120
- "p90": 7.2133818359375,
121
- "p95": 7.2133818359375,
122
- "p99": 7.2133818359375,
123
  "values": [
124
- 7.2133818359375
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 8.947922020834734e-07,
131
- "ram": 4.7455879608104613e-07,
132
- "gpu": 1.736668056000288e-06,
133
- "total": 3.1060190541648075e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 919.494656,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
@@ -145,167 +145,168 @@
145
  },
146
  "latency": {
147
  "unit": "s",
148
- "count": 136,
149
- "total": 0.9969636526107782,
150
- "mean": 0.007330615092726315,
151
- "stdev": 0.0003954046872519939,
152
- "p50": 0.007370752096176148,
153
- "p90": 0.007670783996582031,
154
- "p95": 0.00777187991142273,
155
- "p99": 0.008884480094909669,
156
  "values": [
157
- 0.007670783996582031,
158
- 0.00759500789642334,
159
- 0.007511040210723877,
160
- 0.007554048061370849,
161
- 0.007502848148345947,
162
- 0.007472127914428711,
163
- 0.007444479942321777,
164
- 0.007786496162414551,
165
- 0.0075939841270446775,
166
- 0.007831552028656007,
167
- 0.008451071739196778,
168
- 0.0073175039291381834,
169
- 0.00728985595703125,
170
- 0.007300159931182861,
171
- 0.007354368209838867,
172
- 0.007411712169647216,
173
- 0.007388160228729248,
174
- 0.0073768959045410155,
175
- 0.0073400321006774905,
176
- 0.007357439994812012,
177
- 0.007379968166351319,
178
- 0.007404543876647949,
179
- 0.007373824119567871,
180
- 0.0073697280883789065,
181
  0.007390207767486572,
 
 
 
 
 
182
  0.0073431038856506346,
183
- 0.007366655826568603,
184
- 0.007364607810974121,
185
- 0.007591936111450195,
186
- 0.007573503971099854,
187
- 0.007508992195129394,
188
- 0.007517183780670166,
189
- 0.007541759967803955,
190
- 0.007465983867645264,
191
- 0.007420928001403809,
192
- 0.007536640167236328,
193
- 0.007566336154937744,
194
- 0.007670783996582031,
195
- 0.0076277761459350585,
196
- 0.007683072090148926,
197
- 0.007634943962097168,
198
- 0.007546879768371582,
199
- 0.0075673599243164065,
200
- 0.007655424118041992,
201
- 0.00745472002029419,
202
- 0.007574528217315674,
203
- 0.007516160011291504,
204
- 0.007542784214019775,
205
- 0.007482367992401123,
206
- 0.007511072158813477,
207
- 0.007579648017883301,
208
- 0.00769536018371582,
209
- 0.007547904014587403,
210
- 0.007620672225952149,
211
- 0.00738099193572998,
212
- 0.0075632638931274416,
213
- 0.007526400089263916,
214
- 0.007582719802856445,
215
- 0.00764415979385376,
216
- 0.007767007827758789,
217
- 0.007686143875122071,
218
- 0.007667712211608887,
219
- 0.0076574721336364745,
220
- 0.007620607852935791,
221
- 0.007484447956085205,
222
- 0.007527423858642578,
223
- 0.007935999870300293,
224
- 0.007559167861938477,
225
- 0.007515135765075683,
226
- 0.007371776103973389,
227
- 0.007672832012176513,
228
- 0.007535615921020508,
229
- 0.007388160228729248,
230
- 0.007178239822387696,
231
  0.007229440212249756,
232
- 0.007247871875762939,
233
- 0.007716864109039307,
234
- 0.008867839813232421,
235
- 0.008893440246582032,
236
- 0.008945664405822755,
237
- 0.007661568164825439,
238
- 0.007324672222137451,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
239
  0.007250944137573242,
240
- 0.007252992153167725,
241
- 0.0071526398658752445,
242
- 0.007164927959442138,
243
- 0.007147520065307617,
244
- 0.006895616054534912,
245
- 0.006883327960968018,
246
- 0.007001088142395019,
247
- 0.006971392154693603,
248
- 0.00688640022277832,
249
- 0.006880256175994873,
250
- 0.006896704196929932,
251
- 0.006929408073425293,
252
- 0.0068884482383728025,
253
- 0.0069253120422363285,
254
- 0.006874112129211426,
255
- 0.006897664070129395,
256
- 0.006867968082427979,
257
- 0.006880288124084473,
258
- 0.00687718391418457,
259
- 0.006884352207183838,
260
- 0.006848512172698974,
261
- 0.006870016098022461,
262
- 0.006903808116912841,
263
- 0.006887423992156983,
264
- 0.006929408073425293,
265
- 0.006927360057830811,
266
- 0.0068915200233459475,
267
- 0.006879231929779053,
268
- 0.0069212160110473635,
269
- 0.00690176010131836,
270
- 0.006906879901885986,
271
- 0.006858751773834228,
272
- 0.006915071964263916,
273
- 0.007136256217956543,
274
- 0.007280640125274658,
275
- 0.007120895862579346,
276
- 0.007170048236846924,
277
- 0.007143424034118652,
278
- 0.007136256217956543,
279
- 0.006949952125549316,
280
- 0.006912000179290771,
281
- 0.006915071964263916,
282
- 0.006876160144805908,
283
- 0.006874112129211426,
284
- 0.006896639823913574,
285
- 0.0068884482383728025,
286
- 0.0068853759765625,
287
- 0.006876160144805908,
288
- 0.006915071964263916,
289
- 0.006890495777130127,
290
- 0.006922239780426025,
291
- 0.006882304191589355,
292
- 0.006913023948669434
293
  ]
294
  },
295
  "throughput": {
296
  "unit": "samples/s",
297
- "value": 136.4142009027638
298
  },
299
  "energy": {
300
  "unit": "kWh",
301
- "cpu": 8.22974933908048e-08,
302
- "ram": 4.4910364295373584e-08,
303
- "gpu": 1.6052694834482573e-07,
304
- "total": 2.877348060310041e-07
305
  },
306
  "efficiency": {
307
  "unit": "samples/kWh",
308
- "value": 3475422.4342683367
309
  }
310
  }
311
  }
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 764.125184,
108
  "max_global_vram": 1185.415168,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
 
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.39188525390625,
117
+ "mean": 7.39188525390625,
118
  "stdev": 0.0,
119
+ "p50": 7.39188525390625,
120
+ "p90": 7.39188525390625,
121
+ "p95": 7.39188525390625,
122
+ "p99": 7.39188525390625,
123
  "values": [
124
+ 7.39188525390625
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 8.811242493057216e-07,
131
+ "ram": 4.6715578417844635e-07,
132
+ "gpu": 1.5844457120000807e-06,
133
+ "total": 2.9327257454842483e-06
134
  },
135
  "efficiency": null
136
  },
137
  "forward": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 964.46464,
141
  "max_global_vram": 1195.900928,
142
  "max_process_vram": 0.0,
143
  "max_reserved": 555.74528,
 
145
  },
146
  "latency": {
147
  "unit": "s",
148
+ "count": 137,
149
+ "total": 0.9961949753761289,
150
+ "mean": 0.007271496170628679,
151
+ "stdev": 0.00012064961571531031,
152
+ "p50": 0.007250944137573242,
153
+ "p90": 0.007389593505859375,
154
+ "p95": 0.007487488174438476,
155
+ "p99": 0.007675002841949463,
156
  "values": [
157
+ 0.00753766393661499,
158
+ 0.007481344223022461,
159
+ 0.00744755220413208,
160
+ 0.007316480159759522,
161
+ 0.007377920150756836,
162
+ 0.007314432144165039,
163
+ 0.007323647975921631,
164
+ 0.0072837119102478025,
165
+ 0.0073134078979492185,
166
+ 0.007286784172058106,
167
+ 0.007307263851165771,
168
+ 0.007312384128570557,
169
+ 0.007268352031707764,
170
+ 0.007208960056304932,
171
+ 0.007274496078491211,
172
+ 0.007293951988220215,
173
+ 0.007314367771148682,
174
+ 0.00734822416305542,
 
 
 
 
 
 
175
  0.007390207767486572,
176
+ 0.007356416225433349,
177
+ 0.007323647975921631,
178
+ 0.0072325119972229,
179
+ 0.007305215835571289,
180
+ 0.0073696961402893064,
181
  0.0073431038856506346,
182
+ 0.007337984085083008,
183
+ 0.007365632057189942,
184
+ 0.007279615879058838,
185
+ 0.007331840038299561,
186
+ 0.007289792060852051,
187
+ 0.007308288097381592,
188
+ 0.007327744007110596,
189
+ 0.007222271919250488,
190
+ 0.007156735897064209,
191
+ 0.0072304000854492186,
192
+ 0.007237631797790528,
193
+ 0.007236608028411865,
194
+ 0.007271423816680909,
195
+ 0.007258111953735351,
196
+ 0.00728166389465332,
197
+ 0.007353375911712647,
198
+ 0.00738918399810791,
199
+ 0.007412735939025879,
200
+ 0.007360511779785156,
201
+ 0.00724070405960083,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
202
  0.007229440212249756,
203
+ 0.007301119804382325,
204
+ 0.007398399829864502,
205
+ 0.007327744007110596,
206
+ 0.007287744045257568,
207
+ 0.00727347183227539,
208
+ 0.007217152118682861,
209
+ 0.007186431884765625,
210
+ 0.007299071788787842,
211
+ 0.007193600177764893,
212
+ 0.007314432144165039,
213
+ 0.007238656044006348,
214
+ 0.0071833600997924804,
215
+ 0.007226367950439453,
216
+ 0.007284736156463623,
217
+ 0.007230463981628418,
218
+ 0.0072325119972229,
219
+ 0.007223296165466309,
220
+ 0.007292928218841553,
221
+ 0.0074065918922424315,
222
+ 0.007419904232025146,
223
+ 0.00729702377319336,
224
+ 0.007353343963623047,
225
+ 0.007304192066192627,
226
+ 0.007294976234436036,
227
+ 0.00733081579208374,
228
+ 0.007244800090789795,
229
+ 0.007251967906951904,
230
+ 0.007227392196655274,
231
+ 0.007142399787902832,
232
+ 0.0071833600997924804,
233
+ 0.007017471790313721,
234
+ 0.007057407855987549,
235
+ 0.007196671962738037,
236
+ 0.0070256638526916505,
237
+ 0.006918144226074219,
238
+ 0.006931456089019775,
239
+ 0.0069283838272094726,
240
+ 0.0070225920677185055,
241
+ 0.007351295948028564,
242
+ 0.007549952030181885,
243
+ 0.007677951812744141,
244
+ 0.007669760227203369,
245
+ 0.007748608112335205,
246
+ 0.0075642881393432615,
247
+ 0.007512063980102539,
248
+ 0.007333888053894043,
249
+ 0.007370751857757568,
250
+ 0.007291903972625732,
251
+ 0.007304192066192627,
252
+ 0.007245823860168457,
253
+ 0.007302144050598145,
254
+ 0.007258111953735351,
255
+ 0.007241727828979493,
256
  0.007250944137573242,
257
+ 0.007223296165466309,
258
+ 0.007178239822387696,
259
+ 0.007174143791198731,
260
+ 0.007221248149871826,
261
+ 0.007225344181060791,
262
+ 0.007222208023071289,
263
+ 0.007202816009521484,
264
+ 0.007214079856872559,
265
+ 0.007198719978332519,
266
+ 0.0072120318412780765,
267
+ 0.00722431993484497,
268
+ 0.007174143791198731,
269
+ 0.007226367950439453,
270
+ 0.00719046401977539,
271
+ 0.0072325119972229,
272
+ 0.007215104103088379,
273
+ 0.007208960056304932,
274
+ 0.007202816009521484,
275
+ 0.007218175888061523,
276
+ 0.007218175888061523,
277
+ 0.007214079856872559,
278
+ 0.007211008071899414,
279
+ 0.0072130560874938965,
280
+ 0.007202816009521484,
281
+ 0.0071823358535766605,
282
+ 0.0072427520751953125,
283
+ 0.007209983825683594,
284
+ 0.007214079856872559,
285
+ 0.0072202239036560055,
286
+ 0.0071905279159545895,
287
+ 0.007229440212249756,
288
+ 0.007227392196655274,
289
+ 0.007196703910827637,
290
+ 0.007198719978332519,
291
+ 0.007193600177764893,
292
+ 0.007209983825683594,
293
+ 0.007262207984924316
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
294
  ]
295
  },
296
  "throughput": {
297
  "unit": "samples/s",
298
+ "value": 137.52327946471874
299
  },
300
  "energy": {
301
  "unit": "kWh",
302
+ "cpu": 8.108661075437549e-08,
303
+ "ram": 4.4329133971858914e-08,
304
+ "gpu": 1.4897652709589033e-07,
305
+ "total": 2.743922718221247e-07
306
  },
307
  "efficiency": {
308
  "unit": "samples/kWh",
309
+ "value": 3644417.509864315
310
  }
311
  }
312
  }