IlyasMoutawwakil HF staff commited on
Commit
2439cfe
·
verified ·
1 Parent(s): 18558d8

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.0.dev20231010+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -103,7 +103,7 @@
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
- "peft_version": "0.12.0",
107
  "peft_commit": null
108
  }
109
  },
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 955.37152,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 46054.899712,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.9612412109375,
124
- "mean": 7.9612412109375,
125
  "stdev": 0.0,
126
- "p50": 7.9612412109375,
127
- "p90": 7.9612412109375,
128
- "p95": 7.9612412109375,
129
- "p99": 7.9612412109375,
130
  "values": [
131
- 7.9612412109375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,159 +138,159 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1079.697408,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 234111.123456,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
  "count": 131,
150
- "total": 0.9961052231788636,
151
- "mean": 0.007603856665487509,
152
- "stdev": 0.0004265459507820425,
153
- "p50": 0.007486896991729736,
154
- "p90": 0.008199698448181152,
155
- "p95": 0.008340418338775635,
156
- "p99": 0.00857651443481445,
157
  "values": [
158
- 0.008021777153015136,
159
- 0.007926416873931886,
160
- 0.007594417095184327,
161
- 0.0072937760353088375,
162
- 0.007413296222686767,
163
- 0.007285776138305664,
164
- 0.007247536182403564,
165
- 0.007554896831512451,
166
- 0.008662578582763671,
167
- 0.007443535804748535,
168
- 0.00760785722732544,
169
- 0.007505775928497315,
170
- 0.0075635361671447755,
171
- 0.007335536003112793,
172
- 0.007705457210540771,
173
- 0.00767937707901001,
174
- 0.007368656158447265,
175
- 0.007850098133087157,
176
- 0.007745616912841797,
177
- 0.00742577600479126,
178
- 0.007508176803588867,
179
- 0.0074099359512329105,
180
- 0.007325615882873535,
181
- 0.0077107372283935546,
182
- 0.007360176086425781,
183
- 0.007600337028503418,
184
- 0.007743536949157715,
185
- 0.007393136024475098,
186
- 0.0071633758544921875,
187
- 0.007995378017425538,
188
- 0.0081176176071167,
189
- 0.008199698448181152,
190
- 0.00813937759399414,
191
- 0.00815137767791748,
192
- 0.00812721824645996,
193
- 0.008158098220825196,
194
- 0.00813777732849121,
195
- 0.00814993667602539,
196
- 0.008169297218322755,
197
- 0.00837569808959961,
198
- 0.008165457725524902,
199
- 0.008328337669372559,
200
- 0.008372499465942382,
201
- 0.008357297897338868,
202
- 0.008302417755126953,
203
- 0.008255698204040528,
204
- 0.008290098190307618,
205
- 0.008340818405151367,
206
- 0.008364978790283204,
207
- 0.008340018272399902,
208
- 0.008292338371276856,
209
- 0.009305139541625977,
210
- 0.008039697647094727,
211
- 0.00798401689529419,
212
- 0.008004817962646484,
213
- 0.00801761817932129,
214
- 0.008003856658935547,
215
- 0.00796577787399292,
216
- 0.007941297054290771,
217
- 0.007944337844848633,
218
- 0.00747489595413208,
219
- 0.007304975986480713,
220
- 0.0072344160079956055,
221
  0.007236495971679687,
222
- 0.0071824159622192385,
223
- 0.007302256107330323,
224
- 0.007233615875244141,
225
- 0.007175856113433838,
226
- 0.007179696083068848,
227
- 0.007356815814971924,
228
- 0.007478256225585938,
229
- 0.00723473596572876,
230
- 0.007482097148895263,
231
- 0.007245456218719483,
232
- 0.0072345762252807614,
233
- 0.0074876961708068845,
234
- 0.007244976043701172,
235
- 0.007231855869293213,
236
- 0.0074337759017944335,
237
- 0.0072987360954284665,
238
- 0.0075212960243225095,
239
- 0.007277616024017334,
240
- 0.007147696018218994,
241
- 0.0071203351020812985,
242
- 0.007130895137786865,
243
- 0.007273615837097168,
244
- 0.0072508959770202636,
245
- 0.007359055995941162,
246
- 0.007612977027893066,
247
- 0.007435697078704834,
248
- 0.007696656227111817,
249
- 0.007633296012878418,
250
- 0.007731216907501221,
251
- 0.0072632160186767575,
252
  0.007249616146087647,
253
- 0.007459696769714356,
254
- 0.007401616096496582,
255
- 0.0072864160537719725,
256
- 0.007118095874786377,
257
- 0.007119854927062988,
258
- 0.007117135047912598,
259
- 0.007134096145629883,
260
- 0.00714737606048584,
261
- 0.0072076950073242186,
262
- 0.007109775066375733,
263
- 0.00731601619720459,
264
- 0.008031377792358399,
265
- 0.008019218444824219,
266
- 0.00798721694946289,
267
- 0.007677617073059082,
268
- 0.007495216846466065,
269
- 0.007550415992736817,
270
- 0.007553136825561523,
271
- 0.007124015808105469,
272
- 0.007525135993957519,
273
- 0.008045297622680664,
274
- 0.007475215911865235,
275
- 0.007486896991729736,
276
- 0.007596336841583252,
277
- 0.007579216003417969,
278
- 0.007541457176208496,
279
- 0.007117136001586914,
280
- 0.007163693904876709,
281
- 0.007150735855102539,
282
- 0.007136816024780273,
283
- 0.007164656162261963,
284
- 0.007167695045471191,
285
- 0.007138415813446045,
286
- 0.007142576217651367,
287
- 0.007120336055755615,
288
- 0.007155216217041015
 
 
 
 
 
 
 
 
 
 
 
 
 
289
  ]
290
  },
291
  "throughput": {
292
  "unit": "samples/s",
293
- "value": 131.5122107099696
294
  },
295
  "energy": null,
296
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
+ "peft_version": null,
107
  "peft_commit": null
108
  }
109
  },
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1031.028736,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 43248.467968,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.70135302734375,
124
+ "mean": 7.70135302734375,
125
  "stdev": 0.0,
126
+ "p50": 7.70135302734375,
127
+ "p90": 7.70135302734375,
128
+ "p95": 7.70135302734375,
129
+ "p99": 7.70135302734375,
130
  "values": [
131
+ 7.70135302734375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1148.33408,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 211202.048,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
  "count": 131,
150
+ "total": 1.0011928222179418,
151
+ "mean": 0.007642693299373597,
152
+ "stdev": 0.0018883905585774827,
153
+ "p50": 0.007370256900787354,
154
+ "p90": 0.008073779106140136,
155
+ "p95": 0.008486579418182373,
156
+ "p99": 0.018086472034454286,
157
  "values": [
158
+ 0.008713299751281739,
159
+ 0.008611379623413086,
160
+ 0.008302099227905273,
161
+ 0.008216338157653809,
162
+ 0.008281139373779297,
163
+ 0.00856369972229004,
164
+ 0.008110418319702148,
165
+ 0.008102898597717284,
166
+ 0.008000337600708007,
167
+ 0.007983538150787353,
168
+ 0.007876177787780762,
169
+ 0.007830577850341797,
170
+ 0.014402112007141114,
171
+ 0.007352497100830078,
172
+ 0.007817937850952148,
173
+ 0.00781073808670044,
174
+ 0.00783201789855957,
175
+ 0.007724017143249512,
176
+ 0.007753617763519287,
177
+ 0.00781777811050415,
178
+ 0.007734736919403076,
179
+ 0.007731537818908691,
180
+ 0.007690257072448731,
181
+ 0.011262745857238769,
182
+ 0.008409459114074707,
183
+ 0.00799793815612793,
184
+ 0.007993138790130616,
185
+ 0.007896657943725585,
186
+ 0.007542256832122803,
187
+ 0.007531857013702393,
188
+ 0.007606417179107666,
189
+ 0.007795217990875244,
190
+ 0.007558577060699463,
191
+ 0.007485456943511963,
192
+ 0.007700977802276611,
193
+ 0.007449616909027099,
194
+ 0.007288176059722901,
195
+ 0.007359857082366943,
196
+ 0.007501136779785156,
197
+ 0.007589137077331543,
198
+ 0.007709617137908935,
199
+ 0.007591856956481934,
200
+ 0.007747858047485351,
201
+ 0.007729776859283447,
202
+ 0.007542097091674805,
203
+ 0.007617776870727539,
204
+ 0.007573936939239502,
205
+ 0.00773809814453125,
206
+ 0.007599536895751953,
207
+ 0.007647377014160156,
208
+ 0.007632018089294434,
209
+ 0.007370256900787354,
210
+ 0.007216815948486328,
211
+ 0.007310737133026123,
212
+ 0.0072612957954406735,
213
+ 0.0072440171241760255,
214
+ 0.007186895847320557,
215
+ 0.007284336090087891,
216
+ 0.007280655860900879,
 
 
 
 
217
  0.007236495971679687,
218
+ 0.007180816173553467,
219
+ 0.00732209587097168,
220
+ 0.0072742562294006346,
221
+ 0.0072411360740661625,
222
+ 0.007207376003265381,
223
+ 0.020342605590820313,
224
+ 0.0025214459896087645,
225
+ 0.0043776102066040035,
226
+ 0.007239696025848389,
227
+ 0.007212976932525635,
228
+ 0.007189136028289795,
229
+ 0.007223375797271728,
230
+ 0.007229297161102295,
231
+ 0.007177775859832764,
232
+ 0.007160016059875488,
233
+ 0.007139697074890137,
234
+ 0.007213295936584473,
235
+ 0.007211535930633545,
236
+ 0.00716785717010498,
237
+ 0.00716977596282959,
238
+ 0.007239376068115234,
 
 
 
 
 
 
 
 
 
239
  0.007249616146087647,
240
+ 0.0072276959419250485,
241
+ 0.007265777111053467,
242
+ 0.007277935981750488,
243
+ 0.007252655982971192,
244
+ 0.007018735885620117,
245
+ 0.007043375968933106,
246
+ 0.007052015781402588,
247
+ 0.0070070562362670894,
248
+ 0.0070428957939147946,
249
+ 0.007088016033172608,
250
+ 0.007200175762176513,
251
+ 0.007226096153259277,
252
+ 0.007044976234436035,
253
+ 0.007054736137390136,
254
+ 0.007033616065979004,
255
+ 0.007064335823059082,
256
+ 0.00720033597946167,
257
+ 0.007015535831451416,
258
+ 0.007015855789184571,
259
+ 0.00803489875793457,
260
+ 0.007971377849578857,
261
+ 0.00798545789718628,
262
+ 0.008073779106140136,
263
+ 0.008064977645874024,
264
+ 0.0076849770545959475,
265
+ 0.0075641770362854,
266
+ 0.007361135959625244,
267
+ 0.008009617805480956,
268
+ 0.019665483474731446,
269
+ 0.0025139260292053223,
270
+ 0.00343776798248291,
271
+ 0.007988977909088134,
272
+ 0.00798145818710327,
273
+ 0.007908338069915772,
274
+ 0.007409136772155761,
275
+ 0.007478416919708252,
276
+ 0.007379056930541993,
277
+ 0.007392656803131103,
278
+ 0.007257135868072509,
279
+ 0.007009776115417481,
280
+ 0.007010575771331787,
281
+ 0.007044976234436035,
282
+ 0.0073627371788024904,
283
+ 0.007231056213378906,
284
+ 0.007019375801086426,
285
+ 0.007048495769500732,
286
+ 0.00704753589630127,
287
+ 0.0070523362159729,
288
+ 0.007031374931335449
289
  ]
290
  },
291
  "throughput": {
292
  "unit": "samples/s",
293
+ "value": 130.8439264574389
294
  },
295
  "energy": null,
296
  "efficiency": null