IlyasMoutawwakil HF staff commited on
Commit
de6e982
·
verified ·
1 Parent(s): 39324a2

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1384.710144,
108
- "max_global_vram": 0.0,
109
- "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.37152
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.72025439453125,
117
- "mean": 7.72025439453125,
118
  "stdev": 0.0,
119
- "p50": 7.72025439453125,
120
- "p90": 7.72025439453125,
121
- "p95": 7.72025439453125,
122
- "p99": 7.72025439453125,
123
  "values": [
124
- 7.72025439453125
125
  ]
126
  },
127
  "throughput": null,
@@ -131,149 +131,153 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1510.25664,
135
- "max_global_vram": 0.0,
136
- "max_process_vram": 0.0,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.5072
139
  },
140
  "latency": {
141
  "unit": "s",
142
- "count": 121,
143
- "total": 0.9995409755706787,
144
- "mean": 0.008260669219592386,
145
- "stdev": 0.00037755664228431376,
146
- "p50": 0.008145094871520996,
147
- "p90": 0.008546053886413574,
148
- "p95": 0.008649733543395996,
149
- "p99": 0.009198052406311035,
150
  "values": [
151
- 0.008545093536376953,
152
- 0.008365573883056641,
153
- 0.008546053886413574,
154
- 0.008647334098815919,
155
- 0.00901837158203125,
156
- 0.008649733543395996,
157
- 0.008639814376831055,
158
- 0.008643653869628906,
159
- 0.008459814071655273,
160
- 0.008419014930725098,
161
- 0.008383813858032227,
162
- 0.008406214714050293,
163
- 0.00835069465637207,
164
- 0.008327175140380859,
165
- 0.00824045467376709,
166
- 0.008106855392456055,
167
- 0.008129095077514649,
168
- 0.008127495765686036,
169
- 0.008187335968017578,
170
- 0.008145094871520996,
171
- 0.008198214530944823,
172
- 0.008251495361328125,
173
- 0.008183494567871093,
174
- 0.008185415267944337,
175
- 0.008207815170288086,
176
- 0.008247654914855957,
177
- 0.011565725326538085,
178
- 0.008312294006347656,
179
- 0.00821357536315918,
180
- 0.008197895050048828,
181
- 0.008309735298156738,
182
- 0.008152774810791016,
183
- 0.00815965461730957,
184
- 0.008135496139526367,
185
- 0.008137575149536133,
186
- 0.008166854858398437,
187
- 0.008166375160217286,
188
- 0.008154054641723633,
189
- 0.008111335754394532,
190
- 0.008105895042419434,
191
- 0.008122855186462403,
192
- 0.008099016189575195,
193
- 0.008162694931030274,
194
- 0.008104615211486817,
195
- 0.00809709644317627,
196
- 0.008079976081848144,
197
- 0.008097894668579101,
198
- 0.008120615959167481,
199
- 0.008170056343078613,
200
- 0.008152295112609863,
201
- 0.008117895126342773,
202
- 0.008127175331115722,
203
- 0.00813421630859375,
204
- 0.008136775016784668,
205
- 0.008176295280456543,
206
- 0.008159495353698731,
207
- 0.008179015159606933,
208
- 0.008101896286010742,
209
- 0.008125576019287109,
210
- 0.008128774642944337,
211
- 0.008081255912780761,
212
- 0.008094856262207031,
213
- 0.008088134765625,
214
- 0.008110376358032226,
215
- 0.008112456321716309,
216
- 0.008114375114440919,
217
- 0.008101736068725586,
218
- 0.008118696212768554,
219
- 0.008112935066223145,
220
- 0.008158855438232422,
221
- 0.008136136054992676,
222
- 0.008129096031188965,
223
- 0.008158855438232422,
224
- 0.008120615005493164,
225
- 0.008095975875854491,
226
- 0.008130375862121581,
227
- 0.008114054679870605,
228
- 0.008124935150146484,
229
- 0.008134856224060058,
230
- 0.00819661521911621,
231
- 0.008129255294799805,
232
- 0.008297574996948243,
233
- 0.00834205436706543,
234
- 0.008128456115722656,
235
- 0.008094855308532715,
236
- 0.008146055221557617,
237
- 0.00808445644378662,
238
- 0.008063976287841796,
239
- 0.008102854728698731,
240
- 0.008092776298522949,
241
- 0.008046534538269043,
242
- 0.008044936180114747,
243
- 0.008055974960327148,
244
- 0.00826701545715332,
245
- 0.008296935081481934,
246
- 0.008341733932495116,
247
- 0.00806173610687256,
248
- 0.008066214561462403,
249
- 0.008059656143188477,
250
- 0.008307015419006348,
251
- 0.008125096321105958,
252
- 0.008150215148925782,
253
- 0.008115015029907226,
254
- 0.008079815864562988,
255
- 0.008093894958496094,
256
- 0.008080615043640136,
257
- 0.008118696212768554,
258
- 0.008184294700622558,
259
- 0.00816989517211914,
260
- 0.008299015045166015,
261
- 0.008273573875427246,
262
- 0.008321894645690917,
263
- 0.008097414970397948,
264
- 0.00811181640625,
265
- 0.00813389492034912,
266
- 0.008836133003234863,
267
- 0.009183972358703614,
268
- 0.009201572418212891,
269
- 0.009179812431335449,
270
- 0.008590373992919922,
271
- 0.008632452964782714
 
 
 
 
272
  ]
273
  },
274
  "throughput": {
275
  "unit": "samples/s",
276
- "value": 121.05556746277077
277
  },
278
  "energy": null,
279
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 905.58464,
108
+ "max_global_vram": 841.764864,
109
+ "max_process_vram": 44807.573504,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.37152
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.1981728515625,
117
+ "mean": 7.1981728515625,
118
  "stdev": 0.0,
119
+ "p50": 7.1981728515625,
120
+ "p90": 7.1981728515625,
121
+ "p95": 7.1981728515625,
122
+ "p99": 7.1981728515625,
123
  "values": [
124
+ 7.1981728515625
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1015.414784,
135
+ "max_global_vram": 898.269184,
136
+ "max_process_vram": 218894.667776,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.5072
139
  },
140
  "latency": {
141
  "unit": "s",
142
+ "count": 125,
143
+ "total": 0.9963846902847285,
144
+ "mean": 0.007971077522277832,
145
+ "stdev": 0.0007331559968759307,
146
+ "p50": 0.007840778827667236,
147
+ "p90": 0.008722056388854982,
148
+ "p95": 0.00900333595275879,
149
+ "p99": 0.009678912391662601,
150
  "values": [
151
+ 0.008753096580505371,
152
+ 0.008942537307739258,
153
+ 0.009018535614013673,
154
+ 0.009047016143798828,
155
+ 0.00908973503112793,
156
+ 0.009133416175842285,
157
+ 0.009071334838867188,
158
+ 0.009851174354553223,
159
+ 0.00875389575958252,
160
+ 0.008851016044616699,
161
+ 0.008799016952514649,
162
+ 0.008846055984497071,
163
+ 0.008675496101379394,
164
+ 0.00845213794708252,
165
+ 0.008345577239990234,
166
+ 0.008319977760314942,
167
+ 0.008330376625061035,
168
+ 0.008364458084106446,
169
+ 0.00835085678100586,
170
+ 0.00847997760772705,
171
+ 0.00836589813232422,
172
+ 0.008303977966308594,
173
+ 0.008316938400268554,
174
+ 0.008310697555541991,
175
+ 0.008345577239990234,
176
+ 0.008294537544250488,
177
+ 0.008319658279418945,
178
+ 0.00830253791809082,
179
+ 0.008283978462219238,
180
+ 0.008259178161621094,
181
+ 0.008320937156677246,
182
+ 0.008278377532958984,
183
+ 0.007942698955535888,
184
+ 0.008075018882751465,
185
+ 0.008315977096557617,
186
+ 0.008304298400878906,
187
+ 0.008298057556152344,
188
+ 0.008298856735229492,
189
+ 0.00833421802520752,
190
+ 0.008318697929382324,
191
+ 0.00783117914199829,
192
+ 0.007840778827667236,
193
+ 0.007843499183654784,
194
+ 0.007860138893127441,
195
+ 0.00790685796737671,
196
+ 0.0077487797737121585,
197
+ 0.007526059150695801,
198
+ 0.0074809398651123045,
199
+ 0.007532299995422364,
200
+ 0.0075518198013305666,
201
+ 0.0075018987655639646,
202
+ 0.007574540138244629,
203
+ 0.007522220134735107,
204
+ 0.0073225398063659665,
205
+ 0.007338860034942627,
206
+ 0.007332780838012695,
207
+ 0.007328781127929687,
208
+ 0.007329100131988525,
209
+ 0.007349100112915039,
210
+ 0.00734749984741211,
211
+ 0.007346701145172119,
212
+ 0.007363979816436768,
213
+ 0.007343180179595947,
214
+ 0.007353579998016358,
215
+ 0.00728718090057373,
216
+ 0.0075513401031494145,
217
+ 0.007344620227813721,
218
+ 0.007328620910644531,
219
+ 0.007329740047454834,
220
+ 0.007349899768829346,
221
+ 0.00729725980758667,
222
+ 0.007347980976104737,
223
+ 0.0073580598831176755,
224
+ 0.007381740093231201,
225
+ 0.007370059967041016,
226
+ 0.00741726016998291,
227
+ 0.013565724372863769,
228
+ 0.006860140800476074,
229
+ 0.007444139957427979,
230
+ 0.007424620151519776,
231
+ 0.007429259777069092,
232
+ 0.007548460006713867,
233
+ 0.007392139911651612,
234
+ 0.007597899913787841,
235
+ 0.007853418827056884,
236
+ 0.007895978927612304,
237
+ 0.007876139163970947,
238
+ 0.00805965805053711,
239
+ 0.008069578170776368,
240
+ 0.008212298393249511,
241
+ 0.00815181827545166,
242
+ 0.008466697692871094,
243
+ 0.008303017616271972,
244
+ 0.008314698219299317,
245
+ 0.008339177131652832,
246
+ 0.008228938102722167,
247
+ 0.008234857559204101,
248
+ 0.007965578079223632,
249
+ 0.007995338916778564,
250
+ 0.0077575788497924805,
251
+ 0.00777341890335083,
252
+ 0.00776957893371582,
253
+ 0.008112937927246093,
254
+ 0.00833309841156006,
255
+ 0.008306378364562989,
256
+ 0.007726058959960937,
257
+ 0.007640140056610107,
258
+ 0.007686379909515381,
259
+ 0.0076103792190551755,
260
+ 0.0077166190147399906,
261
+ 0.007756619930267334,
262
+ 0.00777966022491455,
263
+ 0.007792298793792725,
264
+ 0.007614058971405029,
265
+ 0.007431980133056641,
266
+ 0.007464300155639649,
267
+ 0.007458380222320557,
268
+ 0.007469580173492431,
269
+ 0.007443339824676514,
270
+ 0.007457419872283936,
271
+ 0.007420939922332764,
272
+ 0.007431340217590332,
273
+ 0.007448619842529297,
274
+ 0.007426061153411865,
275
+ 0.007421259880065918
276
  ]
277
  },
278
  "throughput": {
279
  "unit": "samples/s",
280
+ "value": 125.45355345060524
281
  },
282
  "energy": null,
283
  "efficiency": null