IlyasMoutawwakil HF staff commited on
Commit
06f8b92
·
verified ·
1 Parent(s): c4fb09d

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1042.898944,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 47311.425536,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 8.1272890625,
124
- "mean": 8.1272890625,
125
  "stdev": 0.0,
126
- "p50": 8.1272890625,
127
- "p90": 8.1272890625,
128
- "p95": 8.1272890625,
129
- "p99": 8.1272890625,
130
  "values": [
131
- 8.1272890625
132
  ]
133
  },
134
  "throughput": null,
@@ -138,153 +138,160 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1160.368128,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 242295.283712,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 125,
150
- "total": 0.9999058513641358,
151
- "mean": 0.007999246810913086,
152
- "stdev": 0.004200741685445301,
153
- "p50": 0.00787277317047119,
154
- "p90": 0.008233635139465333,
155
- "p95": 0.008976192855834961,
156
- "p99": 0.026590379562378037,
157
  "values": [
158
- 0.008843008995056151,
159
- 0.008920289993286134,
160
- 0.009026688575744629,
161
- 0.009074209213256836,
162
- 0.008979168891906739,
163
- 0.008964288711547851,
164
- 0.009096928596496582,
165
- 0.00873548984527588,
166
- 0.00865565013885498,
167
- 0.031488052368164064,
168
- 0.002487191915512085,
169
- 0.0024691109657287596,
170
- 0.004666384220123291,
171
- 0.008293730735778809,
172
- 0.008143491744995117,
173
- 0.011081082344055176,
174
- 0.007774212837219238,
175
- 0.0075887742042541504,
176
- 0.007591494083404541,
177
- 0.007820292949676514,
178
- 0.007962052822113036,
179
- 0.007718052864074707,
180
- 0.007906373023986817,
181
- 0.008040772438049317,
182
- 0.007658214092254639,
183
- 0.007562694072723389,
184
- 0.00762541389465332,
185
- 0.007758533954620362,
186
- 0.007729093074798584,
187
- 0.007545574188232422,
188
- 0.007648293972015381,
189
- 0.007802052974700928,
190
- 0.0076791739463806154,
191
- 0.0074753351211547855,
192
- 0.007713412761688233,
193
- 0.007678214073181152,
194
- 0.007700772762298584,
195
- 0.0075409340858459475,
196
- 0.007701734066009522,
197
- 0.007686373233795166,
198
- 0.007877892971038818,
199
- 0.007612452983856201,
200
- 0.007846532821655273,
201
- 0.007629893779754639,
202
- 0.007510694026947021,
203
- 0.007463174819946289,
204
- 0.007499334812164307,
205
- 0.007454693794250488,
206
- 0.007495654106140137,
207
- 0.007328774929046631,
208
- 0.007534694194793701,
209
- 0.007204774856567383,
210
- 0.007334054946899414,
211
- 0.007266375064849853,
212
- 0.007225415229797363,
213
- 0.007173894882202148,
214
- 0.007237575054168701,
215
- 0.007211336135864258,
216
- 0.007520934104919433,
217
- 0.007954051971435547,
218
- 0.007951013088226318,
219
- 0.007951653003692627,
220
- 0.008022532463073731,
221
- 0.007963172912597656,
222
- 0.007969573020935058,
223
- 0.007971013069152832,
224
- 0.007970531940460206,
225
- 0.007941573143005372,
226
- 0.007973092079162598,
227
- 0.007957733154296875,
228
- 0.007938212871551513,
229
- 0.007953251838684082,
230
- 0.0079564528465271,
231
- 0.007940292835235596,
232
- 0.00793693208694458,
233
- 0.007937092781066895,
234
- 0.007919172763824462,
235
- 0.007927492141723632,
236
- 0.00797613286972046,
237
- 0.007650053977966308,
238
- 0.007431654930114746,
239
- 0.007606694221496582,
240
- 0.007563814163208008,
241
- 0.007771812915802002,
242
- 0.007639974117279053,
243
- 0.00787277317047119,
244
- 0.007581093788146973,
245
- 0.007507015228271485,
246
- 0.007493093967437744,
247
- 0.007948613166809082,
248
- 0.007963653087615968,
249
- 0.007958532810211182,
250
- 0.007945251941680909,
251
- 0.04497840881347656,
252
- 0.0024702320098876954,
253
- 0.0024502320289611816,
254
- 0.0024513518810272217,
255
- 0.002464310884475708,
256
- 0.002443351984024048,
257
- 0.0024588708877563478,
258
- 0.006883976936340332,
259
- 0.008063812255859375,
260
- 0.007969411849975586,
261
- 0.007606373786926269,
262
- 0.007581093788146973,
263
- 0.007571653842926025,
264
- 0.007699812889099121,
265
- 0.00803437328338623,
266
- 0.008029253005981445,
267
- 0.008060452461242675,
268
- 0.00802029323577881,
269
- 0.008025253295898437,
270
- 0.008061092376708984,
271
- 0.008026851654052734,
272
- 0.007973573207855224,
273
- 0.007946053028106689,
274
- 0.008012131690979004,
275
- 0.008058371543884278,
276
- 0.008007493019104004,
277
- 0.00800861167907715,
278
- 0.008032133102416992,
279
- 0.00807581329345703,
280
- 0.0080940523147583,
281
- 0.008020931243896485,
282
- 0.008020453453063965
 
 
 
 
 
 
 
283
  ]
284
  },
285
  "throughput": {
286
  "unit": "samples/s",
287
- "value": 125.01176968758305
288
  },
289
  "energy": null,
290
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.0.dev20231010+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 957.067264,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 45264.879616,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 8.1591474609375,
124
+ "mean": 8.1591474609375,
125
  "stdev": 0.0,
126
+ "p50": 8.1591474609375,
127
+ "p90": 8.1591474609375,
128
+ "p95": 8.1591474609375,
129
+ "p99": 8.1591474609375,
130
  "values": [
131
+ 8.1591474609375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1082.011648,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 224903.74144,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 132,
150
+ "total": 1.0011543745994573,
151
+ "mean": 0.007584502837874673,
152
+ "stdev": 0.0004292566322914227,
153
+ "p50": 0.007838650941848754,
154
+ "p90": 0.007963692092895509,
155
+ "p95": 0.008122539520263669,
156
+ "p99": 0.008589225034713745,
157
  "values": [
158
+ 0.008274251937866212,
159
+ 0.008283371925354005,
160
+ 0.008418412208557128,
161
+ 0.008536011695861816,
162
+ 0.008411372184753418,
163
+ 0.007856491088867188,
164
+ 0.00783889102935791,
165
+ 0.0078384108543396,
166
+ 0.00861313247680664,
167
+ 0.0072984108924865725,
168
+ 0.007314410209655762,
169
+ 0.007367209911346435,
170
+ 0.007254089832305908,
171
+ 0.007227369785308838,
172
+ 0.007247210025787354,
173
+ 0.007194089889526367,
174
+ 0.007196170806884766,
175
+ 0.007229290008544922,
176
+ 0.00724144983291626,
177
+ 0.007247529983520508,
178
+ 0.007233130931854248,
179
+ 0.007244969844818115,
180
+ 0.007247529983520508,
181
+ 0.007240970134735108,
182
+ 0.007234410762786865,
183
+ 0.007214089870452881,
184
+ 0.007253449916839599,
185
+ 0.007285610198974609,
186
+ 0.008720172882080077,
187
+ 0.007212649822235107,
188
+ 0.007037769794464111,
189
+ 0.007066249847412109,
190
+ 0.0070680098533630374,
191
+ 0.007064169883728028,
192
+ 0.007062089920043945,
193
+ 0.007052330017089844,
194
+ 0.0070816102027893065,
195
+ 0.00721136999130249,
196
+ 0.007110889911651611,
197
+ 0.007176169872283936,
198
+ 0.007104810237884522,
199
+ 0.007099370002746582,
200
+ 0.007094890117645264,
201
+ 0.007053770065307617,
202
+ 0.007066569805145264,
203
+ 0.007092010021209716,
204
+ 0.007109449863433838,
205
+ 0.007066249847412109,
206
+ 0.007102250099182129,
207
+ 0.0070590500831604,
208
+ 0.007071209907531738,
209
+ 0.007064010143280029,
210
+ 0.007040649890899658,
211
+ 0.007068809986114502,
212
+ 0.007021289825439453,
213
+ 0.007852651119232178,
214
+ 0.007250569820404053,
215
+ 0.007023849964141846,
216
+ 0.007028329849243164,
217
+ 0.00738336992263794,
218
+ 0.007909292221069337,
219
+ 0.00791697120666504,
220
+ 0.007924971103668212,
221
+ 0.007927530765533448,
222
+ 0.007848491191864013,
223
+ 0.007868171215057373,
224
+ 0.007905611991882324,
225
+ 0.007927051067352295,
226
+ 0.00788081121444702,
227
+ 0.007910891056060791,
228
+ 0.007946731090545655,
229
+ 0.007884811878204346,
230
+ 0.007883211135864258,
231
+ 0.007943370819091796,
232
+ 0.007887371063232422,
233
+ 0.007950571060180664,
234
+ 0.007928492069244385,
235
+ 0.007919530868530273,
236
+ 0.00798737096786499,
237
+ 0.007976810932159423,
238
+ 0.007964332103729248,
239
+ 0.007934091091156005,
240
+ 0.007928811073303222,
241
+ 0.007943850994110107,
242
+ 0.007888011932373047,
243
+ 0.00793121099472046,
244
+ 0.007913451194763183,
245
+ 0.007898090839385986,
246
+ 0.007938090801239014,
247
+ 0.00787585210800171,
248
+ 0.007908811092376709,
249
+ 0.007916491031646728,
250
+ 0.007904651165008545,
251
+ 0.007888970851898193,
252
+ 0.00794353199005127,
253
+ 0.007928811073303222,
254
+ 0.007920011043548584,
255
+ 0.007955690860748292,
256
+ 0.007896811008453368,
257
+ 0.007011849880218506,
258
+ 0.006990409851074218,
259
+ 0.007066090106964112,
260
+ 0.007064330101013184,
261
+ 0.00714976978302002,
262
+ 0.007043529987335205,
263
+ 0.007074570178985596,
264
+ 0.007078410148620606,
265
+ 0.00754417085647583,
266
+ 0.007887051105499268,
267
+ 0.00790337085723877,
268
+ 0.007912812232971191,
269
+ 0.00736097002029419,
270
+ 0.007671531200408936,
271
+ 0.007963691234588623,
272
+ 0.007922251224517823,
273
+ 0.007937451839447022,
274
+ 0.0076105699539184574,
275
+ 0.007929932117462159,
276
+ 0.007933771133422852,
277
+ 0.007998411178588866,
278
+ 0.007883851051330566,
279
+ 0.00796369218826294,
280
+ 0.007943211078643798,
281
+ 0.007977130889892579,
282
+ 0.007946731090545655,
283
+ 0.007974731922149658,
284
+ 0.007945291042327881,
285
+ 0.007930410861968993,
286
+ 0.007345769882202148,
287
+ 0.007323369979858398,
288
+ 0.007384009838104248,
289
+ 0.007315210819244385
290
  ]
291
  },
292
  "throughput": {
293
  "unit": "samples/s",
294
+ "value": 131.84779825070507
295
  },
296
  "energy": null,
297
  "efficiency": null