IlyasMoutawwakil HF staff commited on
Commit
228db39
·
verified ·
1 Parent(s): 7ef8ec1

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1387.597824,
115
- "max_global_vram": 11.137024,
116
- "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.724330078125,
124
- "mean": 7.724330078125,
125
  "stdev": 0.0,
126
- "p50": 7.724330078125,
127
- "p90": 7.724330078125,
128
- "p95": 7.724330078125,
129
- "p99": 7.724330078125,
130
  "values": [
131
- 7.724330078125
132
  ]
133
  },
134
  "throughput": null,
@@ -138,151 +138,156 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1513.275392,
142
- "max_global_vram": 11.137024,
143
- "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 123,
150
- "total": 1.0031430969238282,
151
- "mean": 0.008155634934340067,
152
- "stdev": 0.000756080682784084,
153
- "p50": 0.008066534042358398,
154
- "p90": 0.008382821273803711,
155
- "p95": 0.008452102088928222,
156
- "p99": 0.010802285728454592,
157
  "values": [
158
- 0.008452134132385254,
159
- 0.008078055381774902,
160
- 0.008163015365600585,
161
- 0.00865021324157715,
162
- 0.008270054817199708,
163
- 0.008210535049438477,
164
- 0.008175813674926757,
165
- 0.008211334228515626,
166
- 0.008155653953552246,
167
- 0.00811117458343506,
168
- 0.008066534042358398,
169
- 0.008056614875793457,
170
- 0.007975975036621093,
171
- 0.007956134796142577,
172
- 0.007945895195007324,
173
- 0.007972455024719239,
174
- 0.007935495853424072,
175
- 0.007919976234436035,
176
- 0.007942855834960938,
177
- 0.007962375164031983,
178
- 0.008019174575805664,
179
- 0.007929415225982665,
180
- 0.007950214862823487,
181
- 0.00791773509979248,
182
- 0.007975174903869628,
183
- 0.007947014808654786,
184
- 0.007975494861602784,
185
- 0.011325883865356445,
186
- 0.008217413902282715,
187
- 0.007827176094055175,
188
- 0.007849736213684082,
189
- 0.007926054954528809,
190
- 0.007938694953918458,
191
- 0.007912614822387695,
192
- 0.007939014911651611,
193
- 0.007939014911651611,
194
- 0.007960454940795898,
195
- 0.007965734958648682,
196
- 0.007945895195007324,
197
- 0.007958374977111816,
198
- 0.00797021484375,
199
- 0.007950214862823487,
200
- 0.007962695121765137,
201
- 0.007953414916992188,
202
- 0.00791773509979248,
203
- 0.007903335094451904,
204
- 0.008229093551635742,
205
- 0.00836989402770996,
206
- 0.008275493621826172,
207
- 0.00836125373840332,
208
- 0.008241254806518555,
209
- 0.007936295986175537,
210
- 0.007913256168365478,
211
- 0.007946535110473633,
212
- 0.008145255088806152,
213
- 0.008222694396972657,
214
- 0.007892775058746337,
215
- 0.00805421543121338,
216
- 0.008578693389892578,
217
- 0.008407973289489746,
218
- 0.00844205379486084,
219
- 0.008317733764648437,
220
- 0.008449254035949707,
221
- 0.008271333694458007,
222
- 0.008175813674926757,
223
- 0.008195653915405273,
224
- 0.008250213623046876,
225
- 0.008235013961791993,
226
- 0.008195814132690429,
227
- 0.008343334197998048,
228
- 0.008215014457702636,
229
- 0.008234213829040528,
230
- 0.008240453720092774,
231
- 0.008215173721313477,
232
- 0.00819725513458252,
233
- 0.008281893730163574,
234
- 0.008226055145263672,
235
- 0.008217414855957032,
236
- 0.00819261360168457,
237
- 0.008208933830261231,
238
- 0.00819245433807373,
239
- 0.008246213912963866,
240
- 0.008193254470825195,
241
- 0.008362374305725097,
242
- 0.008296613693237304,
243
- 0.008329093933105469,
244
- 0.008278694152832031,
245
- 0.008289894104003907,
246
- 0.008263653755187988,
247
- 0.008241733551025391,
248
- 0.008279173851013184,
249
- 0.008317573547363281,
250
- 0.008299333572387695,
251
- 0.007913094997406006,
252
- 0.007880935192108154,
253
- 0.007893575191497802,
254
- 0.007916295051574707,
255
- 0.007899335861206055,
256
- 0.007911975860595704,
257
- 0.00794109582901001,
258
- 0.007909095764160157,
259
- 0.007904934883117675,
260
- 0.0079286150932312,
261
- 0.007904934883117675,
262
- 0.007893895149230957,
263
- 0.007906215190887451,
264
- 0.007931015014648438,
265
- 0.007886054992675782,
266
- 0.007906694889068604,
267
- 0.00810013484954834,
268
- 0.00789597511291504,
269
- 0.007890375137329101,
270
- 0.007878695011138916,
271
- 0.00791389513015747,
272
- 0.007885095119476318,
273
- 0.007911015033721923,
274
- 0.008538533210754394,
275
- 0.008945892333984375,
276
- 0.014498515129089356,
277
- 0.0043287849426269535,
278
- 0.008430374145507813,
279
- 0.008451813697814942,
280
- 0.008386053085327149
 
 
 
 
 
281
  ]
282
  },
283
  "throughput": {
284
  "unit": "samples/s",
285
- "value": 122.61461039525028
286
  },
287
  "energy": null,
288
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1032.31488,
115
+ "max_global_vram": 842.375168,
116
+ "max_process_vram": 42494.93504,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.53324609375,
124
+ "mean": 7.53324609375,
125
  "stdev": 0.0,
126
+ "p50": 7.53324609375,
127
+ "p90": 7.53324609375,
128
+ "p95": 7.53324609375,
129
+ "p99": 7.53324609375,
130
  "values": [
131
+ 7.53324609375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1150.091264,
142
+ "max_global_vram": 901.26336,
143
+ "max_process_vram": 217598.58688,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 128,
150
+ "total": 0.9943704147338869,
151
+ "mean": 0.00776851886510849,
152
+ "stdev": 0.0017926209510019933,
153
+ "p50": 0.007607335567474366,
154
+ "p90": 0.008141877937316895,
155
+ "p95": 0.008214917898178102,
156
+ "p99": 0.017659440689086914,
157
  "values": [
158
+ 0.008039174079895019,
159
+ 0.00779165506362915,
160
+ 0.007668615818023682,
161
+ 0.0077193350791931155,
162
+ 0.007897095203399658,
163
+ 0.008219973564147949,
164
+ 0.008006214141845703,
165
+ 0.007959974765777587,
166
+ 0.007768614768981934,
167
+ 0.007706695079803467,
168
+ 0.00767373514175415,
169
+ 0.0076513361930847165,
170
+ 0.00760829496383667,
171
+ 0.007569896221160889,
172
+ 0.007511335849761963,
173
+ 0.007537896156311035,
174
+ 0.007467656135559082,
175
+ 0.0074977359771728515,
176
+ 0.017838504791259764,
177
+ 0.005284783840179444,
178
+ 0.0077135758399963375,
179
+ 0.00772397518157959,
180
+ 0.007687335014343262,
181
+ 0.0076654148101806645,
182
+ 0.007692455768585205,
183
+ 0.007718534946441651,
184
+ 0.008016934394836426,
185
+ 0.008059014320373535,
186
+ 0.008061253547668457,
187
+ 0.008115333557128906,
188
+ 0.0080762939453125,
189
+ 0.008088614463806152,
190
+ 0.008129573822021485,
191
+ 0.008120933532714843,
192
+ 0.008140294075012208,
193
+ 0.008099333763122558,
194
+ 0.008108934402465821,
195
+ 0.008101573944091797,
196
+ 0.008154534339904785,
197
+ 0.008117094039916992,
198
+ 0.008086533546447753,
199
+ 0.008101094245910644,
200
+ 0.00812109375,
201
+ 0.008086693763732911,
202
+ 0.008086693763732911,
203
+ 0.008054694175720216,
204
+ 0.008061253547668457,
205
+ 0.007517416000366211,
206
+ 0.0073060569763183595,
207
+ 0.007281576156616211,
208
+ 0.007431816101074218,
209
+ 0.007100778102874756,
210
+ 0.007135817050933838,
211
+ 0.007145256996154785,
212
+ 0.007144456863403321,
213
+ 0.007137096881866455,
214
+ 0.007160936832427979,
215
+ 0.007224297046661377,
216
+ 0.007212615966796875,
217
+ 0.0071996569633483885,
218
+ 0.007338857173919678,
219
+ 0.007212296962738037,
220
+ 0.007218376159667968,
221
+ 0.0072180571556091306,
222
+ 0.0071868572235107425,
223
+ 0.007220296859741211,
224
+ 0.007234857082366943,
225
+ 0.007274216175079346,
226
+ 0.0072247772216796876,
227
+ 0.007278857231140137,
228
+ 0.007273736000061035,
229
+ 0.017175304412841798,
230
+ 0.0031219100952148437,
231
+ 0.007685895919799805,
232
+ 0.007482056140899658,
233
+ 0.007425896167755127,
234
+ 0.00744157600402832,
235
+ 0.007463975906372071,
236
+ 0.00749021577835083,
237
+ 0.0073978958129882815,
238
+ 0.007424776077270508,
239
+ 0.007387016773223877,
240
+ 0.007334376811981201,
241
+ 0.007647174835205078,
242
+ 0.007319817066192627,
243
+ 0.007430535793304443,
244
+ 0.0072369370460510256,
245
+ 0.007318696022033691,
246
+ 0.007238057136535645,
247
+ 0.007288617134094238,
248
+ 0.0072535772323608395,
249
+ 0.007251817226409912,
250
+ 0.007270856857299805,
251
+ 0.0073031759262084965,
252
+ 0.007346696853637695,
253
+ 0.007315815925598145,
254
+ 0.0075070161819458005,
255
+ 0.0075399761199951175,
256
+ 0.00742301607131958,
257
+ 0.007451655864715576,
258
+ 0.007400936126708985,
259
+ 0.0075783758163452146,
260
+ 0.0074262161254882815,
261
+ 0.00766509485244751,
262
+ 0.008218053817749024,
263
+ 0.008147813796997071,
264
+ 0.008209094047546387,
265
+ 0.008183653831481934,
266
+ 0.00763933515548706,
267
+ 0.007581736087799072,
268
+ 0.0076770949363708495,
269
+ 0.007733574867248535,
270
+ 0.007989095211029053,
271
+ 0.008277092933654785,
272
+ 0.008252774238586426,
273
+ 0.019490018844604492,
274
+ 0.002522552013397217,
275
+ 0.004225107192993164,
276
+ 0.00820013427734375,
277
+ 0.008145573616027833,
278
+ 0.007606376171112061,
279
+ 0.007687495231628418,
280
+ 0.00770653486251831,
281
+ 0.007633255958557129,
282
+ 0.00780205488204956,
283
+ 0.0077266950607299805,
284
+ 0.007784295082092285,
285
+ 0.007599493980407715
286
  ]
287
  },
288
  "throughput": {
289
  "unit": "samples/s",
290
+ "value": 128.72466648583398
291
  },
292
  "energy": null,
293
  "efficiency": null