IlyasMoutawwakil HF staff commited on
Commit
752abda
·
verified ·
1 Parent(s): 60f044c

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -97,7 +97,7 @@
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
- "diffusers_version": "0.30.1",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 908.14464,
115
- "max_global_vram": 841.781248,
116
- "max_process_vram": 44127.731712,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.5796474609375,
124
- "mean": 7.5796474609375,
125
  "stdev": 0.0,
126
- "p50": 7.5796474609375,
127
- "p90": 7.5796474609375,
128
- "p95": 7.5796474609375,
129
- "p99": 7.5796474609375,
130
  "values": [
131
- 7.5796474609375
132
  ]
133
  },
134
  "throughput": null,
@@ -138,148 +138,151 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1019.047936,
142
- "max_global_vram": 897.921024,
143
- "max_process_vram": 227310.44864,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 120,
150
- "total": 0.9992206125259401,
151
- "mean": 0.008326838437716167,
152
- "stdev": 0.0012987484932738425,
153
- "p50": 0.008257853031158446,
154
- "p90": 0.008300876712799072,
155
- "p95": 0.008343941116333008,
156
- "p99": 0.01065097829818726,
157
  "values": [
158
- 0.008558414459228515,
159
- 0.010993617057800293,
160
- 0.008572174072265624,
161
- 0.008276172637939453,
162
- 0.008288972854614258,
163
- 0.00827297306060791,
164
- 0.008248493194580079,
165
- 0.009190255165100098,
166
- 0.0082659330368042,
167
- 0.008259213447570801,
168
- 0.008320173263549805,
169
- 0.008269132614135742,
170
- 0.008260333061218261,
171
- 0.008281292915344239,
172
- 0.008295534133911132,
173
- 0.008300812721252441,
174
- 0.0082756929397583,
175
- 0.008077773094177246,
176
- 0.007896652221679687,
177
- 0.008189453125,
178
- 0.008197612762451172,
179
- 0.008192973136901855,
180
- 0.008215052604675293,
181
- 0.00821249294281006,
182
- 0.008232333183288575,
183
- 0.008187692642211914,
184
- 0.008256973266601563,
185
- 0.008199692726135254,
186
- 0.00821793270111084,
187
- 0.008224971771240234,
188
- 0.008191692352294921,
189
- 0.008228013038635254,
190
- 0.008203212738037109,
191
- 0.008227692604064942,
192
- 0.008218572616577149,
193
- 0.008236332893371582,
194
- 0.008224332809448242,
195
- 0.00825985336303711,
196
- 0.008251532554626465,
197
- 0.008251373291015624,
198
- 0.008262252807617188,
199
- 0.008263052940368653,
200
- 0.00820449161529541,
201
- 0.008221293449401855,
202
- 0.008216492652893067,
203
- 0.008202572822570801,
204
- 0.008186412811279296,
205
- 0.008215052604675293,
206
- 0.008241453170776368,
207
- 0.008232333183288575,
208
- 0.00824785327911377,
209
- 0.008293932914733886,
210
- 0.008234092712402344,
211
- 0.008231693267822266,
212
- 0.008249293327331543,
213
- 0.00827729320526123,
214
- 0.008224493026733398,
215
- 0.008201613426208496,
216
- 0.00824785327911377,
217
- 0.008255693435668945,
218
- 0.008299532890319825,
219
- 0.008243852615356446,
220
- 0.00824529266357422,
221
- 0.008236013412475585,
222
- 0.008208012580871582,
223
- 0.008257613182067872,
224
- 0.008271853446960449,
225
- 0.020822912216186523,
226
- 0.0027374439239501954,
227
- 0.005876169204711914,
228
- 0.008298413276672363,
229
- 0.00830145263671875,
230
- 0.00830641269683838,
231
- 0.008268973350524903,
232
- 0.008231852531433105,
233
- 0.008277453422546387,
234
- 0.008284493446350098,
235
- 0.008290733337402344,
236
- 0.008276812553405762,
237
- 0.008264972686767577,
238
- 0.008258893013000488,
239
- 0.008241613388061524,
240
- 0.008276494026184082,
241
- 0.008284652709960937,
242
- 0.008219053268432617,
243
- 0.008286892890930176,
244
- 0.008270893096923828,
245
- 0.00828225326538086,
246
- 0.008277453422546387,
247
- 0.008332653045654297,
248
- 0.00827201271057129,
249
- 0.008273292541503906,
250
- 0.008264172554016113,
251
- 0.008313773155212402,
252
- 0.008290892601013183,
253
- 0.008251852989196777,
254
- 0.008266253471374512,
255
- 0.008238093376159668,
256
- 0.008256173133850098,
257
- 0.0082756929397583,
258
- 0.008231212615966797,
259
- 0.008270092964172363,
260
- 0.008612013816833497,
261
- 0.00831937313079834,
262
- 0.008186573028564454,
263
- 0.008286092758178711,
264
- 0.00826689338684082,
265
- 0.008244173049926758,
266
- 0.008291852951049805,
267
- 0.008264653205871582,
268
- 0.008284812927246093,
269
- 0.008294093132019042,
270
- 0.008241931915283203,
271
- 0.00825505256652832,
272
- 0.00828161334991455,
273
- 0.008233933448791505,
274
- 0.008244492530822755,
275
- 0.008258732795715332,
276
- 0.008228973388671875,
277
- 0.008258092880249023
 
 
 
278
  ]
279
  },
280
  "throughput": {
281
  "unit": "samples/s",
282
- "value": 120.09359944712386
283
  },
284
  "energy": null,
285
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
+ "diffusers_version": "0.30.2",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1389.109248,
115
+ "max_global_vram": 11.145216,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.75903271484375,
124
+ "mean": 7.75903271484375,
125
  "stdev": 0.0,
126
+ "p50": 7.75903271484375,
127
+ "p90": 7.75903271484375,
128
+ "p95": 7.75903271484375,
129
+ "p99": 7.75903271484375,
130
  "values": [
131
+ 7.75903271484375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1514.917888,
142
+ "max_global_vram": 11.149312,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 123,
150
+ "total": 0.9949127125740053,
151
+ "mean": 0.008088721240439066,
152
+ "stdev": 0.00043408992441727116,
153
+ "p50": 0.007873094081878662,
154
+ "p90": 0.008848355102539063,
155
+ "p95": 0.008872802543640135,
156
+ "p99": 0.00891848575592041,
157
  "values": [
158
+ 0.008542692184448242,
159
+ 0.008315652847290038,
160
+ 0.008163812637329102,
161
+ 0.008611010551452638,
162
+ 0.008234692573547362,
163
+ 0.008172133445739745,
164
+ 0.00811469268798828,
165
+ 0.00808253288269043,
166
+ 0.008071332931518556,
167
+ 0.007990533828735352,
168
+ 0.007913894176483154,
169
+ 0.007922853946685791,
170
+ 0.00791853380203247,
171
+ 0.007877254009246826,
172
+ 0.007904454231262207,
173
+ 0.00791469383239746,
174
+ 0.007924613952636718,
175
+ 0.0078820538520813,
176
+ 0.007952293872833253,
177
+ 0.007915174007415772,
178
+ 0.007896933078765869,
179
+ 0.00790925407409668,
180
+ 0.00798237419128418,
181
+ 0.00805389404296875,
182
+ 0.008050373077392578,
183
+ 0.008026212692260741,
184
+ 0.007968133926391601,
185
+ 0.007934693813323974,
186
+ 0.01092284393310547,
187
+ 0.008631011962890626,
188
+ 0.00828989315032959,
189
+ 0.008307332038879395,
190
+ 0.008849571228027344,
191
+ 0.008843490600585938,
192
+ 0.008862050056457519,
193
+ 0.008865891456604004,
194
+ 0.008873570442199706,
195
+ 0.008922850608825684,
196
+ 0.008892930984497071,
197
+ 0.008887649536132813,
198
+ 0.008855010986328125,
199
+ 0.008903010368347168,
200
+ 0.00858205223083496,
201
+ 0.008314533233642578,
202
+ 0.00804381275177002,
203
+ 0.007860774040222169,
204
+ 0.007877893924713135,
205
+ 0.007860774040222169,
206
+ 0.007839014053344726,
207
+ 0.007850214958190918,
208
+ 0.007824934005737305,
209
+ 0.007840933799743653,
210
+ 0.00785517406463623,
211
+ 0.007863014221191407,
212
+ 0.007859653949737548,
213
+ 0.007902853965759278,
214
+ 0.007860454082489013,
215
+ 0.007873894214630127,
216
+ 0.007851974010467529,
217
+ 0.007838533878326417,
218
+ 0.007873094081878662,
219
+ 0.007865893840789794,
220
+ 0.007844933986663818,
221
+ 0.007844453811645507,
222
+ 0.007861093997955322,
223
+ 0.007855494022369385,
224
+ 0.007845253944396972,
225
+ 0.007834854125976563,
226
+ 0.008394051551818847,
227
+ 0.008207972526550294,
228
+ 0.00832893180847168,
229
+ 0.007927494049072266,
230
+ 0.007837093830108643,
231
+ 0.007835494041442872,
232
+ 0.00784973382949829,
233
+ 0.007825095176696778,
234
+ 0.00786621379852295,
235
+ 0.007846854209899902,
236
+ 0.007857093811035156,
237
+ 0.007832454204559327,
238
+ 0.007818213939666748,
239
+ 0.007808773994445801,
240
+ 0.007834854125976563,
241
+ 0.007815494060516357,
242
+ 0.007824453830718994,
243
+ 0.007817093849182128,
244
+ 0.007833573818206787,
245
+ 0.007813253879547118,
246
+ 0.007829894065856933,
247
+ 0.007843334197998047,
248
+ 0.00784653377532959,
249
+ 0.007830214023590088,
250
+ 0.007829894065856933,
251
+ 0.007825414180755615,
252
+ 0.007810853958129883,
253
+ 0.007853894233703614,
254
+ 0.007839815139770508,
255
+ 0.007838213920593261,
256
+ 0.007796453952789307,
257
+ 0.007827653884887695,
258
+ 0.007825573921203614,
259
+ 0.007785254955291748,
260
+ 0.007803335189819336,
261
+ 0.007830214977264405,
262
+ 0.007829575061798096,
263
+ 0.007829093933105468,
264
+ 0.007807013988494873,
265
+ 0.007810054779052734,
266
+ 0.0078025350570678715,
267
+ 0.00784861421585083,
268
+ 0.007817254066467286,
269
+ 0.007829093933105468,
270
+ 0.00785085391998291,
271
+ 0.00785629415512085,
272
+ 0.008366691589355468,
273
+ 0.008855010986328125,
274
+ 0.008859490394592285,
275
+ 0.008322532653808595,
276
+ 0.008437891960144044,
277
+ 0.008773091316223145,
278
+ 0.008886690139770509,
279
+ 0.00871933078765869,
280
+ 0.008241413116455078
281
  ]
282
  },
283
  "throughput": {
284
  "unit": "samples/s",
285
+ "value": 123.62893593125219
286
  },
287
  "energy": null,
288
  "efficiency": null