IlyasMoutawwakil HF staff commited on
Commit
c2b7c46
·
verified ·
1 Parent(s): 6baf9ec

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1032.994816,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 43087.44192,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.56529248046875,
124
- "mean": 7.56529248046875,
125
  "stdev": 0.0,
126
- "p50": 7.56529248046875,
127
- "p90": 7.56529248046875,
128
- "p95": 7.56529248046875,
129
- "p99": 7.56529248046875,
130
  "values": [
131
- 7.56529248046875
132
  ]
133
  },
134
  "throughput": null,
@@ -138,163 +138,155 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1150.570496,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 182939.971584,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 135,
150
- "total": 0.9965392861366271,
151
- "mean": 0.007381772489900942,
152
- "stdev": 0.0004266711783163913,
153
- "p50": 0.0072833352088928225,
154
- "p90": 0.007777349185943604,
155
- "p95": 0.008137891674041747,
156
- "p99": 0.008593951301574705,
157
  "values": [
158
- 0.007387175083160401,
159
- 0.007376453876495361,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
160
  0.007601093769073486,
161
- 0.0077748532295227055,
162
- 0.007769573211669922,
163
- 0.008051332473754882,
164
- 0.00779597282409668,
165
- 0.00767181396484375,
166
- 0.0075868539810180664,
167
- 0.007576774120330811,
168
- 0.007502694129943848,
169
- 0.007506693840026855,
170
- 0.007488133907318115,
171
- 0.007471814155578614,
172
- 0.007460134029388428,
173
- 0.007393734931945801,
174
- 0.007408453941345215,
175
- 0.00741629409790039,
176
- 0.007407014846801758,
177
- 0.0074583740234375,
178
- 0.007464774131774902,
179
- 0.0074278140068054195,
180
- 0.007405895233154297,
181
- 0.00736429500579834,
182
- 0.007435654163360596,
183
- 0.007484293937683105,
184
- 0.007611174106597901,
185
- 0.0076786928176879885,
186
- 0.00759869384765625,
187
- 0.007599654197692871,
188
- 0.00878125,
189
- 0.007439174175262451,
190
- 0.006938216209411621,
191
- 0.0069118161201477055,
192
- 0.006947656154632568,
193
- 0.006912775993347168,
194
- 0.006930857181549072,
195
- 0.006925416946411133,
196
- 0.0069361371994018555,
197
- 0.006929576873779297,
198
- 0.006980616092681885,
199
- 0.006947015762329102,
200
- 0.006933095932006836,
201
- 0.006950695991516113,
202
- 0.006945896148681641,
203
- 0.006937416076660157,
204
- 0.006930056095123291,
205
- 0.006939335823059082,
206
- 0.0069575757980346676,
207
- 0.006921895980834961,
208
- 0.006927175998687744,
209
- 0.006935815811157226,
210
- 0.006943336009979248,
211
- 0.006937096118927002,
212
- 0.006967175960540772,
213
- 0.006956456184387207,
214
- 0.00696397590637207,
215
- 0.010600443840026855,
216
- 0.007451333999633789,
217
- 0.00724861478805542,
218
- 0.0072642951011657715,
219
- 0.007255815029144287,
220
- 0.0072556548118591305,
221
- 0.0072950139045715335,
222
- 0.007218695163726806,
223
- 0.0072361350059509276,
224
- 0.007240294933319092,
225
- 0.00729661512374878,
226
- 0.007243814945220947,
227
- 0.007294855117797852,
228
- 0.007305893898010254,
229
- 0.00729661512374878,
230
- 0.00728077507019043,
231
- 0.007295814990997315,
232
- 0.007250054836273194,
233
- 0.00727965497970581,
234
- 0.007255974769592285,
235
- 0.007259815216064453,
236
- 0.007309895038604736,
237
- 0.007230374813079834,
238
- 0.007274855136871338,
239
- 0.007274535179138184,
240
- 0.0072482948303222655,
241
- 0.007488614082336426,
242
- 0.007250374794006348,
243
- 0.007263814926147461,
244
- 0.007276295185089111,
245
- 0.007295814990997315,
246
- 0.007286533832550048,
247
- 0.007266214847564697,
248
- 0.007271975040435791,
249
- 0.0072454152107238766,
250
- 0.007277094841003418,
251
- 0.007260934829711914,
252
- 0.007267495155334472,
253
- 0.0072794947624206545,
254
- 0.0072710151672363285,
255
- 0.007243655204772949,
256
- 0.007231015205383301,
257
- 0.0072623748779296875,
258
- 0.007202374935150147,
259
- 0.007244935035705566,
260
- 0.007215495109558105,
261
- 0.00721853494644165,
262
- 0.007159016132354736,
263
- 0.0072058959007263185,
264
- 0.007249575138092041,
265
- 0.007631653785705566,
266
- 0.008177732467651367,
267
- 0.00816285228729248,
268
- 0.008208771705627442,
269
- 0.007648454189300537,
270
- 0.007952773094177246,
271
- 0.008141251564025878,
272
- 0.008230371475219727,
273
- 0.008136451721191406,
274
- 0.007779013156890869,
275
- 0.007781093120574951,
276
- 0.007515974044799804,
277
- 0.007788133144378662,
278
- 0.007639013767242432,
279
- 0.0076292529106140134,
280
  0.0076198139190673825,
281
- 0.007650534152984619,
282
- 0.007472614765167237,
283
- 0.007310214042663574,
284
- 0.007320135116577148,
285
- 0.007288135051727295,
286
- 0.007219494819641114,
287
- 0.007198854923248291,
288
- 0.007243175029754639,
289
- 0.007332133769989014,
290
- 0.0072833352088928225,
291
- 0.007273254871368408,
292
- 0.007303013801574707
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
293
  ]
294
  },
295
  "throughput": {
296
  "unit": "samples/s",
297
- "value": 135.46881881934286
298
  },
299
  "energy": null,
300
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "multiple-choice",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1387.43808,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 7.89338232421875,
124
+ "mean": 7.89338232421875,
125
  "stdev": 0.0,
126
+ "p50": 7.89338232421875,
127
+ "p90": 7.89338232421875,
128
+ "p95": 7.89338232421875,
129
+ "p99": 7.89338232421875,
130
  "values": [
131
+ 7.89338232421875
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1512.984576,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 127,
150
+ "total": 1.002293522357941,
151
+ "mean": 0.007892074979196383,
152
+ "stdev": 0.0004478581805233672,
153
+ "p50": 0.007641573905944824,
154
+ "p90": 0.008383234596252441,
155
+ "p95": 0.008632177925109863,
156
+ "p99": 0.008667272548675537,
157
  "values": [
158
+ 0.008615489959716796,
159
+ 0.008375971794128419,
160
+ 0.008439332008361817,
161
+ 0.008649090766906738,
162
+ 0.008392610549926759,
163
+ 0.008351011276245117,
164
+ 0.0083530912399292,
165
+ 0.008263651847839355,
166
+ 0.00818957233428955,
167
+ 0.008140771865844727,
168
+ 0.008127971649169922,
169
+ 0.008086531639099122,
170
+ 0.008009411811828614,
171
+ 0.007935012817382812,
172
+ 0.007907012939453124,
173
+ 0.00792477321624756,
174
+ 0.007892292976379394,
175
+ 0.007913413047790528,
176
+ 0.007971333026885987,
177
+ 0.007969411849975586,
178
+ 0.007936613082885742,
179
+ 0.007913893222808837,
180
+ 0.007955172061920167,
181
+ 0.007932612895965576,
182
+ 0.007917733192443847,
183
+ 0.008000131607055663,
184
+ 0.007948933124542236,
185
+ 0.00796125316619873,
186
+ 0.007948132991790771,
187
+ 0.011377080917358399,
188
+ 0.007872453212738037,
189
+ 0.007630373954772949,
190
+ 0.007646852970123291,
191
+ 0.007623814105987549,
192
+ 0.007641573905944824,
193
+ 0.00760685396194458,
194
+ 0.007601893901824951,
195
+ 0.0076023731231689455,
196
+ 0.0076241340637207035,
197
+ 0.0076050939559936526,
198
  0.007601093769073486,
199
+ 0.007607493877410889,
200
+ 0.007597733974456787,
201
+ 0.0076124539375305176,
202
+ 0.007599813938140869,
203
+ 0.007598534107208252,
204
+ 0.0075911741256713865,
205
+ 0.007604134082794189,
206
+ 0.007591494083404541,
207
+ 0.00759261417388916,
208
+ 0.0075935740470886234,
209
+ 0.00759645414352417,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
210
  0.0076198139190673825,
211
+ 0.007618053913116455,
212
+ 0.007595493793487549,
213
+ 0.007591973781585693,
214
+ 0.007594054222106934,
215
+ 0.007569414138793945,
216
+ 0.007590373992919922,
217
+ 0.007586854934692383,
218
+ 0.007935973167419433,
219
+ 0.007942052841186523,
220
+ 0.007929412841796876,
221
+ 0.007595173835754395,
222
+ 0.007596774101257324,
223
+ 0.007614533901214599,
224
+ 0.007609414100646973,
225
+ 0.007615173816680908,
226
+ 0.007602374076843262,
227
+ 0.007608774185180664,
228
+ 0.0075935740470886234,
229
+ 0.0075886139869689945,
230
+ 0.0076081337928771975,
231
+ 0.007593894004821777,
232
+ 0.00795357322692871,
233
+ 0.007979172229766846,
234
+ 0.00799565315246582,
235
+ 0.007980291843414306,
236
+ 0.007614694118499756,
237
+ 0.007613574028015137,
238
+ 0.007636134147644043,
239
+ 0.007619974136352539,
240
+ 0.007623653888702392,
241
+ 0.007602374076843262,
242
+ 0.007583174228668213,
243
+ 0.007582374095916748,
244
+ 0.00756429386138916,
245
+ 0.00758445405960083,
246
+ 0.007584934234619141,
247
+ 0.007604934215545654,
248
+ 0.007578053951263428,
249
+ 0.0075905342102050785,
250
+ 0.007613893985748291,
251
+ 0.007614054203033447,
252
+ 0.007601414203643799,
253
+ 0.0075868539810180664,
254
+ 0.007570693969726563,
255
+ 0.007593093872070312,
256
+ 0.007587974071502686,
257
+ 0.007628293037414551,
258
+ 0.0076052541732788085,
259
+ 0.007616454124450683,
260
+ 0.007596933841705322,
261
+ 0.007762692928314209,
262
+ 0.008091651916503906,
263
+ 0.007959012985229491,
264
+ 0.008117732048034667,
265
+ 0.008074531555175781,
266
+ 0.008115971565246582,
267
+ 0.008013092994689941,
268
+ 0.00793837308883667,
269
+ 0.007778532981872558,
270
+ 0.008077252388000488,
271
+ 0.008097731590270996,
272
+ 0.008087171554565429,
273
+ 0.008344451904296874,
274
+ 0.008651009559631348,
275
+ 0.008670849800109863,
276
+ 0.008389090538024902,
277
+ 0.008293572425842286,
278
+ 0.008593889236450195,
279
+ 0.008639329910278321,
280
+ 0.00865709114074707,
281
+ 0.008307331085205078,
282
+ 0.008450051307678223,
283
+ 0.0083793306350708,
284
+ 0.008655810356140136
285
  ]
286
  },
287
  "throughput": {
288
  "unit": "samples/s",
289
+ "value": 126.70938918294792
290
  },
291
  "energy": null,
292
  "efficiency": null