IlyasMoutawwakil HF staff commited on
Commit
1365031
·
verified ·
1 Parent(s): 4c8783a

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1042.92352,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 54984.814592,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 19.712884765625,
124
- "mean": 19.712884765625,
125
  "stdev": 0.0,
126
- "p50": 19.712884765625,
127
- "p90": 19.712884765625,
128
- "p95": 19.712884765625,
129
- "p99": 19.712884765625,
130
  "values": [
131
- 19.712884765625
132
  ]
133
  },
134
  "throughput": null,
@@ -138,150 +138,168 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1159.913472,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 259819.065344,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 122,
150
- "total": 0.9972260696887969,
151
- "mean": 0.008173984177777026,
152
- "stdev": 0.0014704026266823857,
153
- "p50": 0.008039011001586913,
154
- "p90": 0.008192994689941406,
155
- "p95": 0.008282442331314087,
156
- "p99": 0.01653031551361084,
157
  "values": [
158
- 0.007931491851806641,
159
- 0.007937252044677734,
160
- 0.007930051803588868,
161
- 0.007970692157745362,
162
- 0.007935972213745117,
163
- 0.007938052177429199,
164
- 0.008036131858825684,
165
- 0.008231491088867187,
166
- 0.008031810760498047,
167
- 0.007881092071533204,
168
- 0.016546342849731447,
169
- 0.006491977214813233,
170
- 0.00814397144317627,
171
- 0.008284290313720704,
172
- 0.008183650970458985,
173
- 0.008135811805725097,
174
- 0.00810669231414795,
175
- 0.008358530044555664,
176
- 0.008193250656127929,
177
- 0.00803645133972168,
178
- 0.008224451065063476,
179
- 0.008065892219543456,
180
- 0.008093252182006835,
181
- 0.00820509147644043,
182
- 0.00804189109802246,
183
- 0.008125890731811523,
184
- 0.008063331604003906,
185
- 0.008174370765686036,
186
- 0.008166530609130859,
187
- 0.008169570922851562,
188
- 0.007995971202850342,
189
- 0.007967651844024658,
190
- 0.008014851570129395,
191
- 0.008166211128234863,
192
- 0.008157251358032226,
193
- 0.00804525089263916,
194
- 0.008023652076721191,
195
- 0.00798141098022461,
196
- 0.007965732097625732,
197
- 0.008023491859436035,
198
- 0.008051811218261718,
199
- 0.00805277156829834,
200
- 0.00801421070098877,
201
- 0.008014851570129395,
202
- 0.00796701192855835,
203
- 0.007971971035003663,
204
- 0.008006052017211914,
205
- 0.008065251350402832,
206
- 0.007936771869659423,
207
- 0.007972931861877442,
208
- 0.008049250602722168,
209
- 0.007997891902923583,
210
- 0.007995971202850342,
211
- 0.008019971847534179,
212
- 0.008187810897827148,
213
- 0.007954852104187012,
214
- 0.007974371910095215,
215
- 0.007985732078552247,
216
- 0.007988771915435792,
217
- 0.007968451976776123,
218
- 0.007971012115478515,
219
- 0.007981252193450928,
220
- 0.016470022201538086,
221
- 0.004938222885131836,
222
- 0.008062850952148438,
223
- 0.008107651710510254,
224
- 0.00810349178314209,
225
- 0.008059651374816894,
226
- 0.008152131080627442,
227
- 0.008014052391052247,
228
- 0.008179330825805664,
229
- 0.008136771202087402,
230
- 0.008063490867614747,
231
- 0.008139652252197266,
232
- 0.008065892219543456,
233
- 0.008094691276550292,
234
- 0.008207011222839355,
235
- 0.008457249641418457,
236
- 0.008190690994262695,
237
- 0.0080222110748291,
238
- 0.007943332195281982,
239
- 0.007985251903533936,
240
- 0.008100451469421387,
241
- 0.008082691192626954,
242
- 0.007987971782684327,
243
- 0.00801725196838379,
244
- 0.008058851242065429,
245
- 0.007987331867218018,
246
- 0.008115970611572265,
247
- 0.00811245059967041,
248
- 0.008133570671081542,
249
- 0.008012452125549317,
250
- 0.007996931076049804,
251
- 0.008045891761779785,
252
- 0.008083010673522949,
253
- 0.00807229232788086,
254
- 0.008069572448730468,
255
- 0.008065251350402832,
256
- 0.007985732078552247,
257
- 0.007964931964874267,
258
- 0.007996291160583495,
259
- 0.008016291618347169,
260
- 0.007985251903533936,
261
- 0.00800221061706543,
262
- 0.016646820068359373,
263
- 0.002279031991958618,
264
- 0.006942854881286621,
265
- 0.007959971904754639,
266
- 0.007883011817932129,
267
- 0.008041570663452149,
268
- 0.007877731800079345,
269
- 0.007906211853027343,
270
- 0.007830691814422608,
271
- 0.007923171997070313,
272
- 0.007902852058410645,
273
- 0.008086530685424805,
274
- 0.008190690994262695,
275
- 0.008168451309204102,
276
- 0.007945732116699218,
277
- 0.008042531967163086,
278
- 0.008958847999572753,
279
- 0.008247330665588378
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
280
  ]
281
  },
282
  "throughput": {
283
  "unit": "samples/s",
284
- "value": 122.33936086134646
285
  },
286
  "energy": null,
287
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.0.dev20231010+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 955.78112,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 46141.325312,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 13.48337109375,
124
+ "mean": 13.48337109375,
125
  "stdev": 0.0,
126
+ "p50": 13.48337109375,
127
+ "p90": 13.48337109375,
128
+ "p95": 13.48337109375,
129
+ "p99": 13.48337109375,
130
  "values": [
131
+ 13.48337109375
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1082.687488,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 245274.865664,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 140,
150
+ "total": 0.997159558296204,
151
+ "mean": 0.00712256827354431,
152
+ "stdev": 0.0012613846791237773,
153
+ "p50": 0.006859049081802368,
154
+ "p90": 0.0077194820404052735,
155
+ "p95": 0.007879114079475403,
156
+ "p99": 0.013292702598571775,
157
  "values": [
158
+ 0.007718090057373047,
159
+ 0.007677450180053711,
160
+ 0.00784705114364624,
161
+ 0.007869289875030518,
162
+ 0.007898731231689453,
163
+ 0.007889450073242188,
164
+ 0.007930571079254151,
165
+ 0.007878570079803467,
166
+ 0.008015689849853515,
167
+ 0.007808170795440674,
168
+ 0.007778090000152588,
169
+ 0.007714729785919189,
170
+ 0.007692969799041748,
171
+ 0.007700170040130615,
172
+ 0.015660659790039062,
173
+ 0.002416322946548462,
174
+ 0.0056203269958496095,
175
+ 0.007596330165863037,
176
+ 0.007528490066528321,
177
+ 0.007491209983825684,
178
+ 0.007488490104675293,
179
+ 0.007510250091552735,
180
+ 0.007580009937286377,
181
+ 0.007676170825958252,
182
+ 0.007602089881896972,
183
+ 0.007531529903411866,
184
+ 0.007589930057525634,
185
+ 0.007610249996185303,
186
+ 0.007644010066986084,
187
+ 0.007791050910949707,
188
+ 0.007732009887695313,
189
+ 0.007645770072937012,
190
+ 0.007545450210571289,
191
+ 0.007521130084991455,
192
+ 0.0075513701438903805,
193
+ 0.007517769813537598,
194
+ 0.007559209823608398,
195
+ 0.0075683298110961915,
196
+ 0.007549610137939453,
197
+ 0.007589290142059326,
198
+ 0.006831530094146728,
199
+ 0.006917609214782715,
200
+ 0.006875208854675293,
201
+ 0.007174728870391845,
202
+ 0.0068326501846313475,
203
+ 0.007052488803863525,
204
+ 0.007056649208068848,
205
+ 0.006837928771972656,
206
+ 0.006855370044708252,
207
+ 0.006854729175567627,
208
+ 0.006772968769073486,
209
+ 0.006769769191741944,
210
+ 0.006818408966064453,
211
+ 0.006871368885040283,
212
+ 0.006858088970184326,
213
+ 0.00686000919342041,
214
+ 0.006900649070739746,
215
+ 0.006907370090484619,
216
+ 0.006788808822631836,
217
+ 0.00682512903213501,
218
+ 0.006852649211883545,
219
+ 0.006845609188079834,
220
+ 0.01320993709564209,
221
+ 0.0023496038913726805,
222
+ 0.006172167778015137,
223
+ 0.00691248893737793,
224
+ 0.006624168872833252,
225
+ 0.006649128913879395,
226
+ 0.0066684880256652835,
227
+ 0.006633769035339356,
228
+ 0.0066334490776062014,
229
+ 0.006631369113922119,
230
+ 0.006857449054718017,
231
+ 0.00664832878112793,
232
+ 0.006658888816833496,
233
+ 0.006636329174041748,
234
+ 0.006640967845916748,
235
+ 0.00666896915435791,
236
+ 0.0066523289680480956,
237
+ 0.006611369132995606,
238
+ 0.006647688865661621,
239
+ 0.006630408763885498,
240
+ 0.006644167900085449,
241
+ 0.006707367897033692,
242
+ 0.00664880895614624,
243
+ 0.006649448871612549,
244
+ 0.006638729095458985,
245
+ 0.006658248901367188,
246
+ 0.006638568878173828,
247
+ 0.006637767791748047,
248
+ 0.006651847839355469,
249
+ 0.006681128978729248,
250
+ 0.006672648906707763,
251
+ 0.006662408828735352,
252
+ 0.006883849143981933,
253
+ 0.006636969089508057,
254
+ 0.0066878490447998045,
255
+ 0.006702249050140381,
256
+ 0.0066926488876342775,
257
+ 0.006703208923339843,
258
+ 0.006674248218536377,
259
+ 0.0066427278518676755,
260
+ 0.006658729076385498,
261
+ 0.006763848781585693,
262
+ 0.006694249153137207,
263
+ 0.006648969173431396,
264
+ 0.0065979290008544925,
265
+ 0.006645769119262695,
266
+ 0.006675367832183838,
267
+ 0.006626728057861328,
268
+ 0.006629609107971192,
269
+ 0.006637928962707519,
270
+ 0.006632168769836425,
271
+ 0.006622249126434326,
272
+ 0.006623367786407471,
273
+ 0.006628969192504883,
274
+ 0.006629609107971192,
275
+ 0.006639208793640137,
276
+ 0.007574250221252441,
277
+ 0.0074369702339172365,
278
+ 0.007466090202331543,
279
+ 0.007305290222167969,
280
+ 0.00696432876586914,
281
+ 0.006849769115447998,
282
+ 0.006883368968963623,
283
+ 0.006994568824768067,
284
+ 0.0070846500396728514,
285
+ 0.006935688972473144,
286
+ 0.01334561824798584,
287
+ 0.006623528957366944,
288
+ 0.00754272985458374,
289
+ 0.007104329109191895,
290
+ 0.007040008068084717,
291
+ 0.007076809883117676,
292
+ 0.006943368911743164,
293
+ 0.006990569114685059,
294
+ 0.0071206498146057125,
295
+ 0.0071240091323852535,
296
+ 0.007339049816131591,
297
+ 0.007511209964752197
298
  ]
299
  },
300
  "throughput": {
301
  "unit": "samples/s",
302
+ "value": 140.39879459131996
303
  },
304
  "energy": null,
305
  "efficiency": null