IlyasMoutawwakil HF staff commited on
Commit
2af7943
·
verified ·
1 Parent(s): 5aec944

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -71,12 +71,12 @@
71
  "start_method": "spawn"
72
  },
73
  "environment": {
74
- "cpu": " AMD EPYC 7643 48-Core Processor",
75
- "cpu_count": 96,
76
- "cpu_ram_mb": 1082028.982272,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.15.0-84-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -85,16 +85,16 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.4.0",
88
- "optimum_benchmark_commit": "65fa416fd503cfe9a2be7637ee30c70a4a1f96f1",
89
- "transformers_version": "4.43.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.33.0",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.29.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
- "timm_version": "1.0.8",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 903.180288,
108
- "max_global_vram": 841.76896,
109
- "max_process_vram": 45891.321856,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.374592
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 7.16346142578125,
117
- "mean": 7.16346142578125,
118
  "stdev": 0.0,
119
- "p50": 7.16346142578125,
120
- "p90": 7.16346142578125,
121
- "p95": 7.16346142578125,
122
- "p99": 7.16346142578125,
123
  "values": [
124
- 7.16346142578125
125
  ]
126
  },
127
  "throughput": null,
@@ -131,170 +131,146 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1016.774656,
135
- "max_global_vram": 896.335872,
136
- "max_process_vram": 210329.690112,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.4432
139
  },
140
  "latency": {
141
  "unit": "s",
142
- "count": 142,
143
- "total": 0.997850569009781,
144
- "mean": 0.007027116683167471,
145
- "stdev": 0.001211544997441332,
146
- "p50": 0.006848674058914185,
147
- "p90": 0.007212708139419556,
148
- "p95": 0.007453437113761902,
149
- "p99": 0.009323262262344369,
150
  "values": [
151
- 0.007285796165466309,
152
- 0.007286116123199463,
153
- 0.007539078235626221,
154
- 0.007598917961120605,
155
- 0.00789875888824463,
156
- 0.007596037864685059,
157
- 0.007508677005767822,
158
- 0.007455397129058838,
159
- 0.007416196823120117,
160
- 0.007222916126251221,
161
- 0.007205795764923095,
162
- 0.0072241959571838375,
163
- 0.007184676170349121,
164
- 0.020005859375,
165
- 0.0022944109439849854,
166
- 0.006813794136047363,
167
- 0.007060675144195556,
168
- 0.007048514842987061,
169
- 0.007077314853668213,
170
- 0.007049795150756836,
171
- 0.01031317138671875,
172
- 0.006991714954376221,
173
- 0.006806114196777344,
174
- 0.006775232791900634,
175
- 0.006813154220581054,
176
- 0.0068313941955566405,
177
- 0.006873953819274903,
178
- 0.006846913814544678,
179
- 0.006824994087219238,
180
- 0.006829954147338867,
181
- 0.006836674213409424,
182
- 0.006849473953247071,
183
- 0.006835073947906494,
184
- 0.006813154220581054,
185
- 0.006838274002075195,
186
- 0.006835554122924805,
187
- 0.006917954921722412,
188
- 0.006923075199127197,
189
- 0.006845794200897217,
190
- 0.006829954147338867,
191
- 0.006834114074707031,
192
- 0.006842275142669677,
193
- 0.006857793807983398,
194
- 0.006835554122924805,
195
- 0.0068304347991943355,
196
- 0.006846434116363526,
197
- 0.006846593856811523,
198
- 0.0068317151069641115,
199
- 0.006867554187774658,
200
- 0.006852513790130616,
201
- 0.006835874080657959,
202
- 0.0068025941848754885,
203
- 0.006919075012207032,
204
- 0.006840514183044434,
205
- 0.00681075382232666,
206
- 0.006842433929443359,
207
- 0.006852993965148926,
208
- 0.006859873771667481,
209
- 0.00687059497833252,
210
- 0.0068459539413452146,
211
- 0.006898594856262207,
212
- 0.006872513771057129,
213
- 0.0067651538848876955,
214
- 0.006792674064636231,
215
- 0.006834115028381348,
216
- 0.0068693141937255855,
217
- 0.00682659387588501,
218
- 0.00684051513671875,
219
- 0.006867713928222656,
220
- 0.006789793968200683,
221
- 0.0068037137985229494,
222
- 0.006863074779510498,
223
- 0.006852354049682617,
224
- 0.006801154136657715,
225
- 0.006850114822387695,
226
- 0.00683859395980835,
227
- 0.006814434051513672,
228
- 0.006847554206848144,
229
- 0.006829154014587402,
230
- 0.006807074069976806,
231
- 0.006856674194335937,
232
- 0.0068027529716491696,
233
- 0.006847393989562988,
234
- 0.006832193851470947,
235
- 0.006810433864593506,
236
- 0.006839875221252441,
237
- 0.006848194122314453,
238
- 0.0068081941604614254,
239
- 0.006861635208129883,
240
- 0.006808834075927734,
241
- 0.006809154033660889,
242
- 0.006849954128265381,
243
- 0.006803073883056641,
244
- 0.0067771530151367184,
245
- 0.006853154182434082,
246
- 0.006772833824157715,
247
- 0.006796673774719238,
248
- 0.00679219388961792,
249
- 0.006785473823547363,
250
- 0.00682659387588501,
251
- 0.006801794052124024,
252
- 0.006792994022369385,
253
- 0.006853154182434082,
254
- 0.006785793781280518,
255
- 0.006819394111633301,
256
- 0.007172036170959473,
257
- 0.007142274856567383,
258
- 0.007151555061340332,
259
- 0.007213476181030273,
260
- 0.007123875141143799,
261
- 0.007201475143432617,
262
- 0.0071803550720214845,
263
- 0.007174435138702392,
264
- 0.0071400351524353025,
265
- 0.007188035011291504,
266
- 0.007355395793914795,
267
- 0.007162115097045898,
268
- 0.007086115837097168,
269
- 0.006934275150299072,
270
- 0.0071497950553894045,
271
- 0.007167715072631836,
272
- 0.007170595169067383,
273
- 0.007192036151885986,
274
- 0.007140995025634766,
275
- 0.0071569948196411135,
276
- 0.006946273803710937,
277
- 0.006876514911651611,
278
- 0.006818433761596679,
279
- 0.006858273983001709,
280
- 0.006780673027038575,
281
- 0.00682435417175293,
282
- 0.006836353778839111,
283
- 0.00690499496459961,
284
- 0.006848834037780762,
285
- 0.006850112915039062,
286
- 0.006806592941284179,
287
- 0.006835554122924805,
288
- 0.006831714153289795,
289
- 0.006844674110412598,
290
- 0.006848514080047607,
291
- 0.006851714134216308,
292
- 0.006853475093841553
293
  ]
294
  },
295
  "throughput": {
296
  "unit": "samples/s",
297
- "value": 142.30587666138626
298
  },
299
  "energy": null,
300
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
71
  "start_method": "spawn"
72
  },
73
  "environment": {
74
+ "cpu": " AMD EPYC 7763 64-Core Processor",
75
+ "cpu_count": 128,
76
+ "cpu_ram_mb": 1082015.256576,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.15.0-101-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.4.0",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.33.0",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.30.1",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
+ "timm_version": "1.0.9",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1386.233856,
108
+ "max_global_vram": 0.0,
109
+ "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.374592
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 8.3549150390625,
117
+ "mean": 8.3549150390625,
118
  "stdev": 0.0,
119
+ "p50": 8.3549150390625,
120
+ "p90": 8.3549150390625,
121
+ "p95": 8.3549150390625,
122
+ "p99": 8.3549150390625,
123
  "values": [
124
+ 8.3549150390625
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1515.548672,
135
+ "max_global_vram": 0.0,
136
+ "max_process_vram": 0.0,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.4432
139
  },
140
  "latency": {
141
  "unit": "s",
142
+ "count": 118,
143
+ "total": 1.00156640958786,
144
+ "mean": 0.008487850928710678,
145
+ "stdev": 0.0010036197467627685,
146
+ "p50": 0.00836661195755005,
147
+ "p90": 0.008726434707641601,
148
+ "p95": 0.009139761924743651,
149
+ "p99": 0.00937003957748413,
150
  "values": [
151
+ 0.008732930183410644,
152
+ 0.008922050476074219,
153
+ 0.009170530319213867,
154
+ 0.009208930015563966,
155
+ 0.009142210006713867,
156
+ 0.009308287620544434,
157
+ 0.008687971115112304,
158
+ 0.008983169555664063,
159
+ 0.008541090965270996,
160
+ 0.00833853244781494,
161
+ 0.008490212440490722,
162
+ 0.008835809707641602,
163
+ 0.008754850387573242,
164
+ 0.008669570922851562,
165
+ 0.008632451057434083,
166
+ 0.00848893165588379,
167
+ 0.008723650932312012,
168
+ 0.008687331199645995,
169
+ 0.008705571174621582,
170
+ 0.008624610900878906,
171
+ 0.008565731048583984,
172
+ 0.01819289970397949,
173
+ 0.004297585010528565,
174
+ 0.008592291831970215,
175
+ 0.008640931129455567,
176
+ 0.008692931175231934,
177
+ 0.00938268756866455,
178
+ 0.00913932991027832,
179
+ 0.008426371574401856,
180
+ 0.008389251708984375,
181
+ 0.008358531951904297,
182
+ 0.008367332458496093,
183
+ 0.008339971542358398,
184
+ 0.00831661319732666,
185
+ 0.008332612037658692,
186
+ 0.008321252822875977,
187
+ 0.008350531578063964,
188
+ 0.008309412002563476,
189
+ 0.008334531784057617,
190
+ 0.008322372436523437,
191
+ 0.008330371856689453,
192
+ 0.008341571807861329,
193
+ 0.008340291976928711,
194
+ 0.00831789207458496,
195
+ 0.008311652183532714,
196
+ 0.00833757209777832,
197
+ 0.008327972412109376,
198
+ 0.008355972290039062,
199
+ 0.008359332084655762,
200
+ 0.008354211807250976,
201
+ 0.008352611541748047,
202
+ 0.008355972290039062,
203
+ 0.008355812072753906,
204
+ 0.00836589241027832,
205
+ 0.008375811576843261,
206
+ 0.0083966121673584,
207
+ 0.0083582124710083,
208
+ 0.008357091903686524,
209
+ 0.008386212348937988,
210
+ 0.008428132057189942,
211
+ 0.008374531745910645,
212
+ 0.008340771675109864,
213
+ 0.008374852180480957,
214
+ 0.008424132347106934,
215
+ 0.0083814115524292,
216
+ 0.008426692008972168,
217
+ 0.008383332252502441,
218
+ 0.008353571891784668,
219
+ 0.008361251831054687,
220
+ 0.008355011940002441,
221
+ 0.008363012313842774,
222
+ 0.008340291976928711,
223
+ 0.00838429069519043,
224
+ 0.00841069221496582,
225
+ 0.008384772300720215,
226
+ 0.00839165210723877,
227
+ 0.008400450706481934,
228
+ 0.00823037338256836,
229
+ 0.008197731971740722,
230
+ 0.007919973850250244,
231
+ 0.00782461404800415,
232
+ 0.008193412780761718,
233
+ 0.00831421184539795,
234
+ 0.008401090621948243,
235
+ 0.00834525203704834,
236
+ 0.008417572021484375,
237
+ 0.008386052131652832,
238
+ 0.00839261245727539,
239
+ 0.008364932060241699,
240
+ 0.008220932960510253,
241
+ 0.008365411758422852,
242
+ 0.008345412254333497,
243
+ 0.008366691589355468,
244
+ 0.008140772819519042,
245
+ 0.008341092109680176,
246
+ 0.008293891906738281,
247
+ 0.008356931686401368,
248
+ 0.00825581169128418,
249
+ 0.008107333183288575,
250
+ 0.008346052169799804,
251
+ 0.008380612373352051,
252
+ 0.008403331756591796,
253
+ 0.008376132011413575,
254
+ 0.008399811744689942,
255
+ 0.00828685188293457,
256
+ 0.008370532035827636,
257
+ 0.008366532325744629,
258
+ 0.008334853172302247,
259
+ 0.008341092109680176,
260
+ 0.008335492134094239,
261
+ 0.008369412422180176,
262
+ 0.008381252288818359,
263
+ 0.008356931686401368,
264
+ 0.008345731735229492,
265
+ 0.008367491722106933,
266
+ 0.00835565185546875,
267
+ 0.00841069221496582,
268
+ 0.008374052047729492
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
269
  ]
270
  },
271
  "throughput": {
272
  "unit": "samples/s",
273
+ "value": 117.81545274522182
274
  },
275
  "energy": null,
276
  "efficiency": null