IlyasMoutawwakil HF staff commited on
Commit
fbb3ca5
1 Parent(s): ccb53bd

Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_image-classification_google/vit-base-patch16-224",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "image-classification",
9
  "library": "transformers",
@@ -71,12 +71,12 @@
71
  "start_method": "spawn"
72
  },
73
  "environment": {
74
- "cpu": " AMD EPYC 7643 48-Core Processor",
75
- "cpu_count": 96,
76
- "cpu_ram_mb": 1082028.982272,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.15.0-84-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -85,16 +85,16 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.4.0",
88
- "optimum_benchmark_commit": "65fa416fd503cfe9a2be7637ee30c70a4a1f96f1",
89
- "transformers_version": "4.43.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.33.0",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.29.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
- "timm_version": "1.0.8",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 899.444736,
108
- "max_global_vram": 689.364992,
109
- "max_process_vram": 43865.612288,
110
  "max_reserved": 400.556032,
111
  "max_allocated": 346.271744
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 6.88497119140625,
117
- "mean": 6.88497119140625,
118
  "stdev": 0.0,
119
- "p50": 6.88497119140625,
120
- "p90": 6.88497119140625,
121
- "p95": 6.88497119140625,
122
- "p99": 6.88497119140625,
123
  "values": [
124
- 6.88497119140625
125
  ]
126
  },
127
  "throughput": null,
@@ -131,192 +131,192 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1069.453312,
135
- "max_global_vram": 777.691136,
136
- "max_process_vram": 162203.11552,
137
  "max_reserved": 406.847488,
138
- "max_allocated": 355.303424
139
  },
140
  "latency": {
141
  "unit": "s",
142
  "count": 164,
143
- "total": 0.995524206161499,
144
- "mean": 0.006070269549765238,
145
- "stdev": 0.0006449055489030211,
146
- "p50": 0.0058841009140014645,
147
- "p90": 0.006623863363266,
148
- "p95": 0.007850139999389648,
149
- "p99": 0.008218691778182987,
150
  "values": [
151
- 0.005952341079711914,
152
- 0.005876020908355713,
153
- 0.005946901798248291,
154
- 0.005924341201782227,
155
- 0.005972980976104736,
156
- 0.005917942047119141,
157
- 0.0058811411857604984,
158
- 0.005761940956115722,
159
- 0.0057528200149536135,
160
- 0.0057275409698486325,
161
- 0.0056940999031066895,
162
- 0.005669300079345703,
163
- 0.0056269001960754395,
164
- 0.005646580219268798,
165
- 0.005612180233001709,
166
- 0.0056206598281860354,
167
- 0.005597300052642822,
168
- 0.005600339889526367,
169
- 0.005587860107421875,
170
- 0.005573299884796143,
171
- 0.0055729799270629885,
172
- 0.005572019100189209,
173
- 0.0055968198776245115,
174
- 0.005570419788360596,
175
- 0.005566420078277588,
176
- 0.005543379783630371,
177
- 0.005568819999694824,
178
- 0.005578579902648925,
179
- 0.008739871025085449,
180
- 0.005724660873413086,
181
- 0.005691380023956299,
182
- 0.005691860198974609,
183
- 0.005688499927520752,
184
- 0.005788179874420166,
185
- 0.005695061206817627,
186
- 0.005765621185302734,
187
- 0.005979061126708984,
188
- 0.006067701816558838,
189
- 0.006016982078552246,
190
- 0.006151222229003907,
191
- 0.006079541206359863,
192
- 0.006053142070770264,
193
- 0.006041781902313232,
194
- 0.0060245018005371095,
195
- 0.006056500911712647,
196
- 0.0067856240272521975,
197
- 0.00776450777053833,
198
- 0.007917148113250732,
199
- 0.007942108154296875,
200
- 0.007839868068695069,
201
- 0.007897307872772217,
202
- 0.007860348224639893,
203
- 0.007872188091278076,
204
- 0.007851387977600098,
205
- 0.00779666805267334,
206
- 0.00784306812286377,
207
- 0.007810587882995605,
208
- 0.007799228191375732,
209
- 0.007880348205566406,
210
- 0.007629787921905518,
211
- 0.006246421813964844,
212
- 0.006014902114868164,
213
- 0.005963541030883789,
214
- 0.006034901142120361,
215
- 0.0060141019821167,
216
- 0.0059977822303771975,
217
- 0.008689631462097168,
218
- 0.006184182167053223,
219
- 0.006023541927337646,
220
- 0.00603154182434082,
221
- 0.006016981124877929,
222
- 0.006046261787414551,
223
- 0.0060430622100830075,
224
- 0.005998102188110352,
225
- 0.00598914098739624,
226
- 0.006060502052307129,
227
- 0.006207221984863281,
228
- 0.00603938102722168,
229
- 0.006035542011260987,
230
- 0.0060385818481445315,
231
- 0.006038101196289063,
232
- 0.00600338077545166,
233
- 0.006049781799316406,
234
- 0.006032661914825439,
235
- 0.006008660793304444,
236
- 0.006030422210693359,
237
- 0.00602882194519043,
238
- 0.006030422210693359,
239
- 0.006034901142120361,
240
- 0.006001942157745362,
241
- 0.006011702060699463,
242
- 0.006036661148071289,
243
- 0.006046260833740234,
244
- 0.006012182235717773,
245
- 0.0060289821624755855,
246
- 0.00604210090637207,
247
- 0.005893301010131836,
248
- 0.005842260837554932,
249
- 0.005871380805969238,
250
- 0.005768661022186279,
251
- 0.005764180183410644,
252
- 0.005773620128631592,
253
- 0.005752340793609619,
254
- 0.005754580974578857,
255
- 0.0058603410720825195,
256
- 0.005761301040649414,
257
- 0.005789461135864257,
258
- 0.00578722095489502,
259
- 0.005799221038818359,
260
- 0.005768980979919434,
261
- 0.005766260147094727,
262
- 0.005794421195983886,
263
- 0.005754741191864013,
264
- 0.005756021022796631,
265
- 0.005743859767913818,
266
- 0.005775219917297363,
267
- 0.005766261100769043,
268
- 0.005829300880432129,
269
- 0.005874101161956787,
270
- 0.005743061065673828,
271
- 0.005808021068572998,
272
- 0.005759060859680176,
273
- 0.005946260929107666,
274
- 0.0060056219100952145,
275
- 0.005993620872497558,
276
- 0.0059928221702575685,
277
- 0.006016342163085937,
278
- 0.005999220848083496,
279
- 0.005920661926269531,
280
- 0.0058835411071777345,
281
- 0.005927861213684082,
282
- 0.005884020805358887,
283
- 0.005973781108856201,
284
- 0.005915221214294434,
285
- 0.0058272209167480465,
286
- 0.005768180847167969,
287
- 0.005747701168060303,
288
- 0.00576482105255127,
289
- 0.0057692999839782716,
290
- 0.005783860206604004,
291
- 0.0057771410942077635,
292
- 0.005785140991210937,
293
- 0.005755220890045166,
294
- 0.005770579814910888,
295
- 0.0057859401702880855,
296
- 0.005793301105499268,
297
- 0.005766261100769043,
298
- 0.00577138090133667,
299
- 0.0057587409019470214,
300
- 0.0057870597839355465,
301
- 0.0057814598083496095,
302
- 0.005859701156616211,
303
- 0.005859701156616211,
304
- 0.005746739864349365,
305
- 0.005770421028137207,
306
- 0.005796340942382813,
307
- 0.005795701026916504,
308
- 0.005765941143035889,
309
- 0.005808341026306152,
310
- 0.0058982610702514645,
311
- 0.005906421184539795,
312
- 0.005889619827270508,
313
- 0.005884181022644043,
314
- 0.005963380813598633
315
  ]
316
  },
317
  "throughput": {
318
  "unit": "samples/s",
319
- "value": 164.7373303280534
320
  },
321
  "energy": null,
322
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_image-classification_google/vit-base-patch16-224",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "image-classification",
9
  "library": "transformers",
 
71
  "start_method": "spawn"
72
  },
73
  "environment": {
74
+ "cpu": " AMD EPYC 7763 64-Core Processor",
75
+ "cpu_count": 128,
76
+ "cpu_ram_mb": 1082015.256576,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.15.0-101-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.4.0",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.33.0",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.30.1",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
97
+ "timm_version": "1.0.9",
98
  "timm_commit": null,
99
  "peft_version": null,
100
  "peft_commit": null
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1383.40352,
108
+ "max_global_vram": 0.0,
109
+ "max_process_vram": 0.0,
110
  "max_reserved": 400.556032,
111
  "max_allocated": 346.271744
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 7.6373232421875,
117
+ "mean": 7.6373232421875,
118
  "stdev": 0.0,
119
+ "p50": 7.6373232421875,
120
+ "p90": 7.6373232421875,
121
+ "p95": 7.6373232421875,
122
+ "p99": 7.6373232421875,
123
  "values": [
124
+ 7.6373232421875
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1554.47296,
135
+ "max_global_vram": 0.0,
136
+ "max_process_vram": 0.0,
137
  "max_reserved": 406.847488,
138
+ "max_allocated": 354.740224
139
  },
140
  "latency": {
141
  "unit": "s",
142
  "count": 164,
143
+ "total": 0.99387935590744,
144
+ "mean": 0.006060239975045367,
145
+ "stdev": 0.00034437273379574375,
146
+ "p50": 0.005996459007263184,
147
+ "p90": 0.006376425981521607,
148
+ "p95": 0.006418768262863159,
149
+ "p99": 0.006534019680023193,
150
  "values": [
151
+ 0.006515016078948974,
152
+ 0.006229578018188476,
153
+ 0.006443497180938721,
154
+ 0.00656637716293335,
155
+ 0.009208927154541016,
156
+ 0.0064431772232055665,
157
+ 0.006130377769470215,
158
+ 0.005895019054412842,
159
+ 0.006069097995758057,
160
+ 0.00607869815826416,
161
+ 0.006055018901824951,
162
+ 0.005771978855133057,
163
+ 0.005964939117431641,
164
+ 0.006147017955780029,
165
+ 0.005896139144897461,
166
+ 0.005909739017486572,
167
+ 0.006130057811737061,
168
+ 0.005860299110412598,
169
+ 0.005892458915710449,
170
+ 0.0059001388549804685,
171
+ 0.005851658821105957,
172
+ 0.005977098941802979,
173
+ 0.005850059986114502,
174
+ 0.0059014191627502445,
175
+ 0.006074538230895996,
176
+ 0.005891819000244141,
177
+ 0.005837258815765381,
178
+ 0.005991659164428711,
179
+ 0.005762379169464111,
180
+ 0.0058310189247131345,
181
+ 0.00594253921508789,
182
+ 0.0058322992324829105,
183
+ 0.005863179206848144,
184
+ 0.0058572587966918945,
185
+ 0.005874059200286865,
186
+ 0.006143178939819336,
187
+ 0.006079498767852783,
188
+ 0.006067977905273437,
189
+ 0.006289897918701172,
190
+ 0.006133738040924072,
191
+ 0.006231657981872559,
192
+ 0.006242218017578125,
193
+ 0.005974379062652588,
194
+ 0.006132137775421142,
195
+ 0.006240458011627197,
196
+ 0.0060489377975463865,
197
+ 0.006318697929382324,
198
+ 0.0062826981544494625,
199
+ 0.006304458141326904,
200
+ 0.006308298110961914,
201
+ 0.006278538227081299,
202
+ 0.006323816776275634,
203
+ 0.006305257797241211,
204
+ 0.006291337966918945,
205
+ 0.006311017990112305,
206
+ 0.006295816898345947,
207
+ 0.006322378158569336,
208
+ 0.006306057929992676,
209
+ 0.0063161377906799316,
210
+ 0.006350378036499024,
211
+ 0.006309096813201904,
212
+ 0.0063228578567504886,
213
+ 0.006299656867980957,
214
+ 0.006297098159790039,
215
+ 0.006279177188873291,
216
+ 0.006305737972259522,
217
+ 0.006311817169189453,
218
+ 0.00631453800201416,
219
+ 0.006313736915588379,
220
+ 0.006289738178253174,
221
+ 0.006250216960906983,
222
+ 0.005928459167480468,
223
+ 0.006242057800292969,
224
+ 0.005997098922729492,
225
+ 0.0059538989067077636,
226
+ 0.006160778045654297,
227
+ 0.00588477897644043,
228
+ 0.0061975779533386234,
229
+ 0.006221097946166992,
230
+ 0.006077417850494385,
231
+ 0.006350378036499024,
232
+ 0.00620477819442749,
233
+ 0.006144779205322266,
234
+ 0.006089899063110351,
235
+ 0.0058775792121887206,
236
+ 0.006051338195800781,
237
+ 0.00610925817489624,
238
+ 0.005849100112915039,
239
+ 0.005979339122772217,
240
+ 0.005864139080047607,
241
+ 0.0058775792121887206,
242
+ 0.005970219135284424,
243
+ 0.005874699115753174,
244
+ 0.005882538795471191,
245
+ 0.005950539112091064,
246
+ 0.0059159789085388185,
247
+ 0.005755659103393555,
248
+ 0.005995819091796875,
249
+ 0.005808938980102539,
250
+ 0.005947498798370361,
251
+ 0.006060778141021728,
252
+ 0.005829099178314209,
253
+ 0.005856618881225586,
254
+ 0.006213098049163818,
255
+ 0.005945418834686279,
256
+ 0.005962059020996093,
257
+ 0.005838859081268311,
258
+ 0.005853899002075195,
259
+ 0.0061236581802368165,
260
+ 0.005952299118041992,
261
+ 0.006049578189849854,
262
+ 0.005847178936004639,
263
+ 0.005739180088043213,
264
+ 0.005640460014343262,
265
+ 0.005778378963470459,
266
+ 0.005648459911346436,
267
+ 0.0056583800315856934,
268
+ 0.005855659008026123,
269
+ 0.005666860103607178,
270
+ 0.0056479802131652835,
271
+ 0.005883180141448975,
272
+ 0.005611340999603271,
273
+ 0.005637740135192871,
274
+ 0.005826859951019287,
275
+ 0.005640940189361572,
276
+ 0.005643499851226807,
277
+ 0.005950058937072754,
278
+ 0.005868779182434082,
279
+ 0.005706379890441895,
280
+ 0.006008458137512207,
281
+ 0.005776299953460694,
282
+ 0.006388136863708496,
283
+ 0.006396296977996826,
284
+ 0.006383176803588867,
285
+ 0.006369257926940918,
286
+ 0.005948459148406983,
287
+ 0.005642859935760498,
288
+ 0.005852299213409424,
289
+ 0.00562909984588623,
290
+ 0.005706220149993896,
291
+ 0.005683178901672363,
292
+ 0.005700459957122803,
293
+ 0.005934699058532715,
294
+ 0.005673739910125732,
295
+ 0.005637100219726562,
296
+ 0.005878378868103027,
297
+ 0.005626060009002685,
298
+ 0.005838540077209473,
299
+ 0.00593549919128418,
300
+ 0.0060959792137145995,
301
+ 0.005901100158691407,
302
+ 0.006184617996215821,
303
+ 0.006028459072113037,
304
+ 0.005866219043731689,
305
+ 0.006345576763153076,
306
+ 0.00637949800491333,
307
+ 0.006383178234100342,
308
+ 0.0063866968154907226,
309
+ 0.006416457176208496,
310
+ 0.006403656959533691,
311
+ 0.00641917610168457,
312
+ 0.006447017192840576,
313
+ 0.006447816848754883,
314
+ 0.006427337169647216
315
  ]
316
  },
317
  "throughput": {
318
  "unit": "samples/s",
319
+ "value": 165.00996728145472
320
  },
321
  "energy": null,
322
  "efficiency": null