IlyasMoutawwakil HF staff commited on
Commit
bd30878
·
verified ·
1 Parent(s): 321e6dd

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -73,10 +73,10 @@
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 128,
76
- "cpu_ram_mb": 1082015.256576,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.15.0-101-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -91,9 +91,9 @@
91
  ],
92
  "gpu_count": 8,
93
  "gpu_vram_mb": 549621596160,
94
- "optimum_benchmark_version": "0.4.0",
95
  "optimum_benchmark_commit": null,
96
- "transformers_version": "4.44.2",
97
  "transformers_commit": null,
98
  "accelerate_version": "0.34.2",
99
  "accelerate_commit": null,
@@ -103,7 +103,7 @@
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
- "peft_version": "0.12.0",
107
  "peft_commit": null
108
  }
109
  },
@@ -111,25 +111,26 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1078.349824,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 50352.324608,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
122
- "count": 1,
123
- "total": 10.4725830078125,
124
- "mean": 10.4725830078125,
125
- "stdev": 0.0,
126
- "p50": 10.4725830078125,
127
- "p90": 10.4725830078125,
128
- "p95": 10.4725830078125,
129
- "p99": 10.4725830078125,
130
  "values": [
131
- 10.4725830078125
132
- ]
 
 
 
 
 
 
 
 
 
133
  },
134
  "throughput": null,
135
  "energy": null,
@@ -138,145 +139,164 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1196.015616,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 263966.306304,
144
  "max_reserved": 555.74528,
145
- "max_allocated": 499.5072
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 117,
150
- "total": 0.9964194974899292,
151
- "mean": 0.008516405961452388,
152
- "stdev": 0.0012008642340843234,
153
- "p50": 0.008452138900756835,
154
- "p90": 0.008643403434753418,
155
- "p95": 0.00879668254852295,
156
- "p99": 0.014891394386291525,
157
  "values": [
158
- 0.008358379364013672,
159
- 0.008331339836120605,
160
- 0.008399979591369629,
161
- 0.008558539390563965,
162
- 0.008539018630981445,
163
- 0.008523818969726562,
164
- 0.009800135612487793,
165
- 0.00860637855529785,
166
- 0.008530059814453124,
167
- 0.008549739837646484,
168
- 0.008510218620300294,
169
- 0.008570219039916992,
170
- 0.00855069923400879,
171
- 0.00847757911682129,
172
- 0.008595979690551758,
173
- 0.008475339889526368,
174
- 0.008538538932800294,
175
- 0.008474059104919433,
176
- 0.00846205997467041,
177
- 0.008422859191894531,
178
- 0.008503019332885742,
179
- 0.009990376472473144,
180
- 0.008589899063110352,
181
- 0.008393899917602539,
182
- 0.00839181900024414,
183
- 0.008319339752197265,
184
- 0.008334380149841308,
185
- 0.008342378616333008,
186
- 0.008275659561157226,
187
- 0.008506218910217285,
188
- 0.008584939002990722,
189
- 0.008588780403137207,
190
- 0.00867405891418457,
191
- 0.008580458641052247,
192
- 0.00856029987335205,
193
- 0.00850685977935791,
194
- 0.008493899345397949,
195
- 0.008489579200744628,
196
- 0.008456780433654785,
197
- 0.008570219039916992,
198
- 0.00862477970123291,
199
- 0.008792458534240722,
200
- 0.008813578605651856,
201
- 0.00867133903503418,
202
- 0.015824921607971193,
203
- 0.0034425520896911622,
204
- 0.008311819076538086,
205
- 0.00837246036529541,
206
- 0.008364459991455079,
207
- 0.008389419555664062,
208
- 0.008245579719543457,
209
- 0.008274539947509766,
210
- 0.008195659637451173,
211
- 0.008307180404663085,
212
- 0.008265899658203126,
213
- 0.008248140335083008,
214
- 0.008394539833068848,
215
- 0.008452138900756835,
216
- 0.008386380195617675,
217
- 0.008419178962707519,
218
- 0.008732619285583497,
219
- 0.009074698448181153,
220
- 0.008711178779602051,
221
- 0.008572460174560547,
222
- 0.008447978973388672,
223
- 0.008405579566955566,
224
- 0.008435500144958497,
225
- 0.008503819465637207,
226
- 0.00840365982055664,
227
- 0.008389418601989745,
228
- 0.008393420219421386,
229
- 0.008467819213867188,
230
- 0.008546698570251464,
231
- 0.008749098777770995,
232
- 0.008588938713073731,
233
- 0.00855181884765625,
234
- 0.008486220359802246,
235
- 0.00851006031036377,
236
- 0.008534699440002442,
237
- 0.008469578742980956,
238
- 0.008547019004821778,
239
- 0.008405579566955566,
240
- 0.01640523910522461,
241
- 0.004014870166778564,
242
- 0.008542219161987304,
243
- 0.00853853988647461,
244
- 0.008374858856201171,
245
- 0.0083185396194458,
246
- 0.0083087797164917,
247
- 0.008290378570556641,
248
- 0.008266539573669433,
249
- 0.00837950038909912,
250
- 0.008488939285278321,
251
- 0.008489739418029786,
252
- 0.008474860191345215,
253
- 0.008374698638916015,
254
- 0.008293740272521973,
255
- 0.008300620079040527,
256
- 0.008398538589477538,
257
- 0.008294219970703126,
258
- 0.008299019813537598,
259
- 0.008310059547424316,
260
- 0.008506539344787598,
261
- 0.008467979431152343,
262
- 0.008307819366455077,
263
- 0.008244460105895995,
264
- 0.008272780418395995,
265
- 0.008291339874267578,
266
- 0.008272939682006836,
267
- 0.008316458702087403,
268
- 0.008363018989562989,
269
- 0.008314379692077637,
270
- 0.008164620399475098,
271
- 0.008196940422058105,
272
- 0.008123660087585449,
273
- 0.008117580413818359,
274
- 0.008142380714416505
275
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276
  },
277
  "throughput": {
278
  "unit": "samples/s",
279
- "value": 117.42042412330706
280
  },
281
  "energy": null,
282
  "efficiency": null
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
73
  "environment": {
74
  "cpu": " AMD EPYC 7763 64-Core Processor",
75
  "cpu_count": 128,
76
+ "cpu_ram_mb": 1082014.482432,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.15.0-122-generic-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
91
  ],
92
  "gpu_count": 8,
93
  "gpu_vram_mb": 549621596160,
94
+ "optimum_benchmark_version": "0.5.0",
95
  "optimum_benchmark_commit": null,
96
+ "transformers_version": "4.45.1",
97
  "transformers_commit": null,
98
  "accelerate_version": "0.34.2",
99
  "accelerate_commit": null,
 
103
  "optimum_commit": null,
104
  "timm_version": "1.0.9",
105
  "timm_commit": null,
106
+ "peft_version": "0.13.0",
107
  "peft_commit": null
108
  }
109
  },
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1105.55136,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 3736.322048,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.37152
119
  },
120
  "latency": {
121
  "unit": "s",
 
 
 
 
 
 
 
 
122
  "values": [
123
+ 0.06804940795898437
124
+ ],
125
+ "count": 1,
126
+ "total": 0.06804940795898437,
127
+ "mean": 0.06804940795898437,
128
+ "p50": 0.06804940795898437,
129
+ "p90": 0.06804940795898437,
130
+ "p95": 0.06804940795898437,
131
+ "p99": 0.06804940795898437,
132
+ "stdev": 0,
133
+ "stdev_": 0
134
  },
135
  "throughput": null,
136
  "energy": null,
 
139
  "forward": {
140
  "memory": {
141
  "unit": "MB",
142
+ "max_ram": 1223.622656,
143
  "max_global_vram": 68702.69952,
144
+ "max_process_vram": 267460.481024,
145
  "max_reserved": 555.74528,
146
+ "max_allocated": 499.506688
147
  },
148
  "latency": {
149
  "unit": "s",
 
 
 
 
 
 
 
 
150
  "values": [
151
+ 0.007245132923126221,
152
+ 0.006978892803192139,
153
+ 0.006911052227020264,
154
+ 0.006957771778106689,
155
+ 0.0069238519668579105,
156
+ 0.006899532794952393,
157
+ 0.006889613151550293,
158
+ 0.016596830368041993,
159
+ 0.006828173160552979,
160
+ 0.006915213108062744,
161
+ 0.006909933090209961,
162
+ 0.0068904128074646,
163
+ 0.0068956918716430666,
164
+ 0.0069001722335815426,
165
+ 0.006885773181915284,
166
+ 0.006907853126525879,
167
+ 0.006883693218231201,
168
+ 0.007276174068450928,
169
+ 0.007070892810821533,
170
+ 0.006907692909240723,
171
+ 0.0068918528556823734,
172
+ 0.00691329288482666,
173
+ 0.006904812812805175,
174
+ 0.006908651828765869,
175
+ 0.006914093017578125,
176
+ 0.006956973075866699,
177
+ 0.006898892879486084,
178
+ 0.006891053199768066,
179
+ 0.0069128131866455075,
180
+ 0.006923693180084229,
181
+ 0.006834893226623535,
182
+ 0.0071041731834411625,
183
+ 0.0068992128372192384,
184
+ 0.006935853004455567,
185
+ 0.0069102530479431155,
186
+ 0.006915533065795899,
187
+ 0.006915052890777588,
188
+ 0.006924492835998535,
189
+ 0.006902573108673096,
190
+ 0.006897613048553467,
191
+ 0.006905612945556641,
192
+ 0.006920332908630371,
193
+ 0.0069102530479431155,
194
+ 0.00690145206451416,
195
+ 0.0069172921180725095,
196
+ 0.006889613151550293,
197
+ 0.006916333198547364,
198
+ 0.006937932968139648,
199
+ 0.006896492958068848,
200
+ 0.006889933109283447,
201
+ 0.007072173118591309,
202
+ 0.00696401309967041,
203
+ 0.006921613216400146,
204
+ 0.0069102530479431155,
205
+ 0.006916812896728516,
206
+ 0.007013613224029541,
207
+ 0.0069321727752685545,
208
+ 0.0069054532051086425,
209
+ 0.006987853050231934,
210
+ 0.006979692935943604,
211
+ 0.00694769287109375,
212
+ 0.006891693115234375,
213
+ 0.0069555330276489255,
214
+ 0.007356974124908447,
215
+ 0.007895053863525391,
216
+ 0.007885454177856445,
217
+ 0.007913935184478759,
218
+ 0.007393933773040771,
219
+ 0.007446094036102295,
220
+ 0.007863375186920166,
221
+ 0.008727696418762207,
222
+ 0.010171539306640625,
223
+ 0.007526894092559814,
224
+ 0.007621774196624756,
225
+ 0.008070734977722168,
226
+ 0.008190735816955567,
227
+ 0.008064174652099609,
228
+ 0.008054414749145507,
229
+ 0.008120815277099609,
230
+ 0.007964334964752197,
231
+ 0.008109455108642577,
232
+ 0.00809585475921631,
233
+ 0.008022894859313965,
234
+ 0.008092174530029297,
235
+ 0.008063215255737305,
236
+ 0.007843053817749024,
237
+ 0.007609774112701416,
238
+ 0.007667054176330566,
239
+ 0.007640493869781494,
240
+ 0.007657773971557617,
241
+ 0.007538895130157471,
242
+ 0.007552013874053955,
243
+ 0.007388652801513672,
244
+ 0.007175053119659424,
245
+ 0.007241292953491211,
246
+ 0.007391054153442383,
247
+ 0.007162572860717774,
248
+ 0.007117612838745117,
249
+ 0.007217294216156006,
250
+ 0.007178893089294433,
251
+ 0.007258254051208496,
252
+ 0.007111372947692871,
253
+ 0.007098573207855225,
254
+ 0.007122414112091064,
255
+ 0.007106573104858399,
256
+ 0.007282733917236328,
257
+ 0.007555694103240967,
258
+ 0.007251052856445312,
259
+ 0.00818113613128662,
260
+ 0.00759489393234253,
261
+ 0.007807534217834472,
262
+ 0.007092012882232666,
263
+ 0.007161932945251465,
264
+ 0.007818574905395509,
265
+ 0.008147054672241211,
266
+ 0.008197455406188965,
267
+ 0.007712015151977539,
268
+ 0.0072019329071044925,
269
+ 0.007271374225616455,
270
+ 0.007198893070220948,
271
+ 0.007235214233398437,
272
+ 0.007362093925476074,
273
+ 0.00805745506286621,
274
+ 0.007588493824005127,
275
+ 0.007526254177093506,
276
+ 0.007900654792785644,
277
+ 0.008128974914550781,
278
+ 0.007769773960113526,
279
+ 0.007205613136291504,
280
+ 0.007533294200897217,
281
+ 0.007550253868103028,
282
+ 0.007552013874053955,
283
+ 0.007144333839416504,
284
+ 0.007452493190765381,
285
+ 0.007556814193725586
286
+ ],
287
+ "count": 135,
288
+ "total": 0.9972821750640869,
289
+ "mean": 0.007387275370845088,
290
+ "p50": 0.007161932945251465,
291
+ "p90": 0.008063790893554688,
292
+ "p95": 0.00813439884185791,
293
+ "p99": 0.009680632724761959,
294
+ "stdev": 0.0009399251307025448,
295
+ "stdev_": 12.72356969948745
296
  },
297
  "throughput": {
298
  "unit": "samples/s",
299
+ "value": 135.36790627118617
300
  },
301
  "energy": null,
302
  "efficiency": null