IlyasMoutawwakil HF staff commited on
Commit
36033e8
·
verified ·
1 Parent(s): 972f1e0

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "5",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1032.507392,
115
  "max_global_vram": 68702.69952,
116
- "max_process_vram": 46363.738112,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 8.8263759765625,
124
- "mean": 8.8263759765625,
125
  "stdev": 0.0,
126
- "p50": 8.8263759765625,
127
- "p90": 8.8263759765625,
128
- "p95": 8.8263759765625,
129
- "p99": 8.8263759765625,
130
  "values": [
131
- 8.8263759765625
132
  ]
133
  },
134
  "throughput": null,
@@ -138,153 +138,143 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1149.76768,
142
  "max_global_vram": 68702.69952,
143
- "max_process_vram": 235458.269184,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 125,
150
- "total": 1.0006552591323856,
151
- "mean": 0.00800524207305908,
152
- "stdev": 0.0011445919658791043,
153
- "p50": 0.00793453311920166,
154
- "p90": 0.00815363597869873,
155
- "p95": 0.008256003189086914,
156
- "p99": 0.008456642456054689,
157
  "values": [
158
- 0.007987331867218018,
159
- 0.007999331951141357,
160
- 0.008110371589660645,
161
- 0.008148772239685059,
162
- 0.00829613208770752,
163
- 0.008259011268615722,
164
- 0.008507329940795899,
165
- 0.008288612365722656,
166
- 0.008196931838989258,
167
- 0.008147332191467285,
168
- 0.008180610656738281,
169
- 0.008055011749267578,
170
- 0.008048771858215332,
171
- 0.007991013050079345,
172
- 0.007979173183441162,
173
- 0.007978052139282226,
174
- 0.007960931777954101,
175
- 0.007921253204345703,
176
- 0.00798173189163208,
177
- 0.007982691764831544,
178
- 0.007977733135223389,
179
- 0.007910212993621827,
180
- 0.007870852947235107,
181
- 0.00785885190963745,
182
- 0.007867012977600098,
183
- 0.007936612129211425,
184
- 0.007941893100738525,
185
- 0.007871332168579102,
186
- 0.007900612831115723,
187
- 0.007901892185211182,
188
- 0.007871973037719726,
189
- 0.007847012996673584,
190
- 0.007863172054290771,
191
- 0.007862372875213623,
192
- 0.01928521156311035,
193
- 0.002356791973114014,
194
- 0.006408778190612793,
195
- 0.007853093147277833,
196
- 0.007868453025817871,
197
- 0.007820292949676514,
198
- 0.007824132919311524,
199
- 0.007918852806091309,
200
- 0.008019171714782714,
201
- 0.00796317195892334,
202
- 0.007908613204956054,
203
- 0.007877253055572509,
204
- 0.007871013164520263,
205
- 0.0078351731300354,
206
- 0.007864452838897706,
207
- 0.007852453231811524,
208
- 0.007819812774658203,
209
- 0.007895172119140625,
210
- 0.007919013023376465,
211
- 0.007846052169799805,
212
- 0.007829573154449463,
213
- 0.007843812942504884,
214
- 0.00785325288772583,
215
- 0.007921092987060547,
216
- 0.007915653228759766,
217
- 0.008020612716674804,
218
- 0.008059653282165527,
219
- 0.007996932983398437,
220
- 0.008005731582641602,
221
- 0.008114212036132813,
222
- 0.008152291297912598,
223
- 0.008137572288513183,
224
- 0.008021092414855957,
225
- 0.008154532432556153,
226
- 0.008176291465759277,
227
- 0.008106371879577638,
228
- 0.008107172012329102,
229
- 0.008071331977844239,
230
- 0.008041571617126465,
231
- 0.008005891799926758,
232
- 0.008034531593322753,
233
- 0.008014052391052247,
234
- 0.008043492317199707,
235
- 0.007928772926330567,
236
- 0.007911811828613282,
237
- 0.007913413047790528,
238
- 0.007980291843414306,
239
- 0.007979812145233155,
240
- 0.008015492439270019,
241
- 0.007921253204345703,
242
- 0.007924451828002929,
243
- 0.007920773029327392,
244
- 0.007894691944122314,
245
- 0.008016133308410644,
246
- 0.008266692161560058,
247
- 0.008263331413269043,
248
- 0.00824397087097168,
249
- 0.008053892135620118,
250
- 0.007958373069763184,
251
- 0.007917252063751221,
252
- 0.007949732780456542,
253
- 0.007932133197784423,
254
- 0.00795229196548462,
255
- 0.007900292873382569,
256
- 0.00792493200302124,
257
- 0.007947972774505615,
258
- 0.007948771953582763,
259
- 0.007901732921600341,
260
- 0.008022051811218261,
261
- 0.007913891792297362,
262
- 0.007895652770996093,
263
- 0.0081705322265625,
264
- 0.008116771697998047,
265
- 0.008021411895751953,
266
- 0.00799757194519043,
267
- 0.007831653118133546,
268
- 0.007849573135375976,
269
- 0.00793453311920166,
270
- 0.0078636531829834,
271
- 0.00785981321334839,
272
- 0.007856933116912842,
273
- 0.007852932929992677,
274
- 0.007888133049011231,
275
- 0.00784013319015503,
276
- 0.007816292762756348,
277
- 0.0078385329246521,
278
- 0.007828452110290528,
279
- 0.007882372856140137,
280
- 0.00794509220123291,
281
- 0.007850852966308594,
282
- 0.00787453317642212
283
  ]
284
  },
285
  "throughput": {
286
  "unit": "samples/s",
287
- "value": 124.91814624387308
288
  },
289
  "energy": null,
290
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1388.060672,
115
  "max_global_vram": 68702.69952,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 10.5855478515625,
124
+ "mean": 10.5855478515625,
125
  "stdev": 0.0,
126
+ "p50": 10.5855478515625,
127
+ "p90": 10.5855478515625,
128
+ "p95": 10.5855478515625,
129
+ "p99": 10.5855478515625,
130
  "values": [
131
+ 10.5855478515625
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1517.735936,
142
  "max_global_vram": 68702.69952,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 115,
150
+ "total": 1.0000426321029665,
151
+ "mean": 0.00869602288785188,
152
+ "stdev": 0.0007176029213401771,
153
+ "p50": 0.008611010551452638,
154
+ "p90": 0.008837728881835937,
155
+ "p95": 0.008944529628753662,
156
+ "p99": 0.00939737615585327,
157
  "values": [
158
+ 0.008838048934936523,
159
+ 0.00868908977508545,
160
+ 0.008853889465332032,
161
+ 0.0087495698928833,
162
+ 0.008975329399108887,
163
+ 0.008725890159606934,
164
+ 0.008778529167175293,
165
+ 0.008762849807739257,
166
+ 0.008837248802185059,
167
+ 0.009398048400878906,
168
+ 0.0093932466506958,
169
+ 0.008693889617919922,
170
+ 0.008655969619750976,
171
+ 0.008646050453186034,
172
+ 0.008640769958496095,
173
+ 0.008660289764404297,
174
+ 0.00859165096282959,
175
+ 0.008552290916442872,
176
+ 0.008573410034179687,
177
+ 0.008498371124267578,
178
+ 0.008571809768676757,
179
+ 0.008536291122436524,
180
+ 0.008566690444946288,
181
+ 0.01595018482208252,
182
+ 0.006864935874938965,
183
+ 0.008561250686645507,
184
+ 0.008521410942077636,
185
+ 0.008538371086120605,
186
+ 0.008543971061706544,
187
+ 0.008517410278320313,
188
+ 0.008546530723571777,
189
+ 0.008633729934692382,
190
+ 0.008549249649047852,
191
+ 0.008715649604797364,
192
+ 0.008628769874572754,
193
+ 0.008580930709838867,
194
+ 0.008577091217041016,
195
+ 0.008584449768066407,
196
+ 0.008534050941467286,
197
+ 0.008519170761108398,
198
+ 0.00856013011932373,
199
+ 0.008522050857543945,
200
+ 0.008562689781188964,
201
+ 0.008643969535827637,
202
+ 0.008668450355529786,
203
+ 0.008742690086364746,
204
+ 0.00859628963470459,
205
+ 0.008644769668579102,
206
+ 0.008629891395568847,
207
+ 0.008611010551452638,
208
+ 0.008607970237731934,
209
+ 0.008631970405578613,
210
+ 0.00854269027709961,
211
+ 0.008582210540771485,
212
+ 0.008821728706359864,
213
+ 0.008685090065002442,
214
+ 0.00866493034362793,
215
+ 0.008690850257873535,
216
+ 0.008754850387573242,
217
+ 0.008698369979858398,
218
+ 0.008619171142578125,
219
+ 0.008652290344238282,
220
+ 0.0087542085647583,
221
+ 0.008983809471130371,
222
+ 0.009020769119262695,
223
+ 0.00888460922241211,
224
+ 0.008775489807128906,
225
+ 0.008788768768310546,
226
+ 0.008859490394592285,
227
+ 0.008849729537963867,
228
+ 0.00874573040008545,
229
+ 0.008931329727172851,
230
+ 0.008781888961791991,
231
+ 0.008757410049438476,
232
+ 0.008781089782714843,
233
+ 0.008718850135803223,
234
+ 0.008809089660644531,
235
+ 0.008797088623046874,
236
+ 0.008635170936584472,
237
+ 0.008666050910949708,
238
+ 0.008587970733642578,
239
+ 0.008517250061035157,
240
+ 0.00851629066467285,
241
+ 0.008535810470581055,
242
+ 0.008470531463623046,
243
+ 0.008491650581359863,
244
+ 0.008496451377868652,
245
+ 0.008452771186828613,
246
+ 0.008526049613952636,
247
+ 0.008438370704650878,
248
+ 0.008439170837402343,
249
+ 0.008457891464233399,
250
+ 0.00847533130645752,
251
+ 0.008485891342163087,
252
+ 0.008743969917297364,
253
+ 0.008791970252990724,
254
+ 0.00866989040374756,
255
+ 0.008669249534606934,
256
+ 0.008643329620361328,
257
+ 0.008550531387329101,
258
+ 0.00858413028717041,
259
+ 0.008511330604553223,
260
+ 0.008484930992126464,
261
+ 0.008469250679016114,
262
+ 0.008580129623413086,
263
+ 0.008507810592651367,
264
+ 0.008546851158142089,
265
+ 0.008551010131835938,
266
+ 0.008522690773010254,
267
+ 0.008553250312805176,
268
+ 0.008508610725402831,
269
+ 0.008534210205078125,
270
+ 0.008470370292663574,
271
+ 0.008543490409851074,
272
+ 0.008483171463012695
 
 
 
 
 
 
 
 
 
 
273
  ]
274
  },
275
  "throughput": {
276
  "unit": "samples/s",
277
+ "value": 114.99509751716204
278
  },
279
  "energy": null,
280
  "efficiency": null