IlyasMoutawwakil HF staff commited on
Commit
ce2386a
·
verified ·
1 Parent(s): 7e44300

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "6",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -97,7 +97,7 @@
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
- "diffusers_version": "0.30.1",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 908.128256,
115
- "max_global_vram": 841.846784,
116
- "max_process_vram": 40623.259648,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 7.52932861328125,
124
- "mean": 7.52932861328125,
125
  "stdev": 0.0,
126
- "p50": 7.52932861328125,
127
- "p90": 7.52932861328125,
128
- "p95": 7.52932861328125,
129
- "p99": 7.52932861328125,
130
  "values": [
131
- 7.52932861328125
132
  ]
133
  },
134
  "throughput": null,
@@ -138,158 +138,146 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1023.295488,
142
- "max_global_vram": 896.413696,
143
- "max_process_vram": 195502.272512,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 130,
150
- "total": 0.997958024024963,
151
- "mean": 0.00767660018480741,
152
- "stdev": 0.001392735737505683,
153
- "p50": 0.0077181715965271,
154
- "p90": 0.00800095624923706,
155
- "p95": 0.008050756406784059,
156
- "p99": 0.013059410591125453,
157
  "values": [
158
- 0.006945291042327881,
159
- 0.006846089839935302,
160
- 0.007210092067718506,
161
- 0.007150890827178955,
162
- 0.007500010967254639,
163
- 0.007608491897583008,
164
- 0.007941292762756349,
165
- 0.007823212146759033,
166
- 0.008165131568908692,
167
- 0.008097932815551758,
168
- 0.008056013107299805,
169
- 0.008044331550598145,
170
- 0.007990892887115478,
171
- 0.008018571853637696,
172
- 0.007870091915130615,
173
- 0.0076750521659851075,
174
- 0.007863051891326904,
175
- 0.01890418815612793,
176
- 0.0024187240600585936,
177
- 0.0030921640396118165,
178
- 0.007760012149810791,
179
- 0.007760171890258789,
180
- 0.007740491867065429,
181
- 0.0077300920486450195,
182
- 0.007739692211151123,
183
- 0.007742732048034668,
184
- 0.007755531787872315,
185
- 0.007810892105102539,
186
- 0.00784849214553833,
187
- 0.007820012092590332,
188
- 0.0077680120468139645,
189
- 0.007720331192016602,
190
- 0.009096014022827148,
191
- 0.007707372188568116,
192
- 0.007645932197570801,
193
- 0.0076396908760070804,
194
- 0.007633292198181153,
195
- 0.007667212009429931,
196
- 0.007692972183227539,
197
- 0.007683372020721436,
198
- 0.007715211868286133,
199
- 0.007699532032012939,
200
- 0.007756011962890625,
201
- 0.007754571914672851,
202
- 0.007736331939697265,
203
- 0.007723532199859619,
204
- 0.007743691921234131,
205
- 0.007683532238006592,
206
- 0.007693131923675537,
207
- 0.0077134518623352055,
208
- 0.007705932140350342,
209
- 0.007700171947479248,
210
- 0.0077160120010375975,
211
- 0.007725771903991699,
212
- 0.007726731777191162,
213
- 0.007705612182617188,
214
- 0.007713292121887207,
215
- 0.0077160120010375975,
216
- 0.007731691837310791,
217
- 0.007730892181396484,
218
- 0.007727372169494629,
219
- 0.007531851768493652,
220
- 0.006813129901885986,
221
- 0.006924170970916748,
222
- 0.00685313081741333,
223
- 0.006799210071563721,
224
- 0.006829771041870117,
225
- 0.006817609786987305,
226
- 0.006846251010894776,
227
- 0.00683457088470459,
228
- 0.006848330020904541,
229
- 0.0071536111831665035,
230
- 0.0068307299613952635,
231
- 0.00682033109664917,
232
- 0.006826089859008789,
233
- 0.006856811046600342,
234
- 0.006869290828704834,
235
- 0.006825610160827637,
236
- 0.006840970993041992,
237
- 0.0068667311668395995,
238
- 0.006879690170288086,
239
- 0.006824650764465332,
240
- 0.007760811805725098,
241
- 0.0077820920944213865,
242
- 0.007725451946258545,
243
- 0.014678262710571289,
244
- 0.0049688081741333006,
245
- 0.007283851146697998,
246
- 0.0077036919593811036,
247
- 0.0076993718147277835,
248
- 0.0076688117980957034,
249
- 0.00765745210647583,
250
- 0.007656012058258057,
251
- 0.007659532070159912,
252
- 0.007679691791534424,
253
- 0.007696812152862549,
254
- 0.007684971809387207,
255
- 0.007734092235565185,
256
- 0.007715531826019287,
257
- 0.007728012084960938,
258
- 0.007747851848602295,
259
- 0.0077140917778015135,
260
- 0.007706252098083496,
261
- 0.007936811923980713,
262
- 0.007920012950897217,
263
- 0.007956652164459229,
264
- 0.007953773021697998,
265
- 0.007952651977539062,
266
- 0.00793857192993164,
267
- 0.008010892868041993,
268
- 0.007943371772766113,
269
- 0.00799809217453003,
270
- 0.008023852348327637,
271
- 0.00824785327911377,
272
- 0.007999852180480957,
273
- 0.007986732959747314,
274
- 0.007989771842956543,
275
- 0.00799569320678711,
276
- 0.008039212226867676,
277
- 0.007991532802581787,
278
- 0.00797361183166504,
279
- 0.007910572052001952,
280
- 0.007950733184814453,
281
- 0.007946092128753662,
282
- 0.007944171905517577,
283
- 0.008023372650146484,
284
- 0.007913291931152344,
285
- 0.007589292049407959,
286
- 0.0076144118309020995,
287
- 0.0075608110427856446
288
  ]
289
  },
290
  "throughput": {
291
  "unit": "samples/s",
292
- "value": 130.2660000424508
293
  },
294
  "energy": null,
295
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
97
  "transformers_commit": null,
98
  "accelerate_version": "0.33.0",
99
  "accelerate_commit": null,
100
+ "diffusers_version": "0.30.2",
101
  "diffusers_commit": null,
102
  "optimum_version": null,
103
  "optimum_commit": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1389.359104,
115
+ "max_global_vram": 12.070912,
116
+ "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 9.7519384765625,
124
+ "mean": 9.7519384765625,
125
  "stdev": 0.0,
126
+ "p50": 9.7519384765625,
127
+ "p90": 9.7519384765625,
128
+ "p95": 9.7519384765625,
129
+ "p99": 9.7519384765625,
130
  "values": [
131
+ 9.7519384765625
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1518.235648,
142
+ "max_global_vram": 12.156928,
143
+ "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 118,
150
+ "total": 1.0035022182464601,
151
+ "mean": 0.008504256086834406,
152
+ "stdev": 0.0010346381138192993,
153
+ "p50": 0.008405333042144775,
154
+ "p90": 0.008627187728881837,
155
+ "p95": 0.008704540014266968,
156
+ "p99": 0.008969673471450806,
157
  "values": [
158
+ 0.008691971778869628,
159
+ 0.008723172187805175,
160
+ 0.008701251983642579,
161
+ 0.008980771064758301,
162
+ 0.008619973182678223,
163
+ 0.018778820037841797,
164
+ 0.004196146965026856,
165
+ 0.008630211830139161,
166
+ 0.0085657320022583,
167
+ 0.008563013076782226,
168
+ 0.008562691688537597,
169
+ 0.008426053047180176,
170
+ 0.00841981315612793,
171
+ 0.008317092895507812,
172
+ 0.008316292762756347,
173
+ 0.008286213874816895,
174
+ 0.00831981372833252,
175
+ 0.008354533195495605,
176
+ 0.008327973365783692,
177
+ 0.008338692665100098,
178
+ 0.008340132713317872,
179
+ 0.00833997344970703,
180
+ 0.008378533363342286,
181
+ 0.008399012565612794,
182
+ 0.008348133087158203,
183
+ 0.008387493133544923,
184
+ 0.008393893241882324,
185
+ 0.00835677433013916,
186
+ 0.008347973823547364,
187
+ 0.008366052627563476,
188
+ 0.008409893035888672,
189
+ 0.00853453254699707,
190
+ 0.008498533248901367,
191
+ 0.0084313325881958,
192
+ 0.008391813278198242,
193
+ 0.008432772636413574,
194
+ 0.008399653434753417,
195
+ 0.008474212646484375,
196
+ 0.008400293350219726,
197
+ 0.008373252868652344,
198
+ 0.008404932975769042,
199
+ 0.008393253326416016,
200
+ 0.008601411819458008,
201
+ 0.008635811805725098,
202
+ 0.008477732658386231,
203
+ 0.008540452003479004,
204
+ 0.008762051582336425,
205
+ 0.008915491104125976,
206
+ 0.00861021327972412,
207
+ 0.008373252868652344,
208
+ 0.008302693367004394,
209
+ 0.008316452980041505,
210
+ 0.008367012977600098,
211
+ 0.008474533081054687,
212
+ 0.008377893447875977,
213
+ 0.008345892906188965,
214
+ 0.008340292930603028,
215
+ 0.008364612579345704,
216
+ 0.008324133872985839,
217
+ 0.008527973175048827,
218
+ 0.008398853302001952,
219
+ 0.008443013191223144,
220
+ 0.008571011543273926,
221
+ 0.008444453239440918,
222
+ 0.008393412590026855,
223
+ 0.008409893035888672,
224
+ 0.008444613456726074,
225
+ 0.008441252708435059,
226
+ 0.008380773544311524,
227
+ 0.008549893379211426,
228
+ 0.008569092750549316,
229
+ 0.008573573112487793,
230
+ 0.008592612266540527,
231
+ 0.00862589168548584,
232
+ 0.008586213111877442,
233
+ 0.008600452423095702,
234
+ 0.008497093200683594,
235
+ 0.008477412223815917,
236
+ 0.00848253345489502,
237
+ 0.008469093322753906,
238
+ 0.008405733108520507,
239
+ 0.008410853385925293,
240
+ 0.008387972831726074,
241
+ 0.008371012687683105,
242
+ 0.008288772583007812,
243
+ 0.008361892700195312,
244
+ 0.008319812774658203,
245
+ 0.008361092567443848,
246
+ 0.008340933799743652,
247
+ 0.00837837314605713,
248
+ 0.00835965347290039,
249
+ 0.008392132759094239,
250
+ 0.008724612236022949,
251
+ 0.008693092346191407,
252
+ 0.008651171684265137,
253
+ 0.008536131858825683,
254
+ 0.008578372955322265,
255
+ 0.008526371955871583,
256
+ 0.008492293357849122,
257
+ 0.0084596529006958,
258
+ 0.008343812942504882,
259
+ 0.00837229347229004,
260
+ 0.008366852760314942,
261
+ 0.008336612701416016,
262
+ 0.008353572845458984,
263
+ 0.008396773338317872,
264
+ 0.008375172615051269,
265
+ 0.008387493133544923,
266
+ 0.008335653305053712,
267
+ 0.008363013267517089,
268
+ 0.00838045310974121,
269
+ 0.008393732070922852,
270
+ 0.00844109344482422,
271
+ 0.008413573265075684,
272
+ 0.008435973167419434,
273
+ 0.008416132926940918,
274
+ 0.008425413131713867,
275
+ 0.008392932891845704
 
 
 
 
 
 
 
 
 
 
 
 
276
  ]
277
  },
278
  "throughput": {
279
  "unit": "samples/s",
280
+ "value": 117.58818052858477
281
  },
282
  "energy": null,
283
  "efficiency": null