IlyasMoutawwakil HF staff commited on
Commit
1c85739
·
verified ·
1 Parent(s): 931f8eb

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -11,7 +11,7 @@
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
- "device_ids": "0",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
@@ -111,24 +111,24 @@
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
- "max_ram": 1389.883392,
115
- "max_global_vram": 12.120064,
116
- "max_process_vram": 0.0,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
- "total": 10.694306640625,
124
- "mean": 10.694306640625,
125
  "stdev": 0.0,
126
- "p50": 10.694306640625,
127
- "p90": 10.694306640625,
128
- "p95": 10.694306640625,
129
- "p99": 10.694306640625,
130
  "values": [
131
- 10.694306640625
132
  ]
133
  },
134
  "throughput": null,
@@ -138,146 +138,153 @@
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
- "max_ram": 1519.046656,
142
- "max_global_vram": 12.169216,
143
- "max_process_vram": 0.0,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
- "count": 118,
150
- "total": 1.0026790618896484,
151
- "mean": 0.008497280185505494,
152
- "stdev": 0.00010711166726881437,
153
- "p50": 0.008479893684387207,
154
- "p90": 0.00866670913696289,
155
- "p95": 0.008703028774261474,
156
- "p99": 0.008747936124801636,
157
  "values": [
158
- 0.008500773429870606,
159
- 0.0084946928024292,
160
- 0.008568133354187012,
161
- 0.008965250968933105,
162
- 0.008745411872863769,
163
- 0.008681893348693848,
164
- 0.008748453140258789,
165
- 0.008713091850280762,
166
- 0.00871437168121338,
167
- 0.00869021224975586,
168
- 0.008667492866516113,
169
- 0.008592771530151368,
170
- 0.008519172668457031,
171
- 0.00845853328704834,
172
- 0.008454692840576173,
173
- 0.008418853759765626,
174
- 0.008424134254455566,
175
- 0.008494532585144043,
176
- 0.008474852561950684,
177
- 0.008468453407287597,
178
- 0.008443492889404297,
179
- 0.00838973331451416,
180
- 0.00838701343536377,
181
- 0.008363174438476563,
182
- 0.008444293022155762,
183
- 0.008571971893310546,
184
- 0.008581253051757812,
185
- 0.008536932945251464,
186
- 0.008432453155517577,
187
- 0.008525572776794433,
188
- 0.008426372528076171,
189
- 0.008425573348999023,
190
- 0.008423012733459473,
191
- 0.008404772758483886,
192
- 0.008452134132385254,
193
- 0.00842893409729004,
194
- 0.008446212768554687,
195
- 0.00843901252746582,
196
- 0.008416933059692383,
197
- 0.008399012565612794,
198
- 0.008381254196166992,
199
- 0.008403173446655273,
200
- 0.008445412635803222,
201
- 0.008519013404846192,
202
- 0.008426053047180176,
203
- 0.008398053169250488,
204
- 0.008422212600708008,
205
- 0.008404614448547364,
206
- 0.008420774459838867,
207
- 0.008385252952575684,
208
- 0.008435973167419434,
209
- 0.008461413383483886,
210
- 0.008441412925720215,
211
- 0.008391013145446777,
212
- 0.008403654098510743,
213
- 0.00843053436279297,
214
- 0.008369893074035644,
215
- 0.008384773254394531,
216
- 0.008482213020324708,
217
- 0.008632773399353028,
218
- 0.008596772193908692,
219
- 0.008567973136901856,
220
- 0.008484613418579102,
221
- 0.008487492561340333,
222
- 0.008524292945861817,
223
- 0.008458853721618653,
224
- 0.008480294227600098,
225
- 0.008508612632751465,
226
- 0.008642373085021972,
227
- 0.008634692192077636,
228
- 0.008639492988586427,
229
- 0.008657411575317383,
230
- 0.008666373252868653,
231
- 0.008560453414916992,
232
- 0.008521573066711426,
233
- 0.008522533416748046,
234
- 0.008533733367919922,
235
- 0.008527012825012207,
236
- 0.00846717357635498,
237
- 0.008487333297729492,
238
- 0.008441413879394531,
239
- 0.008344293594360352,
240
- 0.008372612953186035,
241
- 0.008366053581237792,
242
- 0.008382694244384765,
243
- 0.008405572891235351,
244
- 0.008353572845458984,
245
- 0.008328133583068848,
246
- 0.008363972663879395,
247
- 0.00835869312286377,
248
- 0.008390374183654785,
249
- 0.008687171936035156,
250
- 0.00868957233428955,
251
- 0.008719653129577636,
252
- 0.008701252937316894,
253
- 0.008557412147521972,
254
- 0.008517413139343262,
255
- 0.008453573226928711,
256
- 0.008437413215637207,
257
- 0.008507013320922852,
258
- 0.008479493141174316,
259
- 0.008471492767333985,
260
- 0.008532933235168457,
261
- 0.00849725341796875,
262
- 0.008530853271484375,
263
- 0.008487492561340333,
264
- 0.008500452995300292,
265
- 0.008497894287109375,
266
- 0.008511173248291015,
267
- 0.008456132888793945,
268
- 0.008474852561950684,
269
- 0.008481093406677246,
270
- 0.008482531547546386,
271
- 0.008446693420410156,
272
- 0.00848045253753662,
273
- 0.008524772644042968,
274
- 0.00859549331665039,
275
- 0.008507972717285156
 
 
 
 
 
 
 
276
  ]
277
  },
278
  "throughput": {
279
  "unit": "samples/s",
280
- "value": 117.684715364074
281
  },
282
  "energy": null,
283
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
11
  "model": "FacebookAI/roberta-base",
12
  "processor": "FacebookAI/roberta-base",
13
  "device": "cuda",
14
+ "device_ids": "4",
15
  "seed": 42,
16
  "inter_op_num_threads": null,
17
  "intra_op_num_threads": null,
 
111
  "load": {
112
  "memory": {
113
  "unit": "MB",
114
+ "max_ram": 1032.974336,
115
+ "max_global_vram": 843.251712,
116
+ "max_process_vram": 48329.367552,
117
  "max_reserved": 555.74528,
118
  "max_allocated": 499.374592
119
  },
120
  "latency": {
121
  "unit": "s",
122
  "count": 1,
123
+ "total": 9.0151884765625,
124
+ "mean": 9.0151884765625,
125
  "stdev": 0.0,
126
+ "p50": 9.0151884765625,
127
+ "p90": 9.0151884765625,
128
+ "p95": 9.0151884765625,
129
+ "p99": 9.0151884765625,
130
  "values": [
131
+ 9.0151884765625
132
  ]
133
  },
134
  "throughput": null,
 
138
  "forward": {
139
  "memory": {
140
  "unit": "MB",
141
+ "max_ram": 1151.393792,
142
+ "max_global_vram": 903.299072,
143
+ "max_process_vram": 243481.837568,
144
  "max_reserved": 555.74528,
145
  "max_allocated": 499.4432
146
  },
147
  "latency": {
148
  "unit": "s",
149
+ "count": 125,
150
+ "total": 0.9961886627674099,
151
+ "mean": 0.007969509302139284,
152
+ "stdev": 0.007743944964142123,
153
+ "p50": 0.007831494808197021,
154
+ "p90": 0.008019397926330567,
155
+ "p95": 0.008041574478149415,
156
+ "p99": 0.008154610786437989,
157
  "values": [
158
+ 0.008008455276489258,
159
+ 0.008075814247131347,
160
+ 0.008017574310302734,
161
+ 0.007979175090789796,
162
+ 0.008028134346008301,
163
+ 0.007969415187835694,
164
+ 0.00817949390411377,
165
+ 0.007851974964141846,
166
+ 0.007816615104675293,
167
+ 0.007785574913024902,
168
+ 0.007766695022583008,
169
+ 0.007797094821929931,
170
+ 0.007811975002288818,
171
+ 0.007773255825042725,
172
+ 0.007819655895233155,
173
+ 0.007773736000061035,
174
+ 0.007773574829101563,
175
+ 0.0077743749618530275,
176
+ 0.007753415107727051,
177
+ 0.00773709487915039,
178
+ 0.007764135837554932,
179
+ 0.0077558159828186035,
180
+ 0.007748455047607422,
181
+ 0.007779815196990967,
182
+ 0.007781575202941895,
183
+ 0.09214866638183594,
184
+ 0.0022166330814361573,
185
+ 0.00237119197845459,
186
+ 0.002261112928390503,
187
+ 0.0022028729915618897,
188
+ 0.002198072910308838,
189
+ 0.0022182331085205077,
190
+ 0.0024436719417572023,
191
+ 0.0022028729915618897,
192
+ 0.0022047929763793946,
193
+ 0.0025163118839263915,
194
+ 0.002317272901535034,
195
+ 0.0022025530338287353,
196
+ 0.003608628034591675,
197
+ 0.007744615077972412,
198
+ 0.007764774799346924,
199
+ 0.007744295120239258,
200
+ 0.007841414928436279,
201
+ 0.00786541509628296,
202
+ 0.0078206148147583,
203
+ 0.007777735233306885,
204
+ 0.007828774929046631,
205
+ 0.007816454887390138,
206
+ 0.007842695236206055,
207
+ 0.007858695030212403,
208
+ 0.00786861515045166,
209
+ 0.007879334926605225,
210
+ 0.007881895065307617,
211
+ 0.007832294940948486,
212
+ 0.007831494808197021,
213
+ 0.007831335067749023,
214
+ 0.007791494846343994,
215
+ 0.007787014961242676,
216
+ 0.007869575023651123,
217
+ 0.00803693389892578,
218
+ 0.00803133487701416,
219
+ 0.007947975158691407,
220
+ 0.007839175224304199,
221
+ 0.007989093780517579,
222
+ 0.00793261480331421,
223
+ 0.007861734867095948,
224
+ 0.007890055179595947,
225
+ 0.007948453903198243,
226
+ 0.008041734695434571,
227
+ 0.007975174903869628,
228
+ 0.00795053482055664,
229
+ 0.00794061517715454,
230
+ 0.00799037504196167,
231
+ 0.008063974380493165,
232
+ 0.008020613670349121,
233
+ 0.007954535007476806,
234
+ 0.007971333980560302,
235
+ 0.007897415161132813,
236
+ 0.007793574810028076,
237
+ 0.00781565523147583,
238
+ 0.007819654941558838,
239
+ 0.007798534870147705,
240
+ 0.007776615142822266,
241
+ 0.007854694843292237,
242
+ 0.007764295101165771,
243
+ 0.007825574874877929,
244
+ 0.0077860550880432125,
245
+ 0.00778733491897583,
246
+ 0.008055174827575683,
247
+ 0.008046854019165038,
248
+ 0.008015494346618652,
249
+ 0.007928935050964356,
250
+ 0.007909894943237305,
251
+ 0.007914374828338623,
252
+ 0.007858695030212403,
253
+ 0.007836295127868651,
254
+ 0.007825415134429931,
255
+ 0.00783757495880127,
256
+ 0.007859014987945556,
257
+ 0.007842374801635741,
258
+ 0.007918213844299317,
259
+ 0.00788941478729248,
260
+ 0.007991654872894287,
261
+ 0.007927334785461426,
262
+ 0.007904294967651368,
263
+ 0.00794061517715454,
264
+ 0.007885094165802003,
265
+ 0.007872135162353515,
266
+ 0.00804093360900879,
267
+ 0.007942375183105468,
268
+ 0.007778055191040039,
269
+ 0.007870695114135743,
270
+ 0.007780934810638428,
271
+ 0.007789414882659912,
272
+ 0.007855655193328858,
273
+ 0.008022053718566895,
274
+ 0.007743015766143799,
275
+ 0.007730535984039307,
276
+ 0.007770374774932861,
277
+ 0.007754694938659668,
278
+ 0.007747654914855957,
279
+ 0.007707975864410401,
280
+ 0.00774477481842041,
281
+ 0.0076959748268127445,
282
+ 0.007734694957733154
283
  ]
284
  },
285
  "throughput": {
286
  "unit": "samples/s",
287
+ "value": 125.47823988756328
288
  },
289
  "energy": null,
290
  "efficiency": null