IlyasMoutawwakil HF staff commited on
Commit
90c8db5
1 Parent(s): 344d165

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.0+rocm6.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
@@ -104,24 +104,24 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1385.537536,
108
- "max_global_vram": 0.0,
109
- "max_process_vram": 0.0,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.374592
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 9.9135771484375,
117
- "mean": 9.9135771484375,
118
  "stdev": 0.0,
119
- "p50": 9.9135771484375,
120
- "p90": 9.9135771484375,
121
- "p95": 9.9135771484375,
122
- "p99": 9.9135771484375,
123
  "values": [
124
- 9.9135771484375
125
  ]
126
  },
127
  "throughput": null,
@@ -131,144 +131,151 @@
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
- "max_ram": 1514.389504,
135
- "max_global_vram": 0.0,
136
- "max_process_vram": 0.0,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.4432
139
  },
140
  "latency": {
141
  "unit": "s",
142
- "count": 116,
143
- "total": 0.9882716417312623,
144
- "mean": 0.008519583118372951,
145
- "stdev": 0.004678037532035842,
146
- "p50": 0.0083514142036438,
147
- "p90": 0.01220212173461914,
148
- "p95": 0.015619991064071656,
149
- "p99": 0.029196933746337878,
150
  "values": [
151
- 0.002330233097076416,
152
- 0.007305576801300049,
153
- 0.008930051803588867,
154
- 0.029582948684692384,
155
- 0.013769717216491699,
156
- 0.0022492730617523194,
157
- 0.011942362785339356,
158
- 0.012329401016235352,
159
- 0.002308633089065552,
160
- 0.0022367930412292483,
161
- 0.004148306846618652,
162
- 0.009434370994567872,
163
- 0.008943653106689453,
164
- 0.008872293472290039,
165
- 0.015518991470336915,
166
- 0.0025353519916534424,
167
- 0.008663172721862794,
168
- 0.008606533050537109,
169
- 0.008599972724914551,
170
- 0.008503172874450684,
171
- 0.012382681846618653,
172
- 0.005279822826385498,
173
- 0.017308425903320313,
174
- 0.013436758041381835,
175
- 0.011740283966064453,
176
- 0.012074842453002929,
177
- 0.010900925636291504,
178
- 0.0022942330837249754,
179
- 0.0022139129638671876,
180
- 0.002212472915649414,
181
- 0.016354028701782226,
182
- 0.002357273101806641,
183
- 0.005990060806274414,
184
- 0.013484758377075195,
185
- 0.007267656803131103,
186
- 0.0270095157623291,
187
- 0.002320152997970581,
188
- 0.002237592935562134,
189
- 0.0022159929275512697,
190
- 0.0049633440971374515,
191
- 0.00839613437652588,
192
- 0.008589253425598145,
193
- 0.008338374137878418,
194
- 0.01592298984527588,
195
- 0.002286552906036377,
196
- 0.007103178024291992,
197
- 0.00864205265045166,
198
- 0.03391141510009766,
199
- 0.0022713530063629152,
200
- 0.00221647310256958,
201
- 0.002210072994232178,
202
- 0.002201112985610962,
203
- 0.008167975425720215,
204
- 0.008337894439697265,
205
- 0.008351814270019532,
206
- 0.008343013763427734,
207
- 0.00833005428314209,
208
- 0.008346054077148437,
209
- 0.008343974113464355,
210
- 0.008364933967590332,
211
- 0.008476452827453613,
212
- 0.008471653938293458,
213
- 0.008467493057250976,
214
- 0.008487813949584961,
215
- 0.008425732612609864,
216
- 0.008422213554382324,
217
- 0.008394054412841797,
218
- 0.008337254524230958,
219
- 0.008351014137268067,
220
- 0.008339333534240722,
221
- 0.008369573593139648,
222
- 0.008339174270629883,
223
- 0.008347654342651367,
224
- 0.008335654258728026,
225
- 0.008313413619995116,
226
- 0.008323493957519532,
227
- 0.008323174476623535,
228
- 0.008358213424682617,
229
- 0.008324133872985839,
230
- 0.008388934135437011,
231
- 0.00831373405456543,
232
- 0.008357574462890625,
233
- 0.008319173812866211,
234
- 0.008336773872375488,
235
- 0.008298213958740235,
236
- 0.008312454223632812,
237
- 0.00833885383605957,
238
- 0.008324773788452148,
239
- 0.008329093933105469,
240
- 0.008354534149169921,
241
- 0.008344453811645508,
242
- 0.008327974319458007,
243
- 0.008308294296264649,
244
- 0.00831965446472168,
245
- 0.008346214294433593,
246
- 0.008333574295043946,
247
- 0.008358373641967773,
248
- 0.008313894271850586,
249
- 0.008353254318237305,
250
- 0.008319173812866211,
251
- 0.008340933799743652,
252
- 0.008406534194946289,
253
- 0.008359494209289551,
254
- 0.00837389373779297,
255
- 0.008362533569335938,
256
- 0.008377734184265137,
257
- 0.00839661407470703,
258
- 0.00840429401397705,
259
- 0.008418534278869629,
260
- 0.008498373985290528,
261
- 0.008541893005371094,
262
- 0.008448613166809083,
263
- 0.008450693130493164,
264
- 0.00838957405090332,
265
- 0.008372453689575196,
266
- 0.008392773628234863
 
 
 
 
 
 
 
267
  ]
268
  },
269
  "throughput": {
270
  "unit": "samples/s",
271
- "value": 117.3766352303606
272
  },
273
  "energy": null,
274
  "efficiency": null
 
3
  "name": "cuda_inference_transformers_text-classification_FacebookAI/roberta-base",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-classification",
9
  "library": "transformers",
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 905.0112,
108
+ "max_global_vram": 841.773056,
109
+ "max_process_vram": 50817.110016,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.374592
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 8.356474609375,
117
+ "mean": 8.356474609375,
118
  "stdev": 0.0,
119
+ "p50": 8.356474609375,
120
+ "p90": 8.356474609375,
121
+ "p95": 8.356474609375,
122
+ "p99": 8.356474609375,
123
  "values": [
124
+ 8.356474609375
125
  ]
126
  },
127
  "throughput": null,
 
131
  "forward": {
132
  "memory": {
133
  "unit": "MB",
134
+ "max_ram": 1020.080128,
135
+ "max_global_vram": 896.344064,
136
+ "max_process_vram": 264395.65312,
137
  "max_reserved": 555.74528,
138
  "max_allocated": 499.4432
139
  },
140
  "latency": {
141
  "unit": "s",
142
+ "count": 123,
143
+ "total": 0.9992745399475101,
144
+ "mean": 0.008124183251605771,
145
+ "stdev": 9.251184838970667e-05,
146
+ "p50": 0.008123177528381348,
147
+ "p90": 0.008261737632751464,
148
+ "p95": 0.008282729148864746,
149
+ "p99": 0.008328210811614991,
150
  "values": [
151
+ 0.008174057960510253,
152
+ 0.008050857543945312,
153
+ 0.007994217872619629,
154
+ 0.007998698234558106,
155
+ 0.007989098072052002,
156
+ 0.007966698169708252,
157
+ 0.007997578144073486,
158
+ 0.008005098342895508,
159
+ 0.008239176750183106,
160
+ 0.007993257999420167,
161
+ 0.007997097969055176,
162
+ 0.007998857975006103,
163
+ 0.00794333791732788,
164
+ 0.007962859153747559,
165
+ 0.00796301794052124,
166
+ 0.007941577911376953,
167
+ 0.007982217788696289,
168
+ 0.008011817932128906,
169
+ 0.007996297836303711,
170
+ 0.007986217975616454,
171
+ 0.007991017818450928,
172
+ 0.0080492582321167,
173
+ 0.008041257858276368,
174
+ 0.008113738059997558,
175
+ 0.008155497550964355,
176
+ 0.008138378143310547,
177
+ 0.008103657722473144,
178
+ 0.008083658218383789,
179
+ 0.008093418121337891,
180
+ 0.00813453769683838,
181
+ 0.00826237678527832,
182
+ 0.008243657112121583,
183
+ 0.008296458244323731,
184
+ 0.008181577682495117,
185
+ 0.008110857963562012,
186
+ 0.00801437759399414,
187
+ 0.008103977203369141,
188
+ 0.008063657760620117,
189
+ 0.00806845760345459,
190
+ 0.008076778411865234,
191
+ 0.008054377555847168,
192
+ 0.008172138214111328,
193
+ 0.008137738227844239,
194
+ 0.00810061740875244,
195
+ 0.008127338409423828,
196
+ 0.0080623779296875,
197
+ 0.008049737930297852,
198
+ 0.00805965805053711,
199
+ 0.008112616539001465,
200
+ 0.0082086181640625,
201
+ 0.008189256668090821,
202
+ 0.008261737823486328,
203
+ 0.00821901798248291,
204
+ 0.008167178153991698,
205
+ 0.008146858215332032,
206
+ 0.008150057792663574,
207
+ 0.008184456825256347,
208
+ 0.008181258201599121,
209
+ 0.00818653678894043,
210
+ 0.008104778289794922,
211
+ 0.008190056800842286,
212
+ 0.00816285800933838,
213
+ 0.008273097038269043,
214
+ 0.008194217681884766,
215
+ 0.008155977249145508,
216
+ 0.008141098022460937,
217
+ 0.00812685775756836,
218
+ 0.008107817649841309,
219
+ 0.008124137878417969,
220
+ 0.008050857543945312,
221
+ 0.00808797836303711,
222
+ 0.00808909797668457,
223
+ 0.008208297729492187,
224
+ 0.008280137062072753,
225
+ 0.008336297035217285,
226
+ 0.008192618370056153,
227
+ 0.008178057670593262,
228
+ 0.008161897659301759,
229
+ 0.008220617294311523,
230
+ 0.0082830171585083,
231
+ 0.008333737373352051,
232
+ 0.008255178451538087,
233
+ 0.008308616638183594,
234
+ 0.008275337219238281,
235
+ 0.008228297233581544,
236
+ 0.008138697624206543,
237
+ 0.008159017562866211,
238
+ 0.008116936683654786,
239
+ 0.00814925765991211,
240
+ 0.00813709831237793,
241
+ 0.00812893772125244,
242
+ 0.00808797836303711,
243
+ 0.007981897830963135,
244
+ 0.008040617942810059,
245
+ 0.008077737808227538,
246
+ 0.008059018135070801,
247
+ 0.008116777420043946,
248
+ 0.008123177528381348,
249
+ 0.00810221767425537,
250
+ 0.008108456611633301,
251
+ 0.008090858459472655,
252
+ 0.008035978317260743,
253
+ 0.008040138244628906,
254
+ 0.008276617050170899,
255
+ 0.008299656867980957,
256
+ 0.008261736869812011,
257
+ 0.008241738319396973,
258
+ 0.008076617240905762,
259
+ 0.008054218292236327,
260
+ 0.008027177810668945,
261
+ 0.008079498291015626,
262
+ 0.008173737525939941,
263
+ 0.008098857879638672,
264
+ 0.008104937553405762,
265
+ 0.00813645839691162,
266
+ 0.008119176864624024,
267
+ 0.00814925765991211,
268
+ 0.008135817527770995,
269
+ 0.008152297019958496,
270
+ 0.008167978286743163,
271
+ 0.008287176132202149,
272
+ 0.008205257415771484,
273
+ 0.008170377731323243
274
  ]
275
  },
276
  "throughput": {
277
  "unit": "samples/s",
278
+ "value": 123.08929636740368
279
  },
280
  "energy": null,
281
  "efficiency": null