IlyasMoutawwakil HF staff commited on
Commit
aece8d2
·
verified ·
1 Parent(s): 9d17d9b

Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -85,7 +85,7 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
- "optimum_benchmark_commit": "c594845efb520077430f4fe3f536bc1756e2b290",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
@@ -104,165 +104,162 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1009.557504,
108
  "max_global_vram": 897.142784,
109
- "max_process_vram": 177571.79904,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.443712
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 137,
116
- "total": 0.9980697560310363,
117
- "mean": 0.0072851807009564696,
118
- "stdev": 0.0003758726213710794,
119
- "p50": 0.007223018169403076,
120
- "p90": 0.007646856594085694,
121
- "p95": 0.007821160697937011,
122
- "p99": 0.007973332366943357,
123
  "values": [
124
- 0.007863656044006347,
125
- 0.008014055252075195,
126
- 0.0077446169853210445,
127
- 0.00781965684890747,
128
- 0.007775496959686279,
129
- 0.007558537006378174,
130
- 0.00744061803817749,
131
- 0.007349256992340088,
132
- 0.007344297885894776,
133
- 0.007188138008117676,
134
- 0.007176619052886963,
135
- 0.007412137985229492,
136
- 0.007293097972869873,
137
- 0.007300457954406738,
138
- 0.007225257873535156,
139
- 0.007351656913757324,
140
- 0.007140139102935791,
141
- 0.007189418792724609,
142
- 0.00708621883392334,
143
- 0.00716989803314209,
144
- 0.007187178134918213,
145
- 0.007236937999725342,
146
- 0.007252458095550537,
147
- 0.007345417976379395,
148
- 0.0072502179145812985,
149
- 0.007332618236541748,
150
- 0.0071926178932189945,
151
- 0.007103658199310302,
152
- 0.007100778102874756,
153
- 0.010521727561950684,
154
- 0.007169258117675781,
155
- 0.006943018913269043,
156
- 0.0069753389358520506,
157
- 0.007223018169403076,
158
- 0.006904139041900635,
159
- 0.006982539176940918,
160
- 0.00699085807800293,
161
- 0.006960618972778321,
162
- 0.006983658790588379,
163
- 0.007009098052978516,
164
- 0.006947338104248047,
165
- 0.007065739154815674,
166
- 0.007399336814880371,
167
- 0.00735277795791626,
168
- 0.007015657901763916,
169
- 0.007187977790832519,
170
- 0.007212138175964356,
171
- 0.007035178184509277,
172
- 0.00702509880065918,
173
- 0.007035337924957276,
174
- 0.007042698860168457,
175
- 0.007019177913665771,
176
- 0.006978538990020752,
177
- 0.006999978065490723,
178
- 0.00702509880065918,
179
- 0.007054537773132324,
180
- 0.007236618041992188,
181
- 0.007082057952880859,
182
- 0.006998379230499268,
183
- 0.007047177791595459,
184
- 0.007014379024505615,
185
- 0.007028937816619873,
186
- 0.007034059047698975,
187
- 0.007002217769622803,
188
- 0.007066218852996826,
189
- 0.0070977392196655276,
190
- 0.007019978046417236,
191
- 0.007041258811950684,
192
- 0.007029098033905029,
193
- 0.007551177024841309,
194
- 0.007401896953582764,
195
- 0.0070228590965271,
196
- 0.006996938228607178,
197
- 0.007366538047790527,
198
- 0.0078054161071777345,
199
- 0.007806215763092041,
200
- 0.007306858062744141,
201
- 0.006968617916107178,
202
- 0.0069822192192077635,
203
- 0.007070218086242676,
204
- 0.0069943790435791015,
205
- 0.007037578105926513,
206
- 0.007084138870239258,
207
- 0.007062698841094971,
208
- 0.007077578067779541,
209
- 0.007052938938140869,
210
- 0.007057738780975342,
211
- 0.007008617877960205,
212
- 0.007030858993530274,
213
- 0.007072457790374756,
214
- 0.0070506992340087895,
215
- 0.007330537796020508,
216
- 0.007841576099395751,
217
- 0.007827176094055175,
218
- 0.007900936126708984,
219
- 0.007840136051177978,
220
- 0.007401576995849609,
221
- 0.007537736892700196,
222
- 0.007515176773071289,
223
- 0.007599496841430664,
224
- 0.00755277681350708,
225
- 0.0074626970291137695,
226
- 0.007279977798461914,
227
- 0.007461577892303467,
228
- 0.007523336887359619,
229
- 0.007393417835235595,
230
- 0.0074577369689941405,
231
- 0.00761133623123169,
232
- 0.007376617908477783,
233
- 0.0073862171173095705,
234
- 0.007361577987670898,
235
- 0.007250536918640136,
236
- 0.007291018009185791,
237
- 0.0072924580574035645,
238
- 0.007551177024841309,
239
- 0.0074674968719482425,
240
- 0.007340617179870605,
241
- 0.007516936779022217,
242
- 0.007334857940673828,
243
- 0.007283176898956299,
244
- 0.0073890981674194335,
245
- 0.007344777107238769,
246
- 0.00707469892501831,
247
- 0.007023338794708252,
248
- 0.007042858123779297,
249
- 0.007193418025970459,
250
- 0.007483336925506592,
251
- 0.0072471780776977536,
252
- 0.007012299060821533,
253
- 0.007141417980194091,
254
- 0.007555656909942627,
255
- 0.007715816020965576,
256
- 0.007339657783508301,
257
- 0.007700137138366699,
258
- 0.007371177196502685,
259
- 0.007140777111053467,
260
- 0.0072924580574035645
261
  ]
262
  },
263
  "throughput": {
264
  "unit": "samples/s",
265
- "value": 137.2649548512517
266
  },
267
  "energy": null,
268
  "efficiency": null
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.1",
88
+ "optimum_benchmark_commit": "156844ab796ad7cf3da92a0bf30b174d1bcc0aa5",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1009.942528,
108
  "max_global_vram": 897.142784,
109
+ "max_process_vram": 199578.570752,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.443712
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 134,
116
+ "total": 0.9950863709449767,
117
+ "mean": 0.007426017693619229,
118
+ "stdev": 0.0004018532667655497,
119
+ "p50": 0.007285921096801758,
120
+ "p90": 0.007909937000274658,
121
+ "p95": 0.007931480956077576,
122
+ "p99": 0.008188913011550902,
123
  "values": [
124
+ 0.008827040672302246,
125
+ 0.007209280967712403,
126
+ 0.00718320083618164,
127
+ 0.00731040096282959,
128
+ 0.007342560768127442,
129
+ 0.007372001171112061,
130
+ 0.0072216010093688965,
131
+ 0.007950241088867188,
132
+ 0.007899200916290284,
133
+ 0.007848960876464844,
134
+ 0.00783712100982666,
135
+ 0.00736336088180542,
136
+ 0.0072614412307739256,
137
+ 0.0073142409324646,
138
+ 0.007400801181793213,
139
+ 0.007210881233215332,
140
+ 0.008231680870056152,
141
+ 0.008102081298828124,
142
+ 0.007395201206207276,
143
+ 0.0070401611328125,
144
+ 0.007402400970458984,
145
+ 0.007244161128997803,
146
+ 0.007036641120910645,
147
+ 0.007023681163787842,
148
+ 0.00703088092803955,
149
+ 0.007251040935516357,
150
+ 0.007001121044158936,
151
+ 0.006983360767364502,
152
+ 0.007025440216064453,
153
+ 0.0069702410697937015,
154
+ 0.007024001121520996,
155
+ 0.007001121044158936,
156
+ 0.0069971208572387695,
157
+ 0.007049760818481445,
158
+ 0.007002080917358398,
159
+ 0.0070222411155700685,
160
+ 0.007003040790557862,
161
+ 0.007228480815887451,
162
+ 0.0070276808738708495,
163
+ 0.007024480819702149,
164
+ 0.006968481063842773,
165
+ 0.007060320854187012,
166
+ 0.0070131211280822755,
167
+ 0.007092161178588867,
168
+ 0.007070081233978271,
169
+ 0.007094241142272949,
170
+ 0.007036160945892334,
171
+ 0.007011360168457032,
172
+ 0.007075840950012207,
173
+ 0.0070796809196472165,
174
+ 0.007100161075592041,
175
+ 0.007052320957183838,
176
+ 0.0070465607643127445,
177
+ 0.007087841033935547,
178
+ 0.0070491209030151365,
179
+ 0.007047360897064209,
180
+ 0.007065280914306641,
181
+ 0.007050721168518066,
182
+ 0.0070790410041809086,
183
+ 0.007064160823822022,
184
+ 0.007076001167297364,
185
+ 0.007103840827941895,
186
+ 0.007078880786895752,
187
+ 0.007082561016082764,
188
+ 0.007366880893707275,
189
+ 0.007902400970458984,
190
+ 0.007968320846557617,
191
+ 0.007204960823059082,
192
+ 0.00704624080657959,
193
+ 0.007215840816497803,
194
+ 0.007794560909271241,
195
+ 0.00791360092163086,
196
+ 0.007879361152648926,
197
+ 0.007853281021118165,
198
+ 0.0078905611038208,
199
+ 0.00785408115386963,
200
+ 0.007869441032409668,
201
+ 0.007835200786590576,
202
+ 0.007959361076354981,
203
+ 0.00789664077758789,
204
+ 0.007922561168670654,
205
+ 0.007916000843048096,
206
+ 0.007906081199645996,
207
+ 0.007907361030578612,
208
+ 0.007876801967620849,
209
+ 0.00788624095916748,
210
+ 0.007930080890655517,
211
+ 0.007908960819244385,
212
+ 0.007934081077575683,
213
+ 0.007923521041870118,
214
+ 0.007909601211547851,
215
+ 0.007869760990142822,
216
+ 0.007849600791931153,
217
+ 0.007847681045532226,
218
+ 0.007866240978240967,
219
+ 0.007842401027679443,
220
+ 0.007873441219329834,
221
+ 0.007892321109771729,
222
+ 0.00791024112701416,
223
+ 0.007877601146697997,
224
+ 0.007898722171783447,
225
+ 0.00786720085144043,
226
+ 0.007861920833587646,
227
+ 0.007865601062774659,
228
+ 0.007832321166992187,
229
+ 0.007862081050872802,
230
+ 0.007855521202087403,
231
+ 0.007910080909729004,
232
+ 0.007480640888214111,
233
+ 0.007563681125640869,
234
+ 0.00786432123184204,
235
+ 0.007860641002655029,
236
+ 0.007444480895996094,
237
+ 0.00744736099243164,
238
+ 0.007353120803833007,
239
+ 0.0073401608467102055,
240
+ 0.007476480960845947,
241
+ 0.007476641178131104,
242
+ 0.007551840782165527,
243
+ 0.007010400772094727,
244
+ 0.007031681060791016,
245
+ 0.007021121025085449,
246
+ 0.007070880889892578,
247
+ 0.007053280830383301,
248
+ 0.007067360877990723,
249
+ 0.007023681163787842,
250
+ 0.007051681041717529,
251
+ 0.007029921054840088,
252
+ 0.007007360935211182,
253
+ 0.007036799907684326,
254
+ 0.0070417609214782715,
255
+ 0.007070879936218261,
256
+ 0.0070809612274169926,
257
+ 0.007218720912933349
 
 
 
258
  ]
259
  },
260
  "throughput": {
261
  "unit": "samples/s",
262
+ "value": 134.66167753131606
263
  },
264
  "energy": null,
265
  "efficiency": null