IlyasMoutawwakil HF staff commited on
Commit
6a310dc
·
verified ·
1 Parent(s): dad3194

Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json CHANGED
@@ -85,7 +85,7 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.3.0",
88
- "optimum_benchmark_commit": "19eeac52e408e408898ac1fac3d9abefced0131b",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
@@ -104,164 +104,164 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1004.220416,
108
- "max_global_vram": 898.4576,
109
- "max_process_vram": 174836.006912,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.507712
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 136,
116
- "total": 1.0000946702957147,
117
- "mean": 0.007353637281586143,
118
- "stdev": 0.00037703269468259565,
119
- "p50": 0.007270452976226806,
120
- "p90": 0.007637732028961182,
121
- "p95": 0.00787725055217743,
122
- "p99": 0.008236936759948731,
123
  "values": [
124
- 0.007567652225494385,
125
- 0.007185254096984863,
126
- 0.0073254132270812984,
127
- 0.007444451808929444,
128
- 0.007500132083892822,
129
- 0.007446211814880371,
130
- 0.007449892044067383,
131
- 0.0073938918113708495,
132
- 0.007305412769317627,
133
- 0.0072207732200622555,
134
- 0.0072231731414794925,
135
- 0.0071846132278442386,
136
- 0.007152933120727539,
137
- 0.007121413230895996,
138
- 0.0071151738166809085,
139
- 0.007078053951263427,
140
- 0.007119332790374756,
141
- 0.007115493774414063,
142
- 0.007096774101257324,
143
- 0.007073412895202637,
144
- 0.007066534042358398,
145
- 0.007104292869567871,
146
- 0.007117894172668457,
147
- 0.0071764540672302245,
148
- 0.007148774147033692,
149
- 0.007087653160095215,
150
- 0.007121892929077148,
151
- 0.007091494083404541,
152
- 0.007087653160095215,
153
- 0.0070937337875366215,
154
- 0.007043334007263183,
155
- 0.0070393338203430175,
156
- 0.007041253089904785,
157
- 0.007043173789978027,
158
- 0.007031333923339844,
159
- 0.00716749382019043,
160
- 0.007278371810913086,
161
- 0.007299653053283691,
162
- 0.007333571910858154,
163
- 0.007317412853240967,
164
- 0.00731277322769165,
165
- 0.007238213062286377,
166
- 0.007367812156677246,
167
- 0.010725239753723144,
168
- 0.007490531921386719,
169
- 0.007280453205108643,
170
- 0.007252933025360108,
171
- 0.0072542128562927246,
172
- 0.007270853042602539,
173
- 0.007249252796173096,
174
- 0.007226212978363037,
175
- 0.007245253086090088,
176
- 0.0072780528068542485,
177
- 0.007209893226623535,
178
- 0.0072879729270935055,
179
- 0.007247172832489014,
180
- 0.007235493183135986,
181
- 0.007203332901000977,
182
- 0.007432611942291259,
183
- 0.007246212959289551,
184
- 0.007256772994995117,
185
- 0.007242533206939697,
186
- 0.007253252983093262,
187
- 0.007234212875366211,
188
- 0.007225412845611573,
189
- 0.007340612888336181,
190
- 0.0072894120216369625,
191
- 0.0072658929824829105,
192
- 0.007288773059844971,
193
- 0.007261413097381592,
194
- 0.0072702131271362306,
195
- 0.00727789306640625,
196
- 0.0072591729164123535,
197
- 0.007251492977142334,
198
- 0.007214693069458008,
199
- 0.007269573211669922,
200
- 0.007224133014678955,
201
- 0.007231973171234131,
202
- 0.0072854127883911135,
203
- 0.007297732830047607,
204
- 0.0072682929039001466,
205
- 0.007290692806243897,
206
- 0.007270692825317383,
207
- 0.0072682929039001466,
208
- 0.0072634930610656736,
209
- 0.007221733093261719,
210
- 0.007246692180633545,
211
- 0.007226852893829345,
212
- 0.007274373054504395,
213
- 0.007249252796173096,
214
- 0.007252613067626953,
215
- 0.007242533206939697,
216
- 0.007231492996215821,
217
- 0.007313252925872803,
218
- 0.008176609039306641,
219
- 0.0082060489654541,
220
- 0.008176449775695802,
221
- 0.007780770778656006,
222
- 0.0081666898727417,
223
- 0.008202049255371094,
224
- 0.008253568649291993,
225
- 0.00775645112991333,
226
- 0.007769411087036133,
227
- 0.007702051162719726,
228
- 0.007640131950378418,
229
- 0.007752450942993164,
230
- 0.007770051002502441,
231
- 0.0075524520874023435,
232
- 0.007191332817077636,
233
- 0.0073820528984069825,
234
- 0.007450532913208008,
235
- 0.0073793320655822755,
236
- 0.00743693208694458,
237
- 0.007351973056793213,
238
- 0.0074084529876708985,
239
- 0.007320452213287354,
240
- 0.007302213191986084,
241
- 0.007279653072357178,
242
- 0.007258052825927734,
243
- 0.007320612907409668,
244
- 0.0073009319305419925,
245
- 0.007280292987823486,
246
- 0.0072658929824829105,
247
- 0.0073199720382690426,
248
- 0.007281892776489258,
249
- 0.007283812999725342,
250
- 0.007496451854705811,
251
- 0.007635332107543945,
252
- 0.007532611846923828,
253
- 0.007252932071685791,
254
- 0.0072446131706237794,
255
- 0.007279973030090332,
256
- 0.007303012847900391,
257
- 0.007296933174133301,
258
- 0.0073102121353149415,
259
- 0.007296932220458984
260
  ]
261
  },
262
  "throughput": {
263
  "unit": "samples/s",
264
- "value": 135.98712605856255
265
  },
266
  "energy": null,
267
  "efficiency": null
 
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.3.0",
88
+ "optimum_benchmark_commit": "57f6495c03ea0fa48e157048c97add150dcd765c",
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.31.0",
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 1005.883392,
108
+ "max_global_vram": 898.473984,
109
+ "max_process_vram": 175601.803264,
110
  "max_reserved": 555.74528,
111
  "max_allocated": 499.507712
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 136,
116
+ "total": 0.9973448486328123,
117
+ "mean": 0.007333418004653033,
118
+ "stdev": 0.0002452375024910205,
119
+ "p50": 0.0072775003910064695,
120
+ "p90": 0.007556140184402466,
121
+ "p95": 0.007932978630065918,
122
+ "p99": 0.008247482204437256,
123
  "values": [
124
+ 0.007991658210754395,
125
+ 0.007382541179656982,
126
+ 0.007444620132446289,
127
+ 0.007689579010009766,
128
+ 0.007338059902191162,
129
+ 0.007279500961303711,
130
+ 0.00726669979095459,
131
+ 0.007239020824432373,
132
+ 0.007247660160064697,
133
+ 0.007199660778045654,
134
+ 0.007185900211334228,
135
+ 0.007147181034088135,
136
+ 0.0071585412025451664,
137
+ 0.007118220806121826,
138
+ 0.0071681408882141116,
139
+ 0.0071425409317016604,
140
+ 0.007184940814971924,
141
+ 0.007171501159667968,
142
+ 0.0071511797904968265,
143
+ 0.007159501075744629,
144
+ 0.007126380920410156,
145
+ 0.007233420848846436,
146
+ 0.0071494197845458985,
147
+ 0.0071358208656311034,
148
+ 0.00718350076675415,
149
+ 0.007194221019744873,
150
+ 0.007153100967407226,
151
+ 0.007169421195983887,
152
+ 0.007136940956115723,
153
+ 0.0071663808822631835,
154
+ 0.007172461032867432,
155
+ 0.0071308608055114745,
156
+ 0.0071425409317016604,
157
+ 0.007142701148986816,
158
+ 0.007135980129241944,
159
+ 0.007131980895996094,
160
+ 0.007259500980377197,
161
+ 0.007183020114898682,
162
+ 0.007249900817871094,
163
+ 0.007343820095062256,
164
+ 0.007303979873657227,
165
+ 0.007249101161956787,
166
+ 0.007249101161956787,
167
+ 0.008466217041015625,
168
+ 0.0073807802200317385,
169
+ 0.007267021179199219,
170
+ 0.0072534198760986325,
171
+ 0.0072252612113952635,
172
+ 0.007251339912414551,
173
+ 0.00721566104888916,
174
+ 0.007277740001678467,
175
+ 0.007232301235198975,
176
+ 0.007254860877990723,
177
+ 0.007266221046447754,
178
+ 0.007273260116577148,
179
+ 0.0072441411018371584,
180
+ 0.007243020057678223,
181
+ 0.007234220981597901,
182
+ 0.007507500171661377,
183
+ 0.007261899948120118,
184
+ 0.007274381160736084,
185
+ 0.007227340221405029,
186
+ 0.007277901172637939,
187
+ 0.007247340202331543,
188
+ 0.007251020908355713,
189
+ 0.007346380233764649,
190
+ 0.007283979892730713,
191
+ 0.007475019931793213,
192
+ 0.007271181106567383,
193
+ 0.0072678208351135256,
194
+ 0.007290219783782959,
195
+ 0.007248140811920166,
196
+ 0.0072649397850036625,
197
+ 0.007254061222076416,
198
+ 0.0072670202255249025,
199
+ 0.007279500961303711,
200
+ 0.0072319798469543456,
201
+ 0.007229741096496582,
202
+ 0.007288619995117187,
203
+ 0.007283181190490723,
204
+ 0.0072913398742675785,
205
+ 0.007268300056457519,
206
+ 0.007295501232147217,
207
+ 0.007283339977264405,
208
+ 0.007267180919647217,
209
+ 0.007245259761810303,
210
+ 0.007284141063690185,
211
+ 0.007304780960083008,
212
+ 0.007285099983215332,
213
+ 0.007305740833282471,
214
+ 0.007313260078430176,
215
+ 0.007315499782562256,
216
+ 0.007348461151123047,
217
+ 0.008228137969970703,
218
+ 0.008257898330688476,
219
+ 0.007913418769836426,
220
+ 0.008070697784423828,
221
+ 0.008225737571716308,
222
+ 0.008199817657470703,
223
+ 0.007908298969268798,
224
+ 0.0077406191825866695,
225
+ 0.007847339153289794,
226
+ 0.007714539051055908,
227
+ 0.007604780197143554,
228
+ 0.0072251009941101075,
229
+ 0.007286059856414795,
230
+ 0.007297101020812988,
231
+ 0.007288461208343506,
232
+ 0.007288459777832031,
233
+ 0.00728334093093872,
234
+ 0.00732013988494873,
235
+ 0.007303659915924072,
236
+ 0.007271501064300537,
237
+ 0.007290540218353271,
238
+ 0.0073009409904479985,
239
+ 0.007292780876159668,
240
+ 0.007309579849243164,
241
+ 0.0072844610214233396,
242
+ 0.007316619873046875,
243
+ 0.007289420127868653,
244
+ 0.007294061183929443,
245
+ 0.007288780212402344,
246
+ 0.007314539909362793,
247
+ 0.007316141128540039,
248
+ 0.007300300121307373,
249
+ 0.007292140007019043,
250
+ 0.007285261154174804,
251
+ 0.007308300018310547,
252
+ 0.007240780830383301,
253
+ 0.007253260135650635,
254
+ 0.007298861026763916,
255
+ 0.007277260780334473,
256
+ 0.00732174015045166,
257
+ 0.00731118106842041,
258
+ 0.007331820011138916,
259
+ 0.007318220138549805
260
  ]
261
  },
262
  "throughput": {
263
  "unit": "samples/s",
264
+ "value": 136.36206191512647
265
  },
266
  "energy": null,
267
  "efficiency": null