IlyasMoutawwakil HF staff commited on
Commit
c4a85b3
·
verified ·
1 Parent(s): 4147368

Upload cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_google-bert/bert-base-uncased/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
- "model": "google-bert/bert-base-uncased",
10
  "library": "transformers",
 
 
11
  "device": "cuda",
12
  "device_ids": "0",
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": true,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -85,7 +83,7 @@
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 68702699520,
87
  "optimum_benchmark_version": "0.2.0",
88
- "optimum_benchmark_commit": "3e2eebdc0f80ae6deeb2e1faad3e889ed5a4df2d",
89
  "transformers_version": "4.40.2",
90
  "transformers_commit": null,
91
  "accelerate_version": "0.30.1",
@@ -104,158 +102,157 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 1001.500672,
108
- "max_global_vram": 930.562048,
109
- "max_process_vram": 236889.645056,
110
  "max_reserved": 589.299712,
111
  "max_allocated": 439.700992
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 130,
116
- "total": 1.006110556125641,
117
- "mean": 0.007739311970197236,
118
- "stdev": 0.0002672115651695585,
119
- "p50": 0.007756392955780029,
120
- "p90": 0.007922631835937501,
121
- "p95": 0.008133086061477661,
122
- "p99": 0.008298235654830932,
123
  "values": [
124
- 0.009126458168029785,
125
- 0.008025190353393555,
126
- 0.007867912769317626,
127
- 0.007892713069915772,
128
- 0.00788183307647705,
129
- 0.007871272087097168,
130
- 0.007851592063903809,
131
- 0.007838152885437011,
132
- 0.007829032897949218,
133
- 0.0077227940559387205,
134
- 0.007691274166107178,
135
- 0.007707434177398682,
136
- 0.0077063140869140625,
137
- 0.007734633922576904,
138
- 0.0076779937744140625,
139
- 0.007695273876190185,
140
- 0.007756234169006348,
141
- 0.007755434036254883,
142
- 0.007781993865966797,
143
- 0.007712553977966308,
144
- 0.007716713905334472,
145
- 0.007735273838043213,
146
- 0.007773993968963623,
147
- 0.00773591423034668,
148
- 0.00775687313079834,
149
- 0.007921031951904297,
150
- 0.007937030792236328,
151
- 0.007821352958679199,
152
- 0.007853353023529052,
153
- 0.0078058338165283205,
154
- 0.007782793998718261,
155
- 0.007750312805175782,
156
- 0.007792393207550049,
157
- 0.007867592811584473,
158
- 0.007799912929534912,
159
- 0.007814952850341797,
160
- 0.0078005528450012206,
161
- 0.007805353164672852,
162
- 0.007817512989044189,
163
- 0.007785353183746338,
164
- 0.00781687307357788,
165
- 0.007788712978363037,
166
- 0.007799273014068604,
167
- 0.00784743309020996,
168
- 0.007830632209777833,
169
- 0.007840713024139405,
170
- 0.007800872802734375,
171
- 0.007909832000732421,
172
- 0.007892073154449463,
173
- 0.007880552768707276,
174
- 0.00787607192993164,
175
- 0.00788487195968628,
176
- 0.007853831768035889,
177
- 0.007852073192596436,
178
- 0.007851431846618653,
179
- 0.008108389854431153,
180
- 0.008240867614746093,
181
- 0.008233508110046387,
182
- 0.008148710250854492,
183
- 0.00811398983001709,
184
- 0.008321667671203613,
185
- 0.008171428680419921,
186
- 0.008195428848266602,
187
- 0.007965510845184326,
188
- 0.00788647222518921,
189
- 0.007875752925872804,
190
- 0.007995269775390624,
191
- 0.007858313083648682,
192
- 0.0077435941696166995,
193
- 0.0077863140106201175,
194
- 0.0077816729545593264,
195
- 0.007749194145202636,
196
- 0.007781033039093018,
197
- 0.007733193874359131,
198
- 0.007756392955780029,
199
- 0.007780872821807862,
200
- 0.007723594188690185,
201
- 0.007722474098205566,
202
- 0.007744393825531006,
203
- 0.007694314002990723,
204
- 0.0076976737976074214,
205
- 0.007747273921966552,
206
- 0.007721834182739258,
207
- 0.007725354194641113,
208
- 0.007753832817077637,
209
- 0.0077655138969421385,
210
- 0.007725994110107422,
211
- 0.0077559127807617185,
212
- 0.007700714111328125,
213
- 0.007712073802947998,
214
- 0.007673355102539062,
215
- 0.007753832817077637,
216
- 0.007700715065002441,
217
- 0.0077203941345214844,
218
- 0.007779913902282715,
219
- 0.007722953796386719,
220
- 0.007794473171234131,
221
- 0.007710154056549072,
222
- 0.007540235996246338,
223
- 0.0077136740684509274,
224
- 0.007233839035034179,
225
- 0.0070602412223815915,
226
- 0.007246319770812988,
227
- 0.007633355140686035,
228
- 0.007285038948059082,
229
- 0.007239919185638427,
230
- 0.007280398845672607,
231
- 0.007586315155029297,
232
- 0.007245679855346679,
233
- 0.007552556037902832,
234
- 0.007920231819152832,
235
- 0.007189199924468994,
236
- 0.006964241981506348,
237
- 0.007221839904785156,
238
- 0.007239599227905274,
239
- 0.007286959171295166,
240
- 0.007038321018218994,
241
- 0.007293999195098877,
242
- 0.007399117946624756,
243
- 0.007499756813049316,
244
- 0.0072893590927124025,
245
- 0.007756392955780029,
246
- 0.007707112789154052,
247
- 0.007617034912109375,
248
- 0.007650475025177002,
249
- 0.007306159019470215,
250
- 0.007813512802124023,
251
- 0.007605035781860351,
252
- 0.0077621531486511235,
253
- 0.007732714176177978
254
  ]
255
  },
256
  "throughput": {
257
  "unit": "samples/s",
258
- "value": 129.21045227932774
259
  },
260
  "energy": null,
261
  "efficiency": null
 
6
  "version": "2.2.2+rocm5.7",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
 
9
  "library": "transformers",
10
+ "model": "google-bert/bert-base-uncased",
11
+ "processor": "google-bert/bert-base-uncased",
12
  "device": "cuda",
13
  "device_ids": "0",
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": true,
21
  "device_map": null,
22
  "torch_dtype": null,
 
83
  "gpu_count": 1,
84
  "gpu_vram_mb": 68702699520,
85
  "optimum_benchmark_version": "0.2.0",
86
+ "optimum_benchmark_commit": "6fd377459e287bb09e9383ba2516b1b2a271a562",
87
  "transformers_version": "4.40.2",
88
  "transformers_commit": null,
89
  "accelerate_version": "0.30.1",
 
102
  "forward": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 1001.435136,
106
+ "max_global_vram": 930.557952,
107
+ "max_process_vram": 262239.711232,
108
  "max_reserved": 589.299712,
109
  "max_allocated": 439.700992
110
  },
111
  "latency": {
112
  "unit": "s",
113
+ "count": 129,
114
+ "total": 1.0045842666625977,
115
+ "mean": 0.007787474935368972,
116
+ "stdev": 0.00013716925129050333,
117
+ "p50": 0.007768144130706787,
118
+ "p90": 0.007841935920715332,
119
+ "p95": 0.007930287075042725,
120
+ "p99": 0.00807528564453125,
121
  "values": [
122
+ 0.00914750099182129,
123
+ 0.007959662914276123,
124
+ 0.007900463104248048,
125
+ 0.007898862838745118,
126
+ 0.007961742877960205,
127
+ 0.008088143348693847,
128
+ 0.00804222297668457,
129
+ 0.007935983180999756,
130
+ 0.007773583889007569,
131
+ 0.00781518316268921,
132
+ 0.007758863925933838,
133
+ 0.007767824172973633,
134
+ 0.007729743957519531,
135
+ 0.007732623100280762,
136
+ 0.007747183799743653,
137
+ 0.007768783092498779,
138
+ 0.007693264007568359,
139
+ 0.007682222843170166,
140
+ 0.007723343849182129,
141
+ 0.007710544109344483,
142
+ 0.007717263221740722,
143
+ 0.0077553439140319825,
144
+ 0.007814542770385742,
145
+ 0.007869263172149659,
146
+ 0.007939504146575928,
147
+ 0.00776110315322876,
148
+ 0.007789103984832763,
149
+ 0.0077543830871582035,
150
+ 0.007759024143218994,
151
+ 0.007723342895507812,
152
+ 0.00768174409866333,
153
+ 0.007709582805633545,
154
+ 0.007723663806915283,
155
+ 0.007759984016418457,
156
+ 0.007741743087768554,
157
+ 0.007731823921203613,
158
+ 0.007743503093719482,
159
+ 0.007758224010467529,
160
+ 0.007693583011627197,
161
+ 0.007738704204559326,
162
+ 0.007778062820434571,
163
+ 0.00777998399734497,
164
+ 0.007782382965087891,
165
+ 0.007820143222808837,
166
+ 0.007820623874664307,
167
+ 0.007818862915039063,
168
+ 0.007855184078216552,
169
+ 0.007808463096618652,
170
+ 0.007802543163299561,
171
+ 0.007774223804473877,
172
+ 0.007765902996063233,
173
+ 0.007806863784790039,
174
+ 0.0077903828620910645,
175
+ 0.007824624061584472,
176
+ 0.007829102993011474,
177
+ 0.007842062950134277,
178
+ 0.007784944057464599,
179
+ 0.007780942916870117,
180
+ 0.007771024227142334,
181
+ 0.007727663040161133,
182
+ 0.007763504028320312,
183
+ 0.007799822807312012,
184
+ 0.007797743797302246,
185
+ 0.007829102993011474,
186
+ 0.00775998306274414,
187
+ 0.007768144130706787,
188
+ 0.007809583187103271,
189
+ 0.007768464088439941,
190
+ 0.0077923030853271485,
191
+ 0.007841904163360596,
192
+ 0.007796943187713623,
193
+ 0.007791502952575684,
194
+ 0.00782590389251709,
195
+ 0.007772462844848633,
196
+ 0.007774543762207031,
197
+ 0.0077763028144836426,
198
+ 0.007826863765716553,
199
+ 0.00776110315322876,
200
+ 0.007757423877716064,
201
+ 0.0077281432151794435,
202
+ 0.007791823863983154,
203
+ 0.007794542789459229,
204
+ 0.00780206298828125,
205
+ 0.007807983875274658,
206
+ 0.007780463218688965,
207
+ 0.007828464031219482,
208
+ 0.007736302852630615,
209
+ 0.007759984016418457,
210
+ 0.007807343006134033,
211
+ 0.007786223888397217,
212
+ 0.00774862289428711,
213
+ 0.007718544006347656,
214
+ 0.007715662956237793,
215
+ 0.007722064018249512,
216
+ 0.007744623184204102,
217
+ 0.007796144008636475,
218
+ 0.0077009439468383786,
219
+ 0.007726542949676514,
220
+ 0.007744143962860108,
221
+ 0.00772510290145874,
222
+ 0.007771503925323486,
223
+ 0.007776622772216797,
224
+ 0.007772784233093261,
225
+ 0.00774910306930542,
226
+ 0.007732783794403076,
227
+ 0.007756783962249756,
228
+ 0.007733583927154541,
229
+ 0.007760303020477295,
230
+ 0.007745584011077881,
231
+ 0.00775726318359375,
232
+ 0.007703023910522461,
233
+ 0.007670382976531983,
234
+ 0.007691984176635742,
235
+ 0.0076937441825866695,
236
+ 0.0077734231948852536,
237
+ 0.007721104145050049,
238
+ 0.007734222888946533,
239
+ 0.007697264194488526,
240
+ 0.007679503917694092,
241
+ 0.007725903034210205,
242
+ 0.007744304180145264,
243
+ 0.007732623100280762,
244
+ 0.0077169442176818846,
245
+ 0.007756624221801758,
246
+ 0.0077476630210876465,
247
+ 0.007770223140716553,
248
+ 0.007921742916107177,
249
+ 0.007788942813873291,
250
+ 0.007811823844909668
 
251
  ]
252
  },
253
  "throughput": {
254
  "unit": "samples/s",
255
+ "value": 128.4113282289004
256
  },
257
  "energy": null,
258
  "efficiency": null