IlyasMoutawwakil HF staff commited on
Commit
870cc25
·
verified ·
1 Parent(s): cd956f3

Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json CHANGED
@@ -84,7 +84,7 @@
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
- "optimum_benchmark_version": "0.2.1",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
@@ -104,7 +104,7 @@
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 976.09728,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
@@ -112,101 +112,104 @@
112
  },
113
  "latency": {
114
  "unit": "s",
115
- "count": 70,
116
- "total": 1.0049732198715209,
117
- "mean": 0.014356760283878873,
118
- "stdev": 0.0005069449326987283,
119
- "p50": 0.014176256179809572,
120
- "p90": 0.014902886390686036,
121
- "p95": 0.015359041786193846,
122
- "p99": 0.016185395374298098,
123
  "values": [
124
- 0.015296544075012207,
125
- 0.015410176277160645,
126
- 0.01510707187652588,
127
- 0.01490124797821045,
128
- 0.014096384048461913,
129
- 0.014072832107543945,
130
- 0.014013440132141113,
131
- 0.014226431846618653,
132
- 0.01417523193359375,
133
- 0.014279680252075195,
134
- 0.014166015625,
135
- 0.014194687843322755,
136
- 0.014110719680786133,
137
- 0.01417523193359375,
138
- 0.014137344360351562,
139
- 0.014078975677490235,
140
- 0.014125056266784668,
141
- 0.014464032173156739,
142
- 0.014375935554504395,
143
- 0.014260224342346191,
144
- 0.014311424255371094,
145
- 0.014268416404724121,
146
- 0.014050304412841797,
147
- 0.014303232192993164,
148
- 0.014225407600402832,
149
- 0.01407487964630127,
150
- 0.014136320114135742,
151
- 0.014467071533203125,
152
- 0.014335040092468262,
153
- 0.014097439765930175,
154
- 0.01406668758392334,
155
- 0.01491763210296631,
156
- 0.01459712028503418,
157
- 0.014526464462280274,
158
- 0.014592000007629394,
159
- 0.015723520278930665,
160
- 0.01721343994140625,
161
- 0.01546444797515869,
162
- 0.014766079902648926,
163
- 0.014264320373535156,
164
- 0.014220288276672363,
165
- 0.014145536422729492,
166
- 0.014108672142028808,
167
- 0.014125056266784668,
168
- 0.014125056266784668,
169
- 0.01426534366607666,
170
- 0.014252032279968262,
171
- 0.014897151947021485,
172
- 0.01417728042602539,
173
- 0.014160896301269531,
174
- 0.014171135902404786,
175
- 0.014114815711975098,
176
- 0.014152704238891601,
177
- 0.01451417636871338,
178
- 0.014017536163330077,
179
- 0.013896703720092773,
180
- 0.013942784309387207,
181
- 0.014303168296813966,
182
- 0.014135295867919923,
183
  0.014082048416137695,
184
- 0.014210047721862793,
185
- 0.014194751739501953,
186
- 0.014118911743164063,
187
- 0.0140830717086792,
188
- 0.014053376197814941,
189
- 0.014156800270080566,
190
- 0.014173184394836426,
 
 
 
 
 
 
 
 
 
 
 
 
191
  0.014101504325866699,
192
- 0.014220288276672363,
193
- 0.013785087585449218
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
194
  ]
195
  },
196
  "throughput": {
197
  "unit": "samples/s",
198
- "value": 69.65359734555814
199
  },
200
  "energy": {
201
  "unit": "kWh",
202
- "cpu": 1.7041056165619502e-07,
203
- "ram": 9.309723819395305e-08,
204
- "gpu": 3.547542520571448e-07,
205
- "total": 6.182620519072929e-07
206
  },
207
  "efficiency": {
208
  "unit": "samples/kWh",
209
- "value": 1617437.1319007427
210
  }
211
  }
212
  }
 
84
  ],
85
  "gpu_count": 1,
86
  "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.3.0",
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.42.3",
90
  "transformers_commit": null,
 
104
  "forward": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 976.187392,
108
  "max_global_vram": 1434.976256,
109
  "max_process_vram": 0.0,
110
  "max_reserved": 794.820608,
 
112
  },
113
  "latency": {
114
  "unit": "s",
115
+ "count": 73,
116
+ "total": 1.0094481916427611,
117
+ "mean": 0.013828057419763852,
118
+ "stdev": 0.000524276007705457,
119
+ "p50": 0.013813759803771973,
120
+ "p90": 0.014313472175598144,
121
+ "p95": 0.01499320316314697,
122
+ "p99": 0.01558552604675293,
123
  "values": [
124
+ 0.015321087837219239,
125
+ 0.015568896293640137,
126
+ 0.01562828826904297,
127
+ 0.015215616226196289,
128
+ 0.014110783576965331,
129
+ 0.013884415626525879,
130
+ 0.01420083236694336,
131
+ 0.014844927787780762,
132
+ 0.014370911598205567,
133
+ 0.014386176109313965,
134
+ 0.014109696388244629,
135
+ 0.013880255699157715,
136
+ 0.013993984222412109,
137
+ 0.014191616058349609,
138
+ 0.014065664291381836,
139
+ 0.013813759803771973,
140
+ 0.013923328399658203,
141
+ 0.014142463684082032,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
  0.014082048416137695,
143
+ 0.013999103546142578,
144
+ 0.014004223823547364,
145
+ 0.01387724781036377,
146
+ 0.013910016059875489,
147
+ 0.01399500846862793,
148
+ 0.013827072143554688,
149
+ 0.013780991554260253,
150
+ 0.014105600357055664,
151
+ 0.014211071968078613,
152
+ 0.014019583702087402,
153
+ 0.013859840393066406,
154
+ 0.013947903633117676,
155
+ 0.013727744102478028,
156
+ 0.013845503807067871,
157
+ 0.013601792335510255,
158
+ 0.013552639961242676,
159
+ 0.01394380760192871,
160
+ 0.014152704238891601,
161
+ 0.014339072227478027,
162
  0.014101504325866699,
163
+ 0.013966336250305175,
164
+ 0.014003199577331543,
165
+ 0.013811712265014648,
166
+ 0.013805567741394043,
167
+ 0.013789183616638183,
168
+ 0.01339907169342041,
169
+ 0.01335807991027832,
170
+ 0.013360128402709961,
171
+ 0.013364224433898926,
172
+ 0.013415424346923829,
173
+ 0.013340671539306641,
174
+ 0.01337548828125,
175
+ 0.013418496131896973,
176
+ 0.013376511573791505,
177
+ 0.013401087760925292,
178
+ 0.013342783927917481,
179
+ 0.013384703636169434,
180
+ 0.013378560066223144,
181
+ 0.013367360115051269,
182
+ 0.013371456146240234,
183
+ 0.013373439788818359,
184
+ 0.013353919982910157,
185
+ 0.013350879669189453,
186
+ 0.01338368034362793,
187
+ 0.013339648246765137,
188
+ 0.013351936340332032,
189
+ 0.013336607933044434,
190
+ 0.013355008125305176,
191
+ 0.013327360153198242,
192
+ 0.013293567657470704,
193
+ 0.013340671539306641,
194
+ 0.013365247726440429,
195
+ 0.013389823913574218,
196
+ 0.013319168090820312
197
  ]
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
+ "value": 72.31673760413685
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
+ "cpu": 1.600326630804274e-07,
206
+ "ram": 8.738030096339772e-08,
207
+ "gpu": 3.615151040266736e-07,
208
+ "total": 6.089280680704987e-07
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
+ "value": 1642230.0965181077
213
  }
214
  }
215
  }