Upload cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json with huggingface_hub
Browse files
cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
-
"version": "2.
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "text-generation",
|
9 |
"library": "transformers",
|
@@ -117,33 +117,33 @@
|
|
117 |
"overall": {
|
118 |
"memory": {
|
119 |
"unit": "MB",
|
120 |
-
"max_ram":
|
121 |
-
"max_global_vram":
|
122 |
-
"max_process_vram":
|
123 |
-
"max_reserved":
|
124 |
"max_allocated": 2506.73664
|
125 |
},
|
126 |
"latency": {
|
127 |
"unit": "s",
|
128 |
"count": 5,
|
129 |
-
"total": 0.
|
130 |
-
"mean": 0.
|
131 |
-
"stdev": 0.
|
132 |
-
"p50": 0.
|
133 |
-
"p90": 0.
|
134 |
-
"p95": 0.
|
135 |
-
"p99": 0.
|
136 |
"values": [
|
137 |
-
0.
|
138 |
-
0.
|
139 |
-
0.
|
140 |
-
0.
|
141 |
-
0.
|
142 |
]
|
143 |
},
|
144 |
"throughput": {
|
145 |
"unit": "samples/s",
|
146 |
-
"value":
|
147 |
},
|
148 |
"energy": null,
|
149 |
"efficiency": null
|
@@ -151,30 +151,30 @@
|
|
151 |
"warmup": {
|
152 |
"memory": {
|
153 |
"unit": "MB",
|
154 |
-
"max_ram":
|
155 |
-
"max_global_vram":
|
156 |
-
"max_process_vram":
|
157 |
-
"max_reserved":
|
158 |
"max_allocated": 2506.73664
|
159 |
},
|
160 |
"latency": {
|
161 |
"unit": "s",
|
162 |
"count": 2,
|
163 |
-
"total": 0.
|
164 |
-
"mean": 0.
|
165 |
-
"stdev": 0.
|
166 |
-
"p50": 0.
|
167 |
-
"p90": 0.
|
168 |
-
"p95": 0.
|
169 |
-
"p99": 0.
|
170 |
"values": [
|
171 |
-
0.
|
172 |
-
0.
|
173 |
]
|
174 |
},
|
175 |
"throughput": {
|
176 |
"unit": "samples/s",
|
177 |
-
"value":
|
178 |
},
|
179 |
"energy": null,
|
180 |
"efficiency": null
|
@@ -182,31 +182,31 @@
|
|
182 |
"train": {
|
183 |
"memory": {
|
184 |
"unit": "MB",
|
185 |
-
"max_ram":
|
186 |
-
"max_global_vram":
|
187 |
-
"max_process_vram":
|
188 |
-
"max_reserved":
|
189 |
"max_allocated": 2506.73664
|
190 |
},
|
191 |
"latency": {
|
192 |
"unit": "s",
|
193 |
"count": 3,
|
194 |
-
"total": 0.
|
195 |
-
"mean": 0.
|
196 |
-
"stdev":
|
197 |
-
"p50": 0.
|
198 |
-
"p90": 0.
|
199 |
-
"p95": 0.
|
200 |
-
"p99": 0.
|
201 |
"values": [
|
202 |
-
0.
|
203 |
-
0.
|
204 |
-
0.
|
205 |
]
|
206 |
},
|
207 |
"throughput": {
|
208 |
"unit": "samples/s",
|
209 |
-
"value":
|
210 |
},
|
211 |
"energy": null,
|
212 |
"efficiency": null
|
|
|
3 |
"name": "cuda_training_transformers_text-generation_openai-community/gpt2",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
+
"version": "2.4.0+rocm6.1",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"task": "text-generation",
|
9 |
"library": "transformers",
|
|
|
117 |
"overall": {
|
118 |
"memory": {
|
119 |
"unit": "MB",
|
120 |
+
"max_ram": 1640.144896,
|
121 |
+
"max_global_vram": 11.239424,
|
122 |
+
"max_process_vram": 0.0,
|
123 |
+
"max_reserved": 2889.875456,
|
124 |
"max_allocated": 2506.73664
|
125 |
},
|
126 |
"latency": {
|
127 |
"unit": "s",
|
128 |
"count": 5,
|
129 |
+
"total": 0.9606382255554198,
|
130 |
+
"mean": 0.19212764511108396,
|
131 |
+
"stdev": 0.2875477069341439,
|
132 |
+
"p50": 0.04804556655883789,
|
133 |
+
"p90": 0.48009199523925783,
|
134 |
+
"p95": 0.6236570205688475,
|
135 |
+
"p99": 0.7385090408325194,
|
136 |
"values": [
|
137 |
+
0.7672220458984375,
|
138 |
+
0.04939691925048828,
|
139 |
+
0.04803228759765625,
|
140 |
+
0.04794140625,
|
141 |
+
0.04804556655883789
|
142 |
]
|
143 |
},
|
144 |
"throughput": {
|
145 |
"unit": "samples/s",
|
146 |
+
"value": 52.04873038556331
|
147 |
},
|
148 |
"energy": null,
|
149 |
"efficiency": null
|
|
|
151 |
"warmup": {
|
152 |
"memory": {
|
153 |
"unit": "MB",
|
154 |
+
"max_ram": 1640.144896,
|
155 |
+
"max_global_vram": 11.239424,
|
156 |
+
"max_process_vram": 0.0,
|
157 |
+
"max_reserved": 2889.875456,
|
158 |
"max_allocated": 2506.73664
|
159 |
},
|
160 |
"latency": {
|
161 |
"unit": "s",
|
162 |
"count": 2,
|
163 |
+
"total": 0.8166189651489257,
|
164 |
+
"mean": 0.40830948257446287,
|
165 |
+
"stdev": 0.3589125633239746,
|
166 |
+
"p50": 0.40830948257446287,
|
167 |
+
"p90": 0.6954395332336425,
|
168 |
+
"p95": 0.7313307895660399,
|
169 |
+
"p99": 0.7600437946319579,
|
170 |
"values": [
|
171 |
+
0.7672220458984375,
|
172 |
+
0.04939691925048828
|
173 |
]
|
174 |
},
|
175 |
"throughput": {
|
176 |
"unit": "samples/s",
|
177 |
+
"value": 9.796490580574565
|
178 |
},
|
179 |
"energy": null,
|
180 |
"efficiency": null
|
|
|
182 |
"train": {
|
183 |
"memory": {
|
184 |
"unit": "MB",
|
185 |
+
"max_ram": 1640.144896,
|
186 |
+
"max_global_vram": 11.239424,
|
187 |
+
"max_process_vram": 0.0,
|
188 |
+
"max_reserved": 2889.875456,
|
189 |
"max_allocated": 2506.73664
|
190 |
},
|
191 |
"latency": {
|
192 |
"unit": "s",
|
193 |
"count": 3,
|
194 |
+
"total": 0.14401926040649415,
|
195 |
+
"mean": 0.04800642013549805,
|
196 |
+
"stdev": 4.6290291877677313e-05,
|
197 |
+
"p50": 0.04803228759765625,
|
198 |
+
"p90": 0.04804291076660156,
|
199 |
+
"p95": 0.04804423866271973,
|
200 |
+
"p99": 0.04804530097961426,
|
201 |
"values": [
|
202 |
+
0.04803228759765625,
|
203 |
+
0.04794140625,
|
204 |
+
0.04804556655883789
|
205 |
]
|
206 |
},
|
207 |
"throughput": {
|
208 |
"unit": "samples/s",
|
209 |
+
"value": 124.98328313306864
|
210 |
},
|
211 |
"energy": null,
|
212 |
"efficiency": null
|