IlyasMoutawwakil HF staff commited on
Commit
f7ac072
·
verified ·
1 Parent(s): 02e7073

Upload cuda_inference_diffusers_text-to-image_CompVis/stable-diffusion-v1-4/benchmark.json with huggingface_hub

Browse files
cuda_inference_diffusers_text-to-image_CompVis/stable-diffusion-v1-4/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_inference_diffusers_text-to-image_CompVis/stable-diffusion-v1-4",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.1+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-to-image",
9
  "library": "diffusers",
@@ -76,7 +76,7 @@
76
  "cpu_ram_mb": 66697.261056,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
- "platform": "Linux-5.10.223-212.873.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
@@ -88,9 +88,9 @@
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
- "accelerate_version": "0.34.0",
92
  "accelerate_commit": null,
93
- "diffusers_version": "0.30.2",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
@@ -104,74 +104,74 @@
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
- "max_ram": 4265.947136,
108
- "max_global_vram": 6212.288512,
109
  "max_process_vram": 0.0,
110
- "max_reserved": 5626.658816,
111
- "max_allocated": 5502.410752
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
- "total": 61.91855078125,
117
- "mean": 61.91855078125,
118
  "stdev": 0.0,
119
- "p50": 61.91855078125,
120
- "p90": 61.91855078125,
121
- "p95": 61.91855078125,
122
- "p99": 61.91855078125,
123
  "values": [
124
- 61.91855078125
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
- "cpu": 0.0006394203940444445,
131
- "ram": 0.0003504430010093095,
132
- "gpu": 0.0008471148443580006,
133
- "total": 0.0018369782394117547
134
  },
135
  "efficiency": null
136
  },
137
  "call": {
138
  "memory": {
139
  "unit": "MB",
140
- "max_ram": 1441.210368,
141
- "max_global_vram": 8577.875968,
142
  "max_process_vram": 0.0,
143
- "max_reserved": 7933.526016,
144
- "max_allocated": 6588.069376
145
  },
146
  "latency": {
147
  "unit": "s",
148
  "count": 2,
149
- "total": 1.1954854736328127,
150
- "mean": 0.5977427368164063,
151
- "stdev": 0.0011124267578124791,
152
- "p50": 0.5977427368164063,
153
- "p90": 0.5986326782226563,
154
- "p95": 0.5987439208984375,
155
- "p99": 0.5988329150390626,
156
  "values": [
157
- 0.5966303100585938,
158
- 0.5988551635742188
159
  ]
160
  },
161
  "throughput": {
162
  "unit": "images/s",
163
- "value": 1.6729605203168616
164
  },
165
  "energy": {
166
  "unit": "kWh",
167
- "cpu": 7.101243098958598e-06,
168
- "ram": 3.870880814651764e-06,
169
- "gpu": 3.8528503044999666e-05,
170
- "total": 4.9500626958610025e-05
171
  },
172
  "efficiency": {
173
  "unit": "images/kWh",
174
- "value": 20201.764329897287
175
  }
176
  }
177
  }
 
3
  "name": "cuda_inference_diffusers_text-to-image_CompVis/stable-diffusion-v1-4",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.6.0.dev20240917+cu124",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-to-image",
9
  "library": "diffusers",
 
76
  "cpu_ram_mb": 66697.261056,
77
  "system": "Linux",
78
  "machine": "x86_64",
79
+ "platform": "Linux-5.10.224-212.876.amzn2.x86_64-x86_64-with-glibc2.35",
80
  "processor": "x86_64",
81
  "python_version": "3.10.12",
82
  "gpu": [
 
88
  "optimum_benchmark_commit": null,
89
  "transformers_version": "4.44.2",
90
  "transformers_commit": null,
91
+ "accelerate_version": "0.34.2",
92
  "accelerate_commit": null,
93
+ "diffusers_version": "0.30.3",
94
  "diffusers_commit": null,
95
  "optimum_version": null,
96
  "optimum_commit": null,
 
104
  "load": {
105
  "memory": {
106
  "unit": "MB",
107
+ "max_ram": 4277.600256,
108
+ "max_global_vram": 6165.102592,
109
  "max_process_vram": 0.0,
110
+ "max_reserved": 5572.132864,
111
+ "max_allocated": 5512.962048
112
  },
113
  "latency": {
114
  "unit": "s",
115
  "count": 1,
116
+ "total": 44.723640625,
117
+ "mean": 44.723640625,
118
  "stdev": 0.0,
119
+ "p50": 44.723640625,
120
+ "p90": 44.723640625,
121
+ "p95": 44.723640625,
122
+ "p99": 44.723640625,
123
  "values": [
124
+ 44.723640625
125
  ]
126
  },
127
  "throughput": null,
128
  "energy": {
129
  "unit": "kWh",
130
+ "cpu": 0.0004285372342708317,
131
+ "ram": 0.00023485715990336945,
132
+ "gpu": 0.0006221054976840029,
133
+ "total": 0.0012854998918582039
134
  },
135
  "efficiency": null
136
  },
137
  "call": {
138
  "memory": {
139
  "unit": "MB",
140
+ "max_ram": 1473.953792,
141
+ "max_global_vram": 8530.690048,
142
  "max_process_vram": 0.0,
143
+ "max_reserved": 7879.000064,
144
+ "max_allocated": 6598.620672
145
  },
146
  "latency": {
147
  "unit": "s",
148
  "count": 2,
149
+ "total": 1.2121260986328126,
150
+ "mean": 0.6060630493164063,
151
+ "stdev": 0.0009461059570312069,
152
+ "p50": 0.6060630493164063,
153
+ "p90": 0.6068199340820313,
154
+ "p95": 0.6069145446777343,
155
+ "p99": 0.6069902331542968,
156
  "values": [
157
+ 0.605116943359375,
158
+ 0.6070091552734375
159
  ]
160
  },
161
  "throughput": {
162
  "unit": "images/s",
163
+ "value": 1.6499933482628995
164
  },
165
  "energy": {
166
  "unit": "kWh",
167
+ "cpu": 7.184782728135513e-06,
168
+ "ram": 3.927934372834812e-06,
169
+ "gpu": 3.983225408799723e-05,
170
+ "total": 5.094497118896756e-05
171
  },
172
  "efficiency": {
173
  "unit": "images/kWh",
174
+ "value": 19629.02278010427
175
  }
176
  }
177
  }