Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -201,7 +201,7 @@ with gr.Blocks(css=STYLE) as hf_endpoint:
|
|
201 |
with gr.Column():
|
202 |
gr.Markdown("""## Target model
|
203 |
|
204 |
-
|
205 |
repository_selector = gr.Textbox(
|
206 |
value="NousResearch/Nous-Hermes-Llama2-70b",
|
207 |
interactive=False,
|
@@ -212,7 +212,7 @@ Import a model from the Hugging Face hub""")
|
|
212 |
with gr.Column():
|
213 |
gr.Markdown("""## Target model version(branch)
|
214 |
|
215 |
-
|
216 |
revision_selector = gr.Textbox(
|
217 |
value=f"main",
|
218 |
interactive=False,
|
@@ -224,14 +224,12 @@ Specify the branch name""")
|
|
224 |
with gr.Column():
|
225 |
gr.Markdown("""## Endpoint name
|
226 |
|
227 |
-
|
228 |
endpoint_name_input = gr.Textbox(show_label=False, elem_classes=["no-label", "small-big"])
|
229 |
|
230 |
with gr.Row():
|
231 |
with gr.Column():
|
232 |
-
gr.Markdown("""## Cloud Provider
|
233 |
-
|
234 |
-
Choose between Amazon Web Services and Microsoft Azure""")
|
235 |
provider_selector = gr.Dropdown(
|
236 |
choices=providers.keys(),
|
237 |
interactive=True,
|
@@ -240,9 +238,7 @@ Choose between Amazon Web Services and Microsoft Azure""")
|
|
240 |
)
|
241 |
|
242 |
with gr.Column():
|
243 |
-
gr.Markdown("""## Cloud Region
|
244 |
-
|
245 |
-
Choose one of the regions from each cloud provider""")
|
246 |
region_selector = gr.Dropdown(
|
247 |
[],
|
248 |
value="",
|
@@ -271,9 +267,7 @@ Choose one of the regions from each cloud provider""")
|
|
271 |
)
|
272 |
|
273 |
with gr.Column():
|
274 |
-
gr.Markdown("""## Select Compute Instance Type
|
275 |
-
|
276 |
-
Select a CPU or GPU accelerated compute option for inference""")
|
277 |
compute_selector = gr.Dropdown(
|
278 |
[],
|
279 |
value="",
|
@@ -285,9 +279,7 @@ Select a CPU or GPU accelerated compute option for inference""")
|
|
285 |
with gr.Row():
|
286 |
with gr.Row():
|
287 |
with gr.Column():
|
288 |
-
gr.Markdown("""## Min Number of Nodes
|
289 |
-
|
290 |
-
Automatically scale the number of replicas based on load and compute usage""")
|
291 |
min_node_selector = gr.Number(
|
292 |
value=1,
|
293 |
interactive=True,
|
@@ -296,9 +288,7 @@ Select a CPU or GPU accelerated compute option for inference""")
|
|
296 |
)
|
297 |
|
298 |
with gr.Column():
|
299 |
-
gr.Markdown("""## Max Number of Nodes
|
300 |
-
|
301 |
-
Automatically scale the number of replicas based on load and compute usage""")
|
302 |
max_node_selector = gr.Number(
|
303 |
value=1,
|
304 |
interactive=True,
|
@@ -307,9 +297,7 @@ Select a CPU or GPU accelerated compute option for inference""")
|
|
307 |
)
|
308 |
|
309 |
with gr.Column():
|
310 |
-
gr.Markdown("""## Security Level
|
311 |
-
|
312 |
-
Choose your endpoint's level of privacy""")
|
313 |
security_selector = gr.Radio(
|
314 |
choices=["Protected", "Public", "Private"],
|
315 |
value="Public",
|
|
|
201 |
with gr.Column():
|
202 |
gr.Markdown("""## Target model
|
203 |
|
204 |
+
Model from the Hugging Face hub""")
|
205 |
repository_selector = gr.Textbox(
|
206 |
value="NousResearch/Nous-Hermes-Llama2-70b",
|
207 |
interactive=False,
|
|
|
212 |
with gr.Column():
|
213 |
gr.Markdown("""## Target model version(branch)
|
214 |
|
215 |
+
Branch name of the Model""")
|
216 |
revision_selector = gr.Textbox(
|
217 |
value=f"main",
|
218 |
interactive=False,
|
|
|
224 |
with gr.Column():
|
225 |
gr.Markdown("""## Endpoint name
|
226 |
|
227 |
+
Name for your new endpoint""")
|
228 |
endpoint_name_input = gr.Textbox(show_label=False, elem_classes=["no-label", "small-big"])
|
229 |
|
230 |
with gr.Row():
|
231 |
with gr.Column():
|
232 |
+
gr.Markdown("""## Cloud Provider""")
|
|
|
|
|
233 |
provider_selector = gr.Dropdown(
|
234 |
choices=providers.keys(),
|
235 |
interactive=True,
|
|
|
238 |
)
|
239 |
|
240 |
with gr.Column():
|
241 |
+
gr.Markdown("""## Cloud Region""")
|
|
|
|
|
242 |
region_selector = gr.Dropdown(
|
243 |
[],
|
244 |
value="",
|
|
|
267 |
)
|
268 |
|
269 |
with gr.Column():
|
270 |
+
gr.Markdown("""## Select Compute Instance Type""")
|
|
|
|
|
271 |
compute_selector = gr.Dropdown(
|
272 |
[],
|
273 |
value="",
|
|
|
279 |
with gr.Row():
|
280 |
with gr.Row():
|
281 |
with gr.Column():
|
282 |
+
gr.Markdown("""## Min Number of Nodes""")
|
|
|
|
|
283 |
min_node_selector = gr.Number(
|
284 |
value=1,
|
285 |
interactive=True,
|
|
|
288 |
)
|
289 |
|
290 |
with gr.Column():
|
291 |
+
gr.Markdown("""## Max Number of Nodes""")
|
|
|
|
|
292 |
max_node_selector = gr.Number(
|
293 |
value=1,
|
294 |
interactive=True,
|
|
|
297 |
)
|
298 |
|
299 |
with gr.Column():
|
300 |
+
gr.Markdown("""## Security Level""")
|
|
|
|
|
301 |
security_selector = gr.Radio(
|
302 |
choices=["Protected", "Public", "Private"],
|
303 |
value="Public",
|