Clémentine
commited on
Commit
•
8dfa543
1
Parent(s):
97b27da
moved the submit to a tab since the results are becoming very long
Browse files
app.py
CHANGED
@@ -335,51 +335,40 @@ with demo:
|
|
335 |
with gr.TabItem("About", elem_id="llm-benchmark-tab-table", id=2):
|
336 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
337 |
|
338 |
-
|
339 |
-
|
340 |
-
gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
|
341 |
-
|
342 |
-
with gr.Column():
|
343 |
-
with gr.Accordion("✅ Finished Evaluations", open=False):
|
344 |
with gr.Row():
|
345 |
-
|
346 |
-
value=finished_eval_queue_df,
|
347 |
-
headers=EVAL_COLS,
|
348 |
-
datatype=EVAL_TYPES,
|
349 |
-
max_rows=5,
|
350 |
-
)
|
351 |
-
with gr.Accordion("🔄 Running Evaluation Queue", open=False):
|
352 |
-
with gr.Row():
|
353 |
-
running_eval_table = gr.components.Dataframe(
|
354 |
-
value=running_eval_queue_df,
|
355 |
-
headers=EVAL_COLS,
|
356 |
-
datatype=EVAL_TYPES,
|
357 |
-
max_rows=5,
|
358 |
-
)
|
359 |
|
360 |
-
|
361 |
-
|
362 |
-
|
363 |
-
|
364 |
-
|
365 |
-
|
366 |
-
|
367 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
368 |
|
369 |
-
with gr.Row():
|
370 |
-
refresh_button = gr.Button("Refresh")
|
371 |
-
refresh_button.click(
|
372 |
-
refresh,
|
373 |
-
inputs=[],
|
374 |
-
outputs=[
|
375 |
-
leaderboard_table,
|
376 |
-
finished_eval_table,
|
377 |
-
running_eval_table,
|
378 |
-
pending_eval_table,
|
379 |
-
],
|
380 |
-
)
|
381 |
-
with gr.Accordion("Submit a new model for evaluation"):
|
382 |
with gr.Row():
|
|
|
383 |
with gr.Column():
|
384 |
model_name_textbox = gr.Textbox(label="Model name")
|
385 |
revision_name_textbox = gr.Textbox(
|
@@ -413,6 +402,19 @@ with demo:
|
|
413 |
submission_result,
|
414 |
)
|
415 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
416 |
with gr.Row():
|
417 |
with gr.Accordion("📙 Citation", open=False):
|
418 |
citation_button = gr.Textbox(
|
|
|
335 |
with gr.TabItem("About", elem_id="llm-benchmark-tab-table", id=2):
|
336 |
gr.Markdown(LLM_BENCHMARKS_TEXT, elem_classes="markdown-text")
|
337 |
|
338 |
+
with gr.TabItem("✉️✨ Submit here! ", elem_id="llm-benchmark-tab-table", id=3):
|
339 |
+
with gr.Column():
|
|
|
|
|
|
|
|
|
340 |
with gr.Row():
|
341 |
+
gr.Markdown(EVALUATION_QUEUE_TEXT, elem_classes="markdown-text")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
342 |
|
343 |
+
with gr.Column():
|
344 |
+
with gr.Accordion("✅ Finished Evaluations", open=False):
|
345 |
+
with gr.Row():
|
346 |
+
finished_eval_table = gr.components.Dataframe(
|
347 |
+
value=finished_eval_queue_df,
|
348 |
+
headers=EVAL_COLS,
|
349 |
+
datatype=EVAL_TYPES,
|
350 |
+
max_rows=5,
|
351 |
+
)
|
352 |
+
with gr.Accordion("🔄 Running Evaluation Queue", open=False):
|
353 |
+
with gr.Row():
|
354 |
+
running_eval_table = gr.components.Dataframe(
|
355 |
+
value=running_eval_queue_df,
|
356 |
+
headers=EVAL_COLS,
|
357 |
+
datatype=EVAL_TYPES,
|
358 |
+
max_rows=5,
|
359 |
+
)
|
360 |
+
|
361 |
+
with gr.Accordion("⏳ Pending Evaluation Queue", open=False):
|
362 |
+
with gr.Row():
|
363 |
+
pending_eval_table = gr.components.Dataframe(
|
364 |
+
value=pending_eval_queue_df,
|
365 |
+
headers=EVAL_COLS,
|
366 |
+
datatype=EVAL_TYPES,
|
367 |
+
max_rows=5,
|
368 |
+
)
|
369 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
370 |
with gr.Row():
|
371 |
+
gr.Markdown("# SUBMIT HERE", elem_classes="markdown-text")
|
372 |
with gr.Column():
|
373 |
model_name_textbox = gr.Textbox(label="Model name")
|
374 |
revision_name_textbox = gr.Textbox(
|
|
|
402 |
submission_result,
|
403 |
)
|
404 |
|
405 |
+
with gr.Row():
|
406 |
+
refresh_button = gr.Button("Refresh")
|
407 |
+
refresh_button.click(
|
408 |
+
refresh,
|
409 |
+
inputs=[],
|
410 |
+
outputs=[
|
411 |
+
leaderboard_table,
|
412 |
+
finished_eval_table,
|
413 |
+
running_eval_table,
|
414 |
+
pending_eval_table,
|
415 |
+
],
|
416 |
+
)
|
417 |
+
|
418 |
with gr.Row():
|
419 |
with gr.Accordion("📙 Citation", open=False):
|
420 |
citation_button = gr.Textbox(
|