run_id
large_stringlengths
64
64
timestamp
unknown
model_name_or_path
large_stringclasses
5 values
unitxt_recipe
large_stringlengths
326
371
quantization_type
large_stringclasses
1 value
quantization_bit_count
large_stringclasses
1 value
inference_runtime_s
float64
1.05
37.4
generation_args
large_stringclasses
1 value
model_args
large_stringclasses
5 values
inference_engine
large_stringclasses
1 value
packages_versions
large_stringclasses
1 value
scores
large_stringlengths
174
240
num_gpu
int64
1
1
device
large_stringclasses
1 value
4bdd29c2978a00ea6376cdca5032c6926a2b6ede7bc6adc14fef715ef7003b97
"2024-12-22T12:34:48.914000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.5454
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.55, 'score': 0.55, 'score_name': 'accuracy', 'accuracy_ci_low': 0.45, 'accuracy_ci_high': 0.64, 'score_ci_low': 0.45, 'score_ci_high': 0.64}
1
a100_80gb
dfb44de0eb91c843ebc4becde08809430b3e49eba890a522fc60b8d3c4e3ecb6
"2024-12-22T12:34:54.910000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.029511
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.49, 'score': 0.49, 'score_name': 'accuracy', 'accuracy_ci_low': 0.3983237850218569, 'accuracy_ci_high': 0.58, 'score_ci_low': 0.3983237850218569, 'score_ci_high': 0.58}
1
a100_80gb
4a5300ae2428612a84fce175d9564d05cb17d30df41978926fd09b534ba65725
"2024-12-22T12:33:42.649000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.399628
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.72, 'score': 0.72, 'score_name': 'accuracy', 'accuracy_ci_low': 0.6258565731259789, 'accuracy_ci_high': 0.8, 'score_ci_low': 0.6258565731259789, 'score_ci_high': 0.8}
1
a100_80gb
4331d2b5d0cb2e6685cc6d9e614d0e6145b6c813d48b6178534ec93901451d54
"2024-12-22T12:33:46.685000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.265079
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
aafbc0a233430ef6af3a9a4b5e7768e19ea41319ccacdb82b2b13d47dd963f2c
"2024-12-22T12:33:50.678000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.217314
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.67, 'score': 0.67, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
42b9aed4b4d9faff7d7bf6a0e61f50daf7e3aed4964b5619a1e5da6690213fa5
"2024-12-22T12:33:53.837000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.384058
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.53, 'score_ci_high': 0.72}
1
a100_80gb
c7b522cb0b1e402d9a3441182ba9b8f84e0843f86c10df868eb066dfbc53f306
"2024-12-22T12:33:56.967000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.361346
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.72, 'score': 0.72, 'score_name': 'accuracy', 'accuracy_ci_low': 0.62, 'accuracy_ci_high': 0.8, 'score_ci_low': 0.62, 'score_ci_high': 0.8}
1
a100_80gb
6d58a4b913d707896cd46b84dbd6e77135293b6dc2b766ec5f17b962a13504b9
"2024-12-22T12:34:00.093000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.361867
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.61, 'score': 0.61, 'score_name': 'accuracy', 'accuracy_ci_low': 0.51, 'accuracy_ci_high': 0.7, 'score_ci_low': 0.51, 'score_ci_high': 0.7}
1
a100_80gb
2f89bd5181e338587a8c91483250a0e0e6ed7f6e4b073b6330762eee7f30d5e4
"2024-12-22T12:34:03.840000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.398619
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.68, 'score': 0.68, 'score_name': 'accuracy', 'accuracy_ci_low': 0.58, 'accuracy_ci_high': 0.76, 'score_ci_low': 0.58, 'score_ci_high': 0.76}
1
a100_80gb
cce88e030a950056879992f497c706cdab0ed4fc7d23a842b6dcb41e58d89bbb
"2024-12-22T12:34:07.931000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.417539
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
abb0fa326d137f680cf68ea94736094e875ae36b730b9757e50d23d055b7ebb2
"2024-12-22T12:34:12.088000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.373913
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.68, 'score': 0.68, 'score_name': 'accuracy', 'accuracy_ci_low': 0.58, 'accuracy_ci_high': 0.76, 'score_ci_low': 0.58, 'score_ci_high': 0.76}
1
a100_80gb
3de5f555f462082c94ea24a72db474db8137ea46239d83d71fb2c9e0b863a738
"2024-12-22T12:34:16.424000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.545616
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.57, 'score_ci_high': 0.74}
1
a100_80gb
bb67e8c4bd70c27642310d96b7bdd050e8fb4048fda4f18d7919e9089d6054ad
"2024-12-22T12:33:52Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.520826
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.57, 'score': 0.57, 'score_name': 'accuracy', 'accuracy_ci_low': 0.47, 'accuracy_ci_high': 0.6649419245613979, 'score_ci_low': 0.47, 'score_ci_high': 0.6649419245613979}
1
a100_80gb
f3144d0fbf79ddc2db3619b2c15cfbd9cd0d39cfe9acf690349b7f563574d7db
"2024-12-22T12:33:55.332000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.56121
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.4969176058823072, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.4969176058823072, 'score_ci_high': 0.69}
1
a100_80gb
ec71ab7a2aa30cf2612311da5d8fd90c7eefffc55c05866bc8b7b2413aa9af2e
"2024-12-22T12:33:58.709000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.592958
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.55, 'score_ci_high': 0.74}
1
a100_80gb
c96852a729547270580d7097f41c7465acfc0908733495b1772d37b1f696220d
"2024-12-22T12:34:02.013000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.545659
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.67, 'score': 0.67, 'score_name': 'accuracy', 'accuracy_ci_low': 0.58, 'accuracy_ci_high': 0.76, 'score_ci_low': 0.58, 'score_ci_high': 0.76}
1
a100_80gb
4e864fe6d92378cec77320504481f123f20cb9879101a990065935fdf76d91a9
"2024-12-22T12:34:05.395000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.607895
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.6, 'score': 0.6, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.5, 'score_ci_high': 0.69}
1
a100_80gb
cde461f831e55066b1fc0b341bd85b7c132cd6f85b2aea545e6b768d2ffec3cb
"2024-12-22T12:34:10.829000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.769473
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.69, 'score': 0.69, 'score_name': 'accuracy', 'accuracy_ci_low': 0.59, 'accuracy_ci_high': 0.78, 'score_ci_low': 0.59, 'score_ci_high': 0.78}
1
a100_80gb
eb33d32ec28d492f61c47455e78019addce152e0c9e3a3d13fbbf0dfaeb543d3
"2024-12-22T12:34:14.218000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.615149
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.64, 'score': 0.64, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.53, 'score_ci_high': 0.72}
1
a100_80gb
9971d5dd77cfc48f4e713aba018df65a7e28fdb511b26193ae054394597a6afe
"2024-12-22T12:34:18.442000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.452065
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5659039697599085, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.5659039697599085, 'score_ci_high': 0.75}
1
a100_80gb
2bb9ebf1f37763faaf1e99892f4a279291fa9629f02a3d9f470d968d20cf5b74
"2024-12-22T12:34:22.448000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.24063
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.56, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.56, 'score_ci_high': 0.75}
1
a100_80gb
0cde18a945adfef008781b6da922713fd57e1184d49072ea74b6f1dfbbefa26e
"2024-12-22T12:34:25.832000Z"
meta-llama/Meta-Llama-3-8B-Instruct
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.569288
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Meta-Llama-3-8B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.56, 'accuracy_ci_high': 0.76, 'score_ci_low': 0.56, 'score_ci_high': 0.76}
1
a100_80gb
74d50df6edb95ea74fe213d3152a15f4086b24246b6aba3c3819e7063b02738a
"2024-12-22T12:33:35.099000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.644391
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.57, 'score': 0.57, 'score_name': 'accuracy', 'accuracy_ci_low': 0.47, 'accuracy_ci_high': 0.67, 'score_ci_low': 0.47, 'score_ci_high': 0.67}
1
a100_80gb
b00a01a2893517202df9925f28e511d82bd518cfeaf2ad8696c755bf09d3607d
"2024-12-22T12:33:37.552000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.741062
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.53, 'score_ci_high': 0.71}
1
a100_80gb
8730d6cc6be21f6abab4849684ff48c884b992071d5c244d08f10e94782cea82
"2024-12-22T12:33:39.960000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.670636
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.49, 'score_ci_high': 0.68}
1
a100_80gb
a197facf856730aad6b0b8ad66456abfb7a42b0fe0393a30ef1e8df72a6d5c14
"2024-12-22T12:33:43.355000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.677186
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.5, 'score_ci_high': 0.69}
1
a100_80gb
d733142a4e87baae4baf1e867024aba49e409db40ff3f44585a13e8eb8da1fa1
"2024-12-22T12:33:46.632000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.548501
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.53, 'score_ci_high': 0.71}
1
a100_80gb
d090df33fa2d7ec332041bd1d973001720ecc73a9ede27f3740e8da2caea8d82
"2024-12-22T12:33:49.888000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.543147
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.55, 'score': 0.55, 'score_name': 'accuracy', 'accuracy_ci_low': 0.46, 'accuracy_ci_high': 0.65, 'score_ci_low': 0.46, 'score_ci_high': 0.65}
1
a100_80gb
2aaf8118d806cefd8da422e535c09f49e35c4b2977a769f5c1e17251e893f1e2
"2024-12-22T12:33:52.750000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.135746
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.61, 'score': 0.61, 'score_name': 'accuracy', 'accuracy_ci_low': 0.515078710624518, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.515078710624518, 'score_ci_high': 0.71}
1
a100_80gb
446633803c8caa345aa609c4dd78dedc1b5e188328055a8ecba1e4a3dedb1013
"2024-12-22T12:33:55.123000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.654844
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.54, 'score': 0.54, 'score_name': 'accuracy', 'accuracy_ci_low': 0.45, 'accuracy_ci_high': 0.6471782200590656, 'score_ci_low': 0.45, 'score_ci_high': 0.6471782200590656}
1
a100_80gb
1722589dcf239e041ac1f2120864262b172907b9ce2c88b0c06dc17a86d7df30
"2024-12-22T12:33:57.540000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.693652
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.54, 'score': 0.54, 'score_name': 'accuracy', 'accuracy_ci_low': 0.4417664948762789, 'accuracy_ci_high': 0.64, 'score_ci_low': 0.4417664948762789, 'score_ci_high': 0.64}
1
a100_80gb
5e7832b4ca1c4be2039a0908d4cb153a109e6422e29fbd51f430d8357f890b78
"2024-12-22T12:34:01.644000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.367253
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.49, 'score_ci_high': 0.68}
1
a100_80gb
d7fffae22f4f797f814d8f83093df856e652cb7a55b21b66e48962730d1c0975
"2024-12-22T12:33:58.250000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.391469
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.67, 'score': 0.67, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.7550250186394021, 'score_ci_low': 0.57, 'score_ci_high': 0.7550250186394021}
1
a100_80gb
68613f6de62b0da53d7066f77099bb0d406e3ef73170ba2d0d8d279a00cb2d31
"2024-12-22T12:34:02.349000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.356816
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.53, 'score_ci_high': 0.72}
1
a100_80gb
29481cdf2f7a5379744272e277c37768273563ad0819ab3bada30269efcbadf9
"2024-12-22T12:34:05.495000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.392672
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.55, 'score_ci_high': 0.74}
1
a100_80gb
707d861e39c2f005714038ca8cb18cf37110edae22790d8ae21e29bf47b80129
"2024-12-22T12:34:08.619000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.378329
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.58, 'score': 0.58, 'score_name': 'accuracy', 'accuracy_ci_low': 0.48, 'accuracy_ci_high': 0.6745702951080432, 'score_ci_low': 0.48, 'score_ci_high': 0.6745702951080432}
1
a100_80gb
fa008c54d7cabfd7cef111867581e14b15ef4411a935d78e9157b5ebae745b6f
"2024-12-22T12:34:11.839000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.449901
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.64, 'score': 0.64, 'score_name': 'accuracy', 'accuracy_ci_low': 0.54, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.54, 'score_ci_high': 0.72}
1
a100_80gb
a27659b95a8220fecf805ad2ef0394ae2bc1470066ac8d563f53e64a6ab3107b
"2024-12-22T12:34:15.024000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.392196
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.52, 'score_ci_high': 0.71}
1
a100_80gb
2f91a4a61add33ee65137f432a8240d1770747c466302b602eec1b1fd3bfbcf0
"2024-12-22T12:34:19.962000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.287892
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.56, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.56, 'score_ci_high': 0.75}
1
a100_80gb
a286396c06506fd7d881600de33f52ad1b02e88e4547a0f61610f5148d280e7c
"2024-12-22T12:34:23.091000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.367312
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.6820475832535976, 'score_ci_low': 0.49, 'score_ci_high': 0.6820475832535976}
1
a100_80gb
81b9b3cc29c4615b5dee1d8eee9396703b09548ffbc307e58985592d41e8648d
"2024-12-22T12:34:27.244000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.402959
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.61, 'score': 0.61, 'score_name': 'accuracy', 'accuracy_ci_low': 0.51, 'accuracy_ci_high': 0.7, 'score_ci_low': 0.51, 'score_ci_high': 0.7}
1
a100_80gb
5c0462c7ca18faa30bce79d5f645dee79e3685b3fd302ac8f4c1cd317ab036a4
"2024-12-22T12:34:30.407000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.astronomy,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_capitals_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.402095
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.6, 'score': 0.6, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.7, 'score_ci_low': 0.5, 'score_ci_high': 0.7}
1
a100_80gb
2cfac69aa40297fa0cc8a3fcfef53473899f34d0454220d0072acf5bf45850a6
"2024-12-22T12:33:50.134000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.506459
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.42, 'score': 0.42, 'score_name': 'accuracy', 'accuracy_ci_low': 0.32, 'accuracy_ci_high': 0.53, 'score_ci_low': 0.32, 'score_ci_high': 0.53}
1
a100_80gb
0cee934ed47bab79b91c9625cb339221cd35f2bbd00c28c7dc0f30797bb8d972
"2024-12-22T12:33:54.366000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.484873
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.28, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.28, 'score_ci_high': 0.47}
1
a100_80gb
1774d3e27263c4ac628651cff02a8a422b878027d0929d4241a52dadb7f61032
"2024-12-22T12:33:57.601000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.487871
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.5, 'score_ci_low': 0.31, 'score_ci_high': 0.5}
1
a100_80gb
696ae7433be855101610140c79e35f2172252dd959d038d03d06f2ebfba92005
"2024-12-22T12:34:00.873000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.503329
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.22, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.22, 'score_ci_high': 0.41}
1
a100_80gb
ec2facd3647bf3cd42755c8b7627dded4b2d1c1b6e5174fe756fa2b99ddfddc7
"2024-12-22T12:34:04.203000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.570994
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.41, 'score': 0.41, 'score_name': 'accuracy', 'accuracy_ci_low': 0.32, 'accuracy_ci_high': 0.51, 'score_ci_low': 0.32, 'score_ci_high': 0.51}
1
a100_80gb
3ac8276356da15078a3510e7fb4d731a075144a86b2da40bc24a13b7c478b988
"2024-12-22T12:34:08.562000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.601345
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.27, 'accuracy_ci_high': 0.46, 'score_ci_low': 0.27, 'score_ci_high': 0.46}
1
a100_80gb
c5194a723a911779fce42a7f4210915e14b49117f3831784ea3857c15fff913d
"2024-12-22T12:34:13.560000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.232348
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.43, 'score': 0.43, 'score_name': 'accuracy', 'accuracy_ci_low': 0.33, 'accuracy_ci_high': 0.52, 'score_ci_low': 0.33, 'score_ci_high': 0.52}
1
a100_80gb
083ec434597e37f8d51036c42807b504e2f6c1143c7c037669c34473310f3c10
"2024-12-22T12:34:16.771000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.458793
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.34, 'score': 0.34, 'score_name': 'accuracy', 'accuracy_ci_low': 0.25, 'accuracy_ci_high': 0.4411435322379134, 'score_ci_low': 0.25, 'score_ci_high': 0.4411435322379134}
1
a100_80gb
40e5b29dd3d35642c0595e4c941ed77c4fab32f75619afcaa76bbff03e646386
"2024-12-22T12:34:20.010000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.450526
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.27, 'score': 0.27, 'score_name': 'accuracy', 'accuracy_ci_low': 0.19, 'accuracy_ci_high': 0.36, 'score_ci_low': 0.19, 'score_ci_high': 0.36}
1
a100_80gb
83e9cb07569a0d8afb9157425569f553bb5ff012159f45dd1c57c0bef7427cd1
"2024-12-22T12:34:24.301000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.high_school_chemistry,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.490048
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.38, 'score': 0.38, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.48, 'score_ci_low': 0.29, 'score_ci_high': 0.48}
1
a100_80gb
1e0698e0c151b625a36fea4c903ce01b081b209839bd340be08fb16b2d174892
"2024-12-22T12:33:32.516000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.264828
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.553669922381089, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.553669922381089, 'score_ci_high': 0.74}
1
a100_80gb
12b2f685637d05458205e0a1f0fdb85bcb25c5db2bcd85533ab77fee0a111b29
"2024-12-22T12:33:35.584000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.317277
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.71, 'score': 0.71, 'score_name': 'accuracy', 'accuracy_ci_low': 0.62, 'accuracy_ci_high': 0.7962425586740983, 'score_ci_low': 0.62, 'score_ci_high': 0.7962425586740983}
1
a100_80gb
fbd25f3d3ea5d4b5b46da0840a52d533332222947832d1a4fae5f720e55cf3ac
"2024-12-22T12:33:39.762000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.305476
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.53, 'score_ci_high': 0.71}
1
a100_80gb
aeb9765ce5910b786c08fed0f4e1198537b3a9091098d0b667322a83a8a49387
"2024-12-22T12:33:43.254000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.769892
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.67, 'score': 0.67, 'score_name': 'accuracy', 'accuracy_ci_low': 0.58, 'accuracy_ci_high': 0.76, 'score_ci_low': 0.58, 'score_ci_high': 0.76}
1
a100_80gb
4cf8bd89032fe5d6fe2f63c21e86e4e964b3b95b2bdfb272998657fd305cb93d
"2024-12-22T12:33:47.284000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.275429
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.53, 'score_ci_high': 0.72}
1
a100_80gb
3095199e69735ff8d20ee742e620484338cd74ffa81fe459e0e86974496c6c19
"2024-12-22T12:33:50.360000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.334187
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.56, 'score': 0.56, 'score_name': 'accuracy', 'accuracy_ci_low': 0.47, 'accuracy_ci_high': 0.67, 'score_ci_low': 0.47, 'score_ci_high': 0.67}
1
a100_80gb
24489199691d6320e29f61ad3758ea4733267698bb019deaea55c5218b6002d2
"2024-12-22T12:33:53.974000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.868292
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.58, 'score': 0.58, 'score_name': 'accuracy', 'accuracy_ci_low': 0.48, 'accuracy_ci_high': 0.67, 'score_ci_low': 0.48, 'score_ci_high': 0.67}
1
a100_80gb
5808ef96ee24268b478c69290d1a89d1dfd3b67a3b7f726b51f54e6f0e90ed78
"2024-12-22T12:33:57.117000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.378497
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.54, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.54, 'score_ci_high': 0.72}
1
a100_80gb
bd588c200b7b88ca2416392f16bf8c8ca451e146f770164e457297d773b5c16a
"2024-12-22T12:34:01.083000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_roman_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.195533
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.56, 'score': 0.56, 'score_name': 'accuracy', 'accuracy_ci_low': 0.4665194662143706, 'accuracy_ci_high': 0.66, 'score_ci_low': 0.4665194662143706, 'score_ci_high': 0.66}
1
a100_80gb
2b29d59480d487af54aba41e257fd5a95f5373fb6d37fe1cf0d97dafd101aa82
"2024-12-22T12:34:05.557000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.logical_fallacies,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopic.enumerator_keyboard_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.721612
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.39, 'score': 0.39, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29762240727713996, 'accuracy_ci_high': 0.48, 'score_ci_low': 0.29762240727713996, 'score_ci_high': 0.48}
1
a100_80gb
dbb2295daeb4d619d06182bf510bc485fd479730d70ab24cb9fa222eaefe74b5
"2024-12-22T12:33:29.228000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.537066
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.59, 'score': 0.59, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.69, 'score_ci_low': 0.49, 'score_ci_high': 0.69}
1
a100_80gb
836842dc8f42366100758bb57ec14f3e5244dca6e08bdcb8ba7aef861647dd25
"2024-12-22T12:33:34.121000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.719037
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.68, 'score': 0.68, 'score_name': 'accuracy', 'accuracy_ci_low': 0.59, 'accuracy_ci_high': 0.77, 'score_ci_low': 0.59, 'score_ci_high': 0.77}
1
a100_80gb
3f8ac97ac20693ee4391421780586bb4987639f6269ab73527a0fd447107918b
"2024-12-22T12:33:38.431000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.109465
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.61, 'score': 0.61, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.52, 'score_ci_high': 0.71}
1
a100_80gb
f0dffae8af2e360e76ac4dfb73b6f570066e07a598fa66da97fe09d5e546a993
"2024-12-22T12:33:44.102000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.515151
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
478d5e889ac4675cf3db50077d42eb8cd4160ad81c4c01cf178c4884f73d53ed
"2024-12-22T12:33:49.089000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.846056
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
2d71fa7c3ea4bdb37ac6c486a8bf73ba001ee8bb078398d34a0cbf5e21192827
"2024-12-22T12:33:53.923000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.647066
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.53, 'score_ci_high': 0.71}
1
a100_80gb
5d034ca52315769e9259487b23c0e1c9ec745bbe03f41b913c8ed3ae19197204
"2024-12-22T12:33:58.107000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.99859
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.58, 'score': 0.58, 'score_name': 'accuracy', 'accuracy_ci_low': 0.49, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.49, 'score_ci_high': 0.68}
1
a100_80gb
4ac6073acf86631df44bf49394df2ada891e704e6d197d52ddb7da9c457e2a24
"2024-12-22T12:34:03.476000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.188039
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.66, 'score': 0.66, 'score_name': 'accuracy', 'accuracy_ci_low': 0.57, 'accuracy_ci_high': 0.75, 'score_ci_low': 0.57, 'score_ci_high': 0.75}
1
a100_80gb
ab83b23eeb0b0d25420ce19ea396988e07eaff7681d9fcaa83af34679bf3726c
"2024-12-22T12:34:07.653000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.984943
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.52, 'score_ci_high': 0.71}
1
a100_80gb
d3910ed2278c26d2e9147d2e00c1e282483419be5d9bc9f232cf84af5ef4e08b
"2024-12-22T12:34:12.400000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.medical_genetics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithTopicHelm.enumerator_capitals_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.543401
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.64, 'score': 0.64, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.7303584532744878, 'score_ci_low': 0.55, 'score_ci_high': 0.7303584532744878}
1
a100_80gb
0d5805e8c264cddc06ecb2cfe94514f707ff1b4e5f13491de3c45d7ce3240b36
"2024-12-22T12:33:25.319000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.833876
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.25776392058592545, 'accuracy_ci_high': 0.45319666903395844, 'score_ci_low': 0.25776392058592545, 'score_ci_high': 0.45319666903395844}
1
a100_80gb
54ee358096b39e90caa8bd40173ff201eefa1f49c65f1cbe789c0a28e1da78dc
"2024-12-22T12:33:30.777000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.093946
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.37, 'score': 0.37, 'score_name': 'accuracy', 'accuracy_ci_low': 0.29, 'accuracy_ci_high': 0.47, 'score_ci_low': 0.29, 'score_ci_high': 0.47}
1
a100_80gb
ba0efd3bc1a8ec35e314cc989fc6a897da21c58d79d2f11edf80ef337a710e32
"2024-12-22T12:33:35.189000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.019069
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.42, 'score': 0.42, 'score_name': 'accuracy', 'accuracy_ci_low': 0.32, 'accuracy_ci_high': 0.52, 'score_ci_low': 0.32, 'score_ci_high': 0.52}
1
a100_80gb
48b3c4e95220f0dffabfe30ccfec90c2a8d0aaaab75dd1362d828738230e5976
"2024-12-22T12:33:40.377000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.128066
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.4, 'score': 0.4, 'score_name': 'accuracy', 'accuracy_ci_low': 0.31, 'accuracy_ci_high': 0.49, 'score_ci_low': 0.31, 'score_ci_high': 0.49}
1
a100_80gb
f9fbbc9c3f07b69f2b7258e092a32a3f28a5b6d64ffbdd36a67fae0af8296a3a
"2024-12-22T12:33:45.306000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.526795
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.28, 'score': 0.28, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.2, 'score_ci_high': 0.37}
1
a100_80gb
7e5ab7ad518412798f6df54e1b243f78dac75b62f514ca153862b64a04e2d3d9
"2024-12-22T12:33:51.744000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.417028
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.41, 'score_ci_low': 0.23, 'score_ci_high': 0.41}
1
a100_80gb
d8ada9b8b833f866352cbec14b04ef03f4b3f969dcec53c04a09bb4446395fcc
"2024-12-22T12:33:56.118000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.995877
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.33, 'score': 0.33, 'score_name': 'accuracy', 'accuracy_ci_low': 0.24, 'accuracy_ci_high': 0.43, 'score_ci_low': 0.24, 'score_ci_high': 0.43}
1
a100_80gb
05b1cddbae3ee04b764762e85c9285230a07f277a5e60c55dc22905dfef4edd0
"2024-12-22T12:34:00.643000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.12793
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.27, 'score': 0.27, 'score_name': 'accuracy', 'accuracy_ci_low': 0.2, 'accuracy_ci_high': 0.37, 'score_ci_low': 0.2, 'score_ci_high': 0.37}
1
a100_80gb
fdd0e3488f9e84ad2a2a0d2afd2caee1fb39ce6312731d85eb4d57adc196905f
"2024-12-22T12:34:05.155000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.109143
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.35, 'score': 0.35, 'score_name': 'accuracy', 'accuracy_ci_low': 0.26, 'accuracy_ci_high': 0.45, 'score_ci_low': 0.26, 'score_ci_high': 0.45}
1
a100_80gb
e3aa2f61dc0fe7625967f5c5b53645974fc8ff843aabc85d09a3c99c3c1273a4
"2024-12-22T12:34:10.739000Z"
meta-llama/Llama-3.2-1B-Instruct
card=cards.mmlu.business_ethics,demos_pool_size=100,num_demos=5,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_roman_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
4.202479
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "meta-llama/Llama-3.2-1B-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.31, 'score': 0.31, 'score_name': 'accuracy', 'accuracy_ci_low': 0.23, 'accuracy_ci_high': 0.4, 'score_ci_low': 0.23, 'score_ci_high': 0.4}
1
a100_80gb
0a4a14f91295a70da0659d0dff7c730f4f4211ad4369961cafb35514be7a9c67
"2024-12-22T12:33:07.112000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.273764
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.7297995533819346, 'score_ci_low': 0.53, 'score_ci_high': 0.7297995533819346}
1
a100_80gb
9b9efb7dece6b86ecd92c607c9783f240c6fa832c46eee2dcebdeb1953819f70
"2024-12-22T12:33:10.162000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.300337
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.52, 'score_ci_high': 0.72}
1
a100_80gb
64ba7fe65336f9a32659b9ac905e2b51d5c014a441c140673eb5b47e4cdee88d
"2024-12-22T12:33:13.298000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.320192
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.63, 'score': 0.63, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5397898919937004, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.5397898919937004, 'score_ci_high': 0.72}
1
a100_80gb
621f11aad76feca7233c7bea0d478c57b206c0c6876bd718cb26d16c3d4ca35d
"2024-12-22T12:33:16.370000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.321378
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.53, 'accuracy_ci_high': 0.7148215943456397, 'score_ci_low': 0.53, 'score_ci_high': 0.7148215943456397}
1
a100_80gb
1186a12118896905117a90020ef92fbfc58adf22c7d498fb898f58bcc92208fa
"2024-12-22T12:33:19.983000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_orLower_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.37016
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.6, 'score': 0.6, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.5, 'score_ci_high': 0.68}
1
a100_80gb
b610c5e32eda362512ee042d6006b77a4f0d4769f53d2fc6a762ece8968ab295
"2024-12-22T12:33:24.390000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_keyboard_choicesSeparator_orLower_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.642099
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.64, 'score': 0.64, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.72, 'score_ci_low': 0.55, 'score_ci_high': 0.72}
1
a100_80gb
c4ec80929acf9fc21a1a21e5e07cbdd83092b18e9b0edd631189073154b62ccc
"2024-12-22T12:33:28.510000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_space_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.348509
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.72, 'score': 0.72, 'score_name': 'accuracy', 'accuracy_ci_low': 0.62, 'accuracy_ci_high': 0.79, 'score_ci_low': 0.62, 'score_ci_high': 0.79}
1
a100_80gb
9655bd4ffabf3801b7405c60dbe7e22ff352880fff29a224c09cd34d538e6d36
"2024-12-22T12:33:32.177000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_space_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.903735
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.64, 'score': 0.64, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.55, 'score_ci_high': 0.74}
1
a100_80gb
234c8e19801bff48cac71bb4cfa324fcd93decc0bf9711eb77b62b59d23f519d
"2024-12-22T12:33:35.318000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.354084
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.68, 'score': 0.68, 'score_name': 'accuracy', 'accuracy_ci_low': 0.58, 'accuracy_ci_high': 0.7548295784212484, 'score_ci_low': 0.58, 'score_ci_high': 0.7548295784212484}
1
a100_80gb
a33dd45b937beb0c95165d1ada6cb05aefec134283cfb9a6364a92011c591057
"2024-12-22T12:33:38.477000Z"
mistralai/Mistral-7B-Instruct-v0.3
card=cards.mmlu.sociology,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_newline_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.391768
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "mistralai/Mistral-7B-Instruct-v0.3", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.68, 'score': 0.68, 'score_name': 'accuracy', 'accuracy_ci_low': 0.59, 'accuracy_ci_high': 0.77, 'score_ci_low': 0.59, 'score_ci_high': 0.77}
1
a100_80gb
3d63cebfdc0665988bacab1070ddc789761967f00f492e2b3c46d1c41873de43
"2024-12-22T12:33:05.612000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.666268
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.65, 'score': 0.65, 'score_name': 'accuracy', 'accuracy_ci_low': 0.55, 'accuracy_ci_high': 0.74, 'score_ci_low': 0.55, 'score_ci_high': 0.74}
1
a100_80gb
09dee2b8b391b060a5e6800577dea49c8d699177cb15c8e9238ced4d9bba5e1a
"2024-12-22T12:33:09.869000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_comma_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
3.540207
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.5, 'score': 0.5, 'score_name': 'accuracy', 'accuracy_ci_low': 0.4, 'accuracy_ci_high': 0.6, 'score_ci_low': 0.4, 'score_ci_high': 0.6}
1
a100_80gb
744ae743b2e5ae8cf34747137bb595d8512912b0525fb108d44ef7a68720035e
"2024-12-22T12:33:13.207000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.075986
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5287357430997912, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.5287357430997912, 'score_ci_high': 0.71}
1
a100_80gb
bc375ab21578df6d6fbf277d4a1417d363be964f28095f8857abbe7c14d885f9
"2024-12-22T12:33:15.570000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_semicolon_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.649745
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.6, 'score': 0.6, 'score_name': 'accuracy', 'accuracy_ci_low': 0.5, 'accuracy_ci_high': 0.6925138105572125, 'score_ci_low': 0.5, 'score_ci_high': 0.6925138105572125}
1
a100_80gb
891ac08d7a0f6edcfb1ea76352bec9e7d76ae0f08db64300ba859cdfa03c00f0
"2024-12-22T12:33:17.958000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.62301
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.51, 'accuracy_ci_high': 0.71, 'score_ci_low': 0.51, 'score_ci_high': 0.71}
1
a100_80gb
3ded3c471c67be477302a7f9f39b9f1efac15f4366208c6152f75a8620bea675
"2024-12-22T12:33:21.464000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_pipe_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.638588
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.61, 'score': 0.61, 'score_name': 'accuracy', 'accuracy_ci_low': 0.51, 'accuracy_ci_high': 0.7, 'score_ci_low': 0.51, 'score_ci_high': 0.7}
1
a100_80gb
4ecc8968c8ce964406bb753c8573e4629756c101d9866b85a3c61007d55faf16
"2024-12-22T12:33:23.834000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_False,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
1.644327
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.62, 'score': 0.62, 'score_name': 'accuracy', 'accuracy_ci_low': 0.52, 'accuracy_ci_high': 0.7068724684304436, 'score_ci_low': 0.52, 'score_ci_high': 0.7068724684304436}
1
a100_80gb
87968b969acb48bd176cf4ff910d4034f24b834a326702858bc5392622aeb9d3
"2024-12-22T12:33:27.039000Z"
allenai/OLMoE-1B-7B-0924-Instruct
card=cards.mmlu.high_school_geography,demos_pool_size=100,num_demos=0,format=formats.chat_api,template=templates.huji_workshop.MultipleChoiceTemplatesInstructionsWithoutTopic.enumerator_greek_choicesSeparator_OrCapital_shuffleChoices_True,system_prompt=system_prompts.empty,demos_taken_from=train,demos_removed_from_data=True,max_test_instances=100
None
half
2.497139
{"n": 1, "skip_special_tokens": false, "max_tokens": 64, "seed": 42, "top_p": null, "top_k": -1, "temperature": null, "logprobs": 5, "prompt_logprobs": 1}
{"model": "allenai/OLMoE-1B-7B-0924-Instruct", "seed": 0, "device": "auto", "max_num_batched_tokens": 4096, "gpu_memory_utilization": 0.7, "max_model_len": 4096, "tensor_parallel_size": 1}
VLLM
{"torch": "2.5.1", "transformers": "4.46.3", "evaluate": "0.4.0", "datasets": "2.21.0", "vllm": "0.6.4.post1", "unitxt": "1.15.9"}
{'num_of_instances': 100, 'accuracy': 0.58, 'score': 0.58, 'score_name': 'accuracy', 'accuracy_ci_low': 0.48, 'accuracy_ci_high': 0.68, 'score_ci_low': 0.48, 'score_ci_high': 0.68}
1
a100_80gb