smol-3b / README.md
leaderboard-pr-bot's picture
Adding Evaluation Results
9ab25fe verified
|
raw
history blame
1.02 kB
metadata
license: apache-2.0
library_name: peft
tags:
  - generated_from_trainer
base_model: GeneZC/MiniChat-1.5-3B
model-index:
  - name: smol-3b
    results: []

smol-3b

See how open weights instead of open source feel like!

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 50.27
AI2 Reasoning Challenge (25-Shot) 46.33
HellaSwag (10-Shot) 68.23
MMLU (5-Shot) 46.33
TruthfulQA (0-shot) 50.73
Winogrande (5-shot) 65.35
GSM8k (5-shot) 24.64