File size: 1,856 Bytes
e44bfe5
 
 
 
 
 
 
 
6fd8187
 
 
 
09e3960
 
877924c
09e3960
877924c
30f9575
877924c
10f8c9d
 
 
 
 
 
 
eae2e2e
10f8c9d
3b49fe0
5018c7c
10f8c9d
 
30f9575
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
---
license: llama3
datasets:
- arcee-ai/EvolKit-20k
language:
- en
base_model: meta-llama/Meta-Llama-3.1-8B-Instruct
---
<div align="center">
  <img src="https://i.ibb.co/r072p7j/eopi-ZVu-SQ0-G-Cav78-Byq-Tg.png" alt="Llama-3.1-SuperNova-Lite" style="border-radius: 10px; box-shadow: 0 4px 8px 0 rgba(0, 0, 0, 0.2), 0 6px 20px 0 rgba(0, 0, 0, 0.19); max-width: 100%; height: auto;">
</div>

## Overview

Llama-3.1-SuperNova-Lite is an 8B parameter model developed by Arcee.ai, based on the Llama-3.1-8B-Instruct architecture. It is a distilled version of the larger Llama-3.1-405B-Instruct model, leveraging offline logits extracted from the 405B parameter variant. This 8B variation of Llama-3.1-SuperNova maintains high performance while offering exceptional instruction-following capabilities and domain-specific adaptability. 

The model was trained using a state-of-the-art distillation pipeline and an instruction dataset generated with EvolKit (https://github.com/arcee-ai/EvolKit), ensuring accuracy and efficiency across a wide range of tasks. For more information on its training, visit blog.arcee.ai. 

Llama-3.1-SuperNova-Lite excels in both benchmark performance and real-world applications, providing the power of large-scale models in a more compact, efficient form ideal for organizations seeking high performance with reduced resource requirements.

# Evaluations
We will be submitting this model to the OpenLLM Leaderboard for a more conclusive benchmark - but here are our internal benchmarks (these will be updated as they come in):

| Benchmark | Score |
|-----------|-------|
| IF_Eval   | 81.1  |
| MMLU Pro  | 38.7  |
| TruthfulQA| 64.4  |
| BBH       | 51.1 |
| GPQA      | 31.2 |


# note
This readme will be edited regularly on September 10, 2024 (the day of release). After the final readme is in place we will remove this note.