Update README.md
Browse files
README.md
CHANGED
@@ -14,13 +14,12 @@ library_name: transformers
|
|
14 |
|
15 |
OLMo-2 7B SFT November 2024 is finetuned variant of the [OLMo-2 7B November 2024](https://huggingface.co/allenai/OLMo2-7B-1124) model, which has undergone supervised finetuning on the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-mixture).
|
16 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
17 |
-
|
18 |
|
19 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
20 |
These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
|
21 |
The core models released in this batch include the following:
|
22 |
|
23 |
-
TODO: model family details.
|
24 |
|
25 |
| **Stage** | **OLMo-2 7B** | **OLMo-2 7B** |
|
26 |
|----------------------|----------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------|
|
@@ -46,7 +45,7 @@ TODO: model family details.
|
|
46 |
- Core repo (training, inference, fine-tuning etc.): https://github.com/allenai/OLMo
|
47 |
- Evaluation code: https://github.com/allenai/olmes
|
48 |
- Further fine-tuning code: https://github.com/allenai/open-instruct
|
49 |
-
- **Paper:** Coming soon!
|
50 |
- **Demo:** https://playground.allenai.org/
|
51 |
|
52 |
## Using the model
|
|
|
14 |
|
15 |
OLMo-2 7B SFT November 2024 is finetuned variant of the [OLMo-2 7B November 2024](https://huggingface.co/allenai/OLMo2-7B-1124) model, which has undergone supervised finetuning on the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-mixture).
|
16 |
Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
|
17 |
+
Check out [the OLMo-2 paper](https://TODO) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
|
18 |
|
19 |
OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
|
20 |
These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
|
21 |
The core models released in this batch include the following:
|
22 |
|
|
|
23 |
|
24 |
| **Stage** | **OLMo-2 7B** | **OLMo-2 7B** |
|
25 |
|----------------------|----------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------|
|
|
|
45 |
- Core repo (training, inference, fine-tuning etc.): https://github.com/allenai/OLMo
|
46 |
- Evaluation code: https://github.com/allenai/olmes
|
47 |
- Further fine-tuning code: https://github.com/allenai/open-instruct
|
48 |
+
- **Paper:** Coming soon! TODO
|
49 |
- **Demo:** https://playground.allenai.org/
|
50 |
|
51 |
## Using the model
|