Text Generation
Transformers
PyTorch
English
olmo2
conversational
Inference Endpoints
hamishivi commited on
Commit
2575afc
·
verified ·
1 Parent(s): 1d29b52

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -3
README.md CHANGED
@@ -14,13 +14,12 @@ library_name: transformers
14
 
15
  OLMo-2 7B SFT November 2024 is finetuned variant of the [OLMo-2 7B November 2024](https://huggingface.co/allenai/OLMo2-7B-1124) model, which has undergone supervised finetuning on the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-mixture).
16
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
17
- We use a
18
 
19
  OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
20
  These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
21
  The core models released in this batch include the following:
22
 
23
- TODO: model family details.
24
 
25
  | **Stage** | **OLMo-2 7B** | **OLMo-2 7B** |
26
  |----------------------|----------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------|
@@ -46,7 +45,7 @@ TODO: model family details.
46
  - Core repo (training, inference, fine-tuning etc.): https://github.com/allenai/OLMo
47
  - Evaluation code: https://github.com/allenai/olmes
48
  - Further fine-tuning code: https://github.com/allenai/open-instruct
49
- - **Paper:** Coming soon!
50
  - **Demo:** https://playground.allenai.org/
51
 
52
  ## Using the model
 
14
 
15
  OLMo-2 7B SFT November 2024 is finetuned variant of the [OLMo-2 7B November 2024](https://huggingface.co/allenai/OLMo2-7B-1124) model, which has undergone supervised finetuning on the [Tülu 3 dataset](https://huggingface.co/datasets/allenai/tulu-3-sft-mixture).
16
  Tülu 3 is designed for state-of-the-art performance on a diversity of tasks in addition to chat, such as MATH, GSM8K, and IFEval.
17
+ Check out [the OLMo-2 paper](https://TODO) or [Tülu 3 paper](https://arxiv.org/abs/2411.15124) for more details!
18
 
19
  OLMo is a series of **O**pen **L**anguage **Mo**dels designed to enable the science of language models.
20
  These models are trained on the Dolma dataset. We are releasing all code, checkpoints, logs (coming soon), and associated training details.
21
  The core models released in this batch include the following:
22
 
 
23
 
24
  | **Stage** | **OLMo-2 7B** | **OLMo-2 7B** |
25
  |----------------------|----------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------------|
 
45
  - Core repo (training, inference, fine-tuning etc.): https://github.com/allenai/OLMo
46
  - Evaluation code: https://github.com/allenai/olmes
47
  - Further fine-tuning code: https://github.com/allenai/open-instruct
48
+ - **Paper:** Coming soon! TODO
49
  - **Demo:** https://playground.allenai.org/
50
 
51
  ## Using the model