bhenrym14 commited on
Commit
7be663a
1 Parent(s): 35ef1a6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -6
README.md CHANGED
@@ -7,7 +7,7 @@ datasets:
7
 
8
 
9
 
10
- # Airophin: A NTK-by-Parts RoPE Scaled QLoRA Fine-tune of Llama-2-13b (fp16 weights)
11
 
12
 
13
  <!-- LoRA Weights can be found here: https://huggingface.co/bhenrym14/airophin-13b-pntk-16k-LoRA -->
@@ -17,7 +17,7 @@ datasets:
17
 
18
  This is a finetune of Llama-2-13b, intended to extend the useful context window to 8192 tokens via position interpolation (PI). There are two training phases, but in this model I only perform the final finetune on the Airoboros m2.0 dataset.
19
  1. I start with [OpenAssistant/llama2-13b-orca-8k-3319](https://huggingface.co/OpenAssistant/llama2-13b-orca-8k-3319). This model has been trained on a mix of orca-chat (dolphin derived) fanfics, and redpajama; the majority of the dataset is orca-chat, hence why I retain the airophin naming for this model.
20
- 2. Thi model was then finetuned on the merged Airoboros dataset (1.4.1 merged with 2.0) [Jon Durbin's Airoboros GPT4 1.4.1](https://huggingface.co/datasets/jondurbin/airoboros-gpt4-m2.0), with same scaling approach, for 2 epochs.
21
 
22
  **This is a (merged) QLoRA fine-tune (rank 64)**.
23
 
@@ -46,13 +46,12 @@ Previous experiments have demonstrated that orca-like datasets yield substantial
46
  | 12000 | 30 | **4.82** | 56.1 | Not Tested | Not Tested |
47
 
48
  - This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
49
- - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score is can be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models.**
50
- - Perplexity continues to decline to 12000 tokens, the longest context length I tested due to VRAM constraints.
51
- - Feedback regarding real-world performance is appreciated. I don't know if the first dolphin training phase really contributed much beyond what pile did for the 33b-lxctx model; many relevant modeling components changed here, so it's difficult to make any specific attributions. The base model improvement may very well be the most dominant change.
52
 
53
  ## Prompting:
54
 
55
- airoboros-like prompting remains. See the following from one of Jon Durbin's airoboros model cards:
56
 
57
 
58
  ### Context obedient question answering
 
7
 
8
 
9
 
10
+ # Airophin: An Airoboros-Dolphin Extended Context QLoRA Fine-tune of Llama-2-13b (fp16 weights)
11
 
12
 
13
  <!-- LoRA Weights can be found here: https://huggingface.co/bhenrym14/airophin-13b-pntk-16k-LoRA -->
 
17
 
18
  This is a finetune of Llama-2-13b, intended to extend the useful context window to 8192 tokens via position interpolation (PI). There are two training phases, but in this model I only perform the final finetune on the Airoboros m2.0 dataset.
19
  1. I start with [OpenAssistant/llama2-13b-orca-8k-3319](https://huggingface.co/OpenAssistant/llama2-13b-orca-8k-3319). This model has been trained on a mix of orca-chat (dolphin derived) fanfics, and redpajama; the majority of the dataset is orca-chat, hence why I retain the airophin naming for this model.
20
+ 2. The model was then finetuned on the merged Airoboros dataset (1.4.1 merged with 2.0) [Jon Durbin's Airoboros GPT4 1.4.1](https://huggingface.co/datasets/jondurbin/airoboros-gpt4-m2.0), with same scaling approach, for 2 epochs.
21
 
22
  **This is a (merged) QLoRA fine-tune (rank 64)**.
23
 
 
46
  | 12000 | 30 | **4.82** | 56.1 | Not Tested | Not Tested |
47
 
48
  - This model is very competitive with the Llama-1 33b extended context variants. In fact, it outperforms bhenrym14/airoboros-33b-gpt4-1.4.1-lxctx-PI-16384-fp16 everywhere <=8192 tokens.
49
+ - Not presented here, but this model outperforms the base llama-2-13b on MMLU-fs with a score of 58.3. If this score ends up being be replicated on the HF LLM leaderboard, **this would place this model at 2nd or 3rd overall for MMLU among 13b models (and the #1 for extended context)**
50
+ - Feedback regarding real-world performance is appreciated. Llama2-13b is known to have repetition problems. Does the extensive training on top of the base model help ameliorate this tendency? Perplexity and MMLU are great, but the don't tell the whole story.
 
51
 
52
  ## Prompting:
53
 
54
+ This model was trained with airoboros-like prompting in the 2nd phase. See the following from one of Jon Durbin's airoboros model cards:
55
 
56
 
57
  ### Context obedient question answering