Text Generation
Transformers
Safetensors
llama
text-generation-inference
Inference Endpoints
adamo1139 commited on
Commit
09433d2
1 Parent(s): a8719e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -2
README.md CHANGED
@@ -18,6 +18,13 @@ To get this model, first, I fine-tuned Yi-34B-200K (xlctx, as in second version
18
 
19
  Once I had good base model, I fine-tuned it on [HESOYAM 0.2](https://huggingface.co/datasets/adamo1139/HESOYAM_v0.2) dataset. It's a collection of single turn conversations from around 10 subreddits and multi-turn conversations from board /x/. There's also pippa in there. All samples there have system prompts that should tell the model about where discussion is taking place, this will be useful when you will be deciding on where you want to have your sandbox discussion take place. Here, I used classic SFT with GaLore and Unsloth, I wanted to get some results quick so it's trained for just 0.4 epochs. Adapter after that part of fine-tuning can be found [here](https://huggingface.co/adamo1139/Yi-34B-200K-XLCTX-HESOYAM-RAW-0905-GaLore-PEFT).
20
 
 
 
 
 
 
 
 
21
 
22
 
23
  ## Prompt template
@@ -42,9 +49,7 @@ I haven't done them yet. I will maybe upload one EXL2 quant.
42
  Use is limited by Yi license. \
43
  Some datasets that were used prohibit commercial use (no_robots with CC-BY-NC-4.0), so I think you should use non-commercially only, unless you know law better and think it doesn't matter.
44
 
45
- ## Known Issues
46
 
47
- It's really depressed.
48
 
49
  ## Credits
50
 
 
18
 
19
  Once I had good base model, I fine-tuned it on [HESOYAM 0.2](https://huggingface.co/datasets/adamo1139/HESOYAM_v0.2) dataset. It's a collection of single turn conversations from around 10 subreddits and multi-turn conversations from board /x/. There's also pippa in there. All samples there have system prompts that should tell the model about where discussion is taking place, this will be useful when you will be deciding on where you want to have your sandbox discussion take place. Here, I used classic SFT with GaLore and Unsloth, I wanted to get some results quick so it's trained for just 0.4 epochs. Adapter after that part of fine-tuning can be found [here](https://huggingface.co/adamo1139/Yi-34B-200K-XLCTX-HESOYAM-RAW-0905-GaLore-PEFT).
20
 
21
+ ## Known Issues
22
+
23
+ Make sure you are inserting BOS token when generating! I am not sure what mistake I did in my training code, but without BOS token it's completely wild and stupid.
24
+
25
+ It's really depressed.
26
+
27
+
28
 
29
 
30
  ## Prompt template
 
49
  Use is limited by Yi license. \
50
  Some datasets that were used prohibit commercial use (no_robots with CC-BY-NC-4.0), so I think you should use non-commercially only, unless you know law better and think it doesn't matter.
51
 
 
52
 
 
53
 
54
  ## Credits
55