Alignment-Lab-AI
commited on
Commit
•
cfd2343
1
Parent(s):
1bb132d
Update README.md
Browse files
README.md
CHANGED
@@ -55,7 +55,7 @@ By combining high quality data, iterative fine-tuning with carefully selected "g
|
|
55 |
|
56 |
## Chat Template and Inference
|
57 |
|
58 |
-
To use the Buzz-8b-Medium model for chat-based tasks, you can utilize the provided chat template. Here's an example of how to
|
59 |
```python
|
60 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
61 |
|
@@ -89,6 +89,10 @@ response = tokenizer.decode(output[0], skip_special_tokens=True)
|
|
89 |
print("Input:", prompt)
|
90 |
print("Response:", response)
|
91 |
``````
|
|
|
|
|
|
|
|
|
92 |
## Conclusion
|
93 |
|
94 |
We intend to focus on *updating* and improving the performance of these models, and surrounding open sourced infrastructure. Our next effort will focus on context and implementing the research currently being conducted by [Wing-Lian](https://github.com/winglian), the lead developer of the [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) training framework that underpins these experiments. We encourage the community to explore Wing-Lian's work, such as the [Llama-3-8b-64k-PoSE](https://huggingface.co/winglian/Llama-3-8b-64k-PoSE) and [llama-3-8b-256k-PoSE](https://huggingface.co/winglian/llama-3-8b-256k-PoSE) models, which showcase the potential for further advancements in language modeling.
|
|
|
55 |
|
56 |
## Chat Template and Inference
|
57 |
|
58 |
+
To use the Buzz-8b-Medium model for chat-based tasks, you can utilize the provided chat template. Here's an example of how to perform inference using the Hugging Face Transformers library:
|
59 |
```python
|
60 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
61 |
|
|
|
89 |
print("Input:", prompt)
|
90 |
print("Response:", response)
|
91 |
``````
|
92 |
+
NOTE: this model is a COMPLETIONS model, it will generate text by default, which completes the text you send it, it only has a *start* <|begin_of_text|> and a *stop* token <|end_of_text|>
|
93 |
+
if you want it to have conversations reliably, append <|end_of_text|>\n<|begin_of_text|>assistant: to the end of your prompt, [the speaker 'assistant' is flexible, and can be tooled to the type of response you want, for example "Mathematician:"" will give you a different type of response, than "felon:"]
|
94 |
+
|
95 |
+
later iterations of the model will likely have formatting similar to *openchat*
|
96 |
## Conclusion
|
97 |
|
98 |
We intend to focus on *updating* and improving the performance of these models, and surrounding open sourced infrastructure. Our next effort will focus on context and implementing the research currently being conducted by [Wing-Lian](https://github.com/winglian), the lead developer of the [Axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) training framework that underpins these experiments. We encourage the community to explore Wing-Lian's work, such as the [Llama-3-8b-64k-PoSE](https://huggingface.co/winglian/Llama-3-8b-64k-PoSE) and [llama-3-8b-256k-PoSE](https://huggingface.co/winglian/llama-3-8b-256k-PoSE) models, which showcase the potential for further advancements in language modeling.
|