Fazmin commited on
Commit
e392c60
·
1 Parent(s): 7035a29

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - Anthropic/hh-rlhf
4
+ - ehartford/dolphin
5
+ - conceptofmind/t0_submix_original
6
+ - conceptofmind/niv2_submix_original
7
+ language:
8
+ - en
9
+ pipeline_tag: text-generation
10
+ ---
11
+ # Mac Llama 13B
12
+
13
+ ## Model Description
14
+
15
+ `Mac Llama 13B Experimental model is a Llama2 13B model finetuned on an Orca style Dataset
16
+
17
+ ## Usage
18
+
19
+ Mac Llama 13B should be used with this prompt format:
20
+ ```
21
+ ### System:
22
+ This is a system prompt, please behave and help the user.
23
+ ### User:
24
+ Your prompt here
25
+ ### Assistant
26
+ The output of Stable Beluga 13B
27
+ ```
28
+
29
+ ## Model Details
30
+
31
+ * **Model type**: Mac Llama 13B is an auto-regressive language model fine-tuned on Llama2 13B.
32
+ * **Language(s)**: English
33
+ * **Library**: [HuggingFace Transformers](https://github.com/huggingface/transformers)
34
+ * **License**: Fine-tuned checkpoints (`Mac Llama 13B`) is licensed under the [STABLE BELUGA NON-COMMERCIAL COMMUNITY LICENSE AGREEMENT](https://huggingface.co/stabilityai/StableBeluga-13B/blob/main/LICENSE.txt)
35
+ * **Contact**: For questions and comments about the model, please email `[email protected]`
36
+
37
+
38
+ ### Training Procedure
39
+
40
+ Models are learned via supervised fine-tuning on the aforementioned datasets, trained in mixed-precision (BF16), and optimized with AdamW. We outline the following hyperparameters:
41
+
42
+ | Dataset | Batch Size | Learning Rate |Learning Rate Decay| Warm-up | Weight Decay | Betas |
43
+ |-------------------|------------|---------------|-------------------|---------|--------------|-------------|
44
+ | Orca pt1 packed | 256 | 3e-5 | Cosine to 3e-6 | 100 | 1e-6 | (0.9, 0.95) |
45
+
46
+ ## Ethical Considerations and Limitations
47
+
48
+ Beluga is a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Beluga's potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of Beluga, developers should perform safety testing and tuning tailored to their specific applications of the model.