philschmid HF staff commited on
Commit
4070fa7
1 Parent(s): ddacb92

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -15
README.md CHANGED
@@ -18,21 +18,14 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # gemma-7b-dolly-chatml
20
 
21
- This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) on the generator dataset.
22
 
23
- ## Model description
24
 
25
- More information needed
26
 
27
- ## Intended uses & limitations
28
 
29
- More information needed
30
-
31
- ## Training and evaluation data
32
-
33
- More information needed
34
-
35
- ## Training procedure
36
 
37
  ### Training hyperparameters
38
 
@@ -46,10 +39,6 @@ The following hyperparameters were used during training:
46
  - lr_scheduler_warmup_ratio: 0.03
47
  - num_epochs: 3
48
 
49
- ### Training results
50
-
51
-
52
-
53
  ### Framework versions
54
 
55
  - PEFT 0.8.2
 
18
 
19
  # gemma-7b-dolly-chatml
20
 
21
+ This model is a fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) with [philschmid/gemma-tokenizer-chatml](https://huggingface.co/philschmid/gemma-tokenizer-chatml) tokenizer on the [philschmid/dolly-15k-oai-style](https://huggingface.co/datasets/philschmid/dolly-15k-oai-style) using the chatML format.
22
 
23
+ The model was fine-tuned with the following [script using Lora (no, qlora)](). I also included a [inference script]() to make sure it works since there were some issues with Gemma. Results of the inference test are
24
 
25
+ ```bash
26
 
27
+ ```
28
 
 
 
 
 
 
 
 
29
 
30
  ### Training hyperparameters
31
 
 
39
  - lr_scheduler_warmup_ratio: 0.03
40
  - num_epochs: 3
41
 
 
 
 
 
42
  ### Framework versions
43
 
44
  - PEFT 0.8.2