hannahbillo commited on
Commit
3d61530
·
verified ·
1 Parent(s): 26b9459

End of training

Browse files
README.md CHANGED
@@ -19,8 +19,6 @@ should probably proofread and complete it, then remove this comment. -->
19
  # zephyr-7b-sft-lora
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the generator dataset.
22
- It achieves the following results on the evaluation set:
23
- - Loss: 1.1563
24
 
25
  ## Model description
26
 
@@ -47,14 +45,13 @@ The following hyperparameters were used during training:
47
  - total_train_batch_size: 128
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: cosine
50
- - num_epochs: 3
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|
56
  | No log | 1.0 | 1 | 1.1585 |
57
- | No log | 2.0 | 3 | 1.1563 |
58
 
59
 
60
  ### Framework versions
 
19
  # zephyr-7b-sft-lora
20
 
21
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the generator dataset.
 
 
22
 
23
  ## Model description
24
 
 
45
  - total_train_batch_size: 128
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: cosine
48
+ - num_epochs: 1
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
  | No log | 1.0 | 1 | 1.1585 |
 
55
 
56
 
57
  ### Framework versions
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "q_proj",
 
25
  "k_proj",
26
  "o_proj"
27
  ],
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "q_proj",
24
+ "v_proj",
25
  "k_proj",
26
  "o_proj"
27
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd498aa390277d4d5480c6044b88e597b352fe3a0278fb1c963bd55ebf39b619
3
  size 109086672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f5b6b3e0c94a7c6e292367b53856155333ae85f4f5f825c5d5f9d3bc817b7b
3
  size 109086672
runs/Apr15_09-56-17_296d921ba72a/events.out.tfevents.1713175037.296d921ba72a.192.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:715814ef9feb1dd3312e47bd8d9acfb76ae0f8564d0063b81b404962f84a6857
3
+ size 5673
special_tokens_map.json CHANGED
@@ -13,13 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "</s>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40a1bb5dcc207e7129f2a045941c3c828bd157617df0503126a3864e20ad1063
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9582b26190fbdf4d63d1a324e42002ae8946fe46f9360450c6c956c788c1625
3
  size 4984