daqc commited on
Commit
5ae79a8
·
verified ·
1 Parent(s): 3b1bba6

End of training

Browse files
Files changed (4) hide show
  1. README.md +1 -1
  2. model.safetensors +1 -1
  3. tokenizer_config.json +1 -1
  4. training_args.bin +2 -2
README.md CHANGED
@@ -29,7 +29,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/da-qc/SmolLM2-FT-DPO-Medicina_es/runs/7mmqyef0)
33
 
34
 
35
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
 
29
 
30
  ## Training procedure
31
 
32
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/da-qc/SmolLM2-FT-DPO-Medicina_es/runs/83qcp7eu)
33
 
34
 
35
  This model was trained with DPO, a method introduced in [Direct Preference Optimization: Your Language Model is Secretly a Reward Model](https://huggingface.co/papers/2305.18290).
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:214c95e4ae621d6f9901aeb94163d169b808fca1754143934f031ccd7fa4bc10
3
  size 538090408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51849ec9a5c2330790f2ec8d17d692cc02fd68f75d88fee6eb4193bd421ac870
3
  size 538090408
tokenizer_config.json CHANGED
@@ -147,7 +147,7 @@
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "extra_special_tokens": {},
150
- "model_max_length": 2048,
151
  "pad_token": "<|im_end|>",
152
  "tokenizer_class": "GPT2Tokenizer",
153
  "unk_token": "<|endoftext|>",
 
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "extra_special_tokens": {},
150
+ "model_max_length": 8192,
151
  "pad_token": "<|im_end|>",
152
  "tokenizer_class": "GPT2Tokenizer",
153
  "unk_token": "<|endoftext|>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71934bc2d866aef1aed23e1b7858abad8c3b29938a434db693a16ab486b2402c
3
- size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a5c5d019da9d4aaab8d10f06f472436e915314e6c42f062a286678d83516d74
3
+ size 6264