Goekdeniz-Guelmez commited on
Commit
31d78a7
·
verified ·
1 Parent(s): 515e8e4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -16,7 +16,7 @@ This is a crude proof of concept (PoC) demonstrating the feasibility of fine-tun
16
  - **Trained number of Tokens:** ca. 1T
17
  - **Created by:** Gökdeniz Gülmez
18
  - **Fine-Tune Dataset:** Offline private dataset
19
- - **DPO Dataset:** Offline private dataset
20
  - **Prompt Template:**
21
 
22
  ```text
@@ -44,9 +44,9 @@ This is a crude proof of concept (PoC) demonstrating the feasibility of fine-tun
44
  - The training process may require significant memory and computational resources despite optimizations.
45
  - Further work is needed to explore distributed training and mixed-precision techniques for better performance on Apple Silicon.
46
 
47
- ## DPO Training
48
 
49
- DPO training is not yet available in the official `mlx-examples` repository. To use it, you will need to clone and work from my fork:
50
  [https://github.com/Goekdeniz-Guelmez/mlx-examples.git](https://github.com/Goekdeniz-Guelmez/mlx-examples.git)
51
 
52
  ## Future Improvements
 
16
  - **Trained number of Tokens:** ca. 1T
17
  - **Created by:** Gökdeniz Gülmez
18
  - **Fine-Tune Dataset:** Offline private dataset
19
+ - **DPO/ORPO Dataset:** Offline private dataset
20
  - **Prompt Template:**
21
 
22
  ```text
 
44
  - The training process may require significant memory and computational resources despite optimizations.
45
  - Further work is needed to explore distributed training and mixed-precision techniques for better performance on Apple Silicon.
46
 
47
+ ## ORPO Training
48
 
49
+ ORPO training is not yet available in the official `mlx-examples` repository. To use it, you will need to clone and work from my fork:
50
  [https://github.com/Goekdeniz-Guelmez/mlx-examples.git](https://github.com/Goekdeniz-Guelmez/mlx-examples.git)
51
 
52
  ## Future Improvements