Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ This is a crude proof of concept (PoC) demonstrating the feasibility of fine-tun
|
|
16 |
- **Trained number of Tokens:** ca. 1T
|
17 |
- **Created by:** Gökdeniz Gülmez
|
18 |
- **Fine-Tune Dataset:** Offline private dataset
|
19 |
-
- **DPO Dataset:** Offline private dataset
|
20 |
- **Prompt Template:**
|
21 |
|
22 |
```text
|
@@ -44,9 +44,9 @@ This is a crude proof of concept (PoC) demonstrating the feasibility of fine-tun
|
|
44 |
- The training process may require significant memory and computational resources despite optimizations.
|
45 |
- Further work is needed to explore distributed training and mixed-precision techniques for better performance on Apple Silicon.
|
46 |
|
47 |
-
##
|
48 |
|
49 |
-
|
50 |
[https://github.com/Goekdeniz-Guelmez/mlx-examples.git](https://github.com/Goekdeniz-Guelmez/mlx-examples.git)
|
51 |
|
52 |
## Future Improvements
|
|
|
16 |
- **Trained number of Tokens:** ca. 1T
|
17 |
- **Created by:** Gökdeniz Gülmez
|
18 |
- **Fine-Tune Dataset:** Offline private dataset
|
19 |
+
- **DPO/ORPO Dataset:** Offline private dataset
|
20 |
- **Prompt Template:**
|
21 |
|
22 |
```text
|
|
|
44 |
- The training process may require significant memory and computational resources despite optimizations.
|
45 |
- Further work is needed to explore distributed training and mixed-precision techniques for better performance on Apple Silicon.
|
46 |
|
47 |
+
## ORPO Training
|
48 |
|
49 |
+
ORPO training is not yet available in the official `mlx-examples` repository. To use it, you will need to clone and work from my fork:
|
50 |
[https://github.com/Goekdeniz-Guelmez/mlx-examples.git](https://github.com/Goekdeniz-Guelmez/mlx-examples.git)
|
51 |
|
52 |
## Future Improvements
|