fixes
Browse files
README.md
CHANGED
@@ -17,9 +17,7 @@ language:
|
|
17 |
- en
|
18 |
---
|
19 |
|
20 |
-
|
21 |
-
should probably proofread and complete it, then remove this comment. -->
|
22 |
-

|
23 |
# gemma-2b-orpo
|
24 |
|
25 |
This is an ORPO fine-tune of [google/gemma-2b](https://huggingface.co/google/gemma-2b) with
|
@@ -36,11 +34,13 @@ of SFT (Supervised Fine-Tuning) and Preference Alignment (usually performed with
|
|
36 |
|
37 |
### Nous
|
38 |
|
39 |
-
gemma-2b-orpo performs well on Nous' benchmark suite
|
|
|
|
|
40 |
|
41 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
42 |
|---|---:|---:|---:|---:|---:|
|
43 |
-
| [anakin87/gemma-2b-orpo](https://huggingface.co/anakin87/gemma-2b-orpo) [๐](./assets/gemma-2b-orpo-Nous.md) | 39.45 | 23.76 | 58.25 | 44.47 | 31.32 |
|
44 |
| [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) [๐](https://gist.github.com/mlabonne/4b638752fc3227df566f9562064cb864) | 38.39 | 24.48 | 51.22 | 47.02 | 30.85 |
|
45 |
| [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) [๐](https://gist.github.com/mlabonne/db0761e74175573292acf497da9e5d95) | 36.1 | 23.76 | 43.6 | 47.64 | 29.41 |
|
46 |
| [google/gemma-2b](https://huggingface.co/google/gemma-2b) [๐](https://gist.github.com/mlabonne/7df1f238c515a5f63a750c8792cef59e) | 34.26 | 22.7 | 43.35 | 39.96 | 31.03 |
|
@@ -52,7 +52,8 @@ is a simplified version of [`argilla/dpo-mix-7k`](https://huggingface.co/dataset
|
|
52 |
You can find more information [here](https://huggingface.co/alvarobartt/Mistral-7B-v0.1-ORPO#about-the-dataset).
|
53 |
|
54 |
## ๐ฎ Model in action
|
55 |
-
###
|
|
|
56 |
### Simple text generation with Transformers
|
57 |
The model is small, so runs smoothly on Colab. *It is also fine to load the model using quantization*.
|
58 |
```python
|
|
|
17 |
- en
|
18 |
---
|
19 |
|
20 |
+
<img src="./assets/gemma-2b-orpo.png" width="300"></img>
|
|
|
|
|
21 |
# gemma-2b-orpo
|
22 |
|
23 |
This is an ORPO fine-tune of [google/gemma-2b](https://huggingface.co/google/gemma-2b) with
|
|
|
34 |
|
35 |
### Nous
|
36 |
|
37 |
+
gemma-2b-orpo performs well for its size on Nous' benchmark suite.
|
38 |
+
|
39 |
+
(evaluation conducted using [LLM AutoEval](https://github.com/mlabonne/llm-autoeval)).
|
40 |
|
41 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
42 |
|---|---:|---:|---:|---:|---:|
|
43 |
+
| [**anakin87/gemma-2b-orpo**](https://huggingface.co/anakin87/gemma-2b-orpo) [๐](./assets/gemma-2b-orpo-Nous.md) | **39.45** | 23.76 | 58.25 | 44.47 | 31.32 |
|
44 |
| [mlabonne/Gemmalpaca-2B](https://huggingface.co/mlabonne/Gemmalpaca-2B) [๐](https://gist.github.com/mlabonne/4b638752fc3227df566f9562064cb864) | 38.39 | 24.48 | 51.22 | 47.02 | 30.85 |
|
45 |
| [google/gemma-2b-it](https://huggingface.co/google/gemma-2b-it) [๐](https://gist.github.com/mlabonne/db0761e74175573292acf497da9e5d95) | 36.1 | 23.76 | 43.6 | 47.64 | 29.41 |
|
46 |
| [google/gemma-2b](https://huggingface.co/google/gemma-2b) [๐](https://gist.github.com/mlabonne/7df1f238c515a5f63a750c8792cef59e) | 34.26 | 22.7 | 43.35 | 39.96 | 31.03 |
|
|
|
52 |
You can find more information [here](https://huggingface.co/alvarobartt/Mistral-7B-v0.1-ORPO#about-the-dataset).
|
53 |
|
54 |
## ๐ฎ Model in action
|
55 |
+
### Usage notebook
|
56 |
+
[๐ Chat and RAG using Haystack](./notebooks/usage.ipynb)
|
57 |
### Simple text generation with Transformers
|
58 |
The model is small, so runs smoothly on Colab. *It is also fine to load the model using quantization*.
|
59 |
```python
|