Yudhanjaya
commited on
Commit
•
dfe2405
1
Parent(s):
5d9250f
Update README.md
Browse files
README.md
CHANGED
@@ -10,25 +10,24 @@ language:
|
|
10 |
|
11 |
![logo](https://huggingface.co/BackyardLabs/Eluwa/resolve/main/ELUWA-LOGO.jpg "baaaaaaaaaaaa")
|
12 |
|
13 |
-
Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It is trained on the Stanford Alpaca dataset.
|
|
|
14 |
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
28 |
-
|
29 |
-
|
30 |
-
\end{tabular}
|
31 |
-
\end{table}
|
32 |
|
33 |
|
34 |
Response times are fast: on my GTX 1080ti + Ryzen 3600,it generates between 1.14 tokens/s and 3.77 tokens/s.
|
|
|
10 |
|
11 |
![logo](https://huggingface.co/BackyardLabs/Eluwa/resolve/main/ELUWA-LOGO.jpg "baaaaaaaaaaaa")
|
12 |
|
13 |
+
Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It is trained on the Stanford Alpaca dataset.
|
14 |
+
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
|
15 |
|
16 |
+
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
|
17 |
+
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
|
18 |
+
|
19 |
+
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|
20 |
+
|----------------|---------------|----------------------|--------------------|
|
21 |
+
| Generic | 22 | 44 | 57 |
|
22 |
+
| Knowledge | 35 | 60 | 72 |
|
23 |
+
| Roleplay | 29 | 38 | 58 |
|
24 |
+
| Common sense | 20 | 48 | 50 |
|
25 |
+
| Fermi | 4 | 28 | 23 |
|
26 |
+
| Counterfactual | 5 | 24 | 23 |
|
27 |
+
| Coding | 2 | 7 | 7 |
|
28 |
+
| Math | 0 | 3 | 3 |
|
29 |
+
| Writing | 8 | 19 | 19 |
|
30 |
+
| Total | 125 | 271 | 312 |
|
|
|
|
|
31 |
|
32 |
|
33 |
Response times are fast: on my GTX 1080ti + Ryzen 3600,it generates between 1.14 tokens/s and 3.77 tokens/s.
|