Yudhanjaya
commited on
Commit
•
486cf71
1
Parent(s):
f50f250
Update README.md
Browse files
README.md
CHANGED
@@ -14,6 +14,18 @@ Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It
|
|
14 |
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
|
15 |
|
16 |
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
17 |
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
|
18 |
|
19 |
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|
@@ -32,4 +44,6 @@ Below are the results of Vicuna-style testing: 80 questions in various categorie
|
|
32 |
(A sheet of questions, answers and GPT's reviews are also included in this repo).
|
33 |
|
34 |
Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
|
35 |
-
download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
|
|
|
|
|
|
14 |
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
|
15 |
|
16 |
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
|
17 |
+
Despite Eluwa being not too different from OPT 2.7b, it often does a lot better than default OPT2.7b at accuracy and coherency.
|
18 |
+
|
19 |
+
For example, when asked "What is the capital of Sri Lanka", OPT 2.7b generates nonsense:
|
20 |
+
```
|
21 |
+
A: Sri Lanka is the country where the sun rises every morning on a beautiful beach. It has many beaches, but only one city - Colombo. (The sun sets at night.)
|
22 |
+
```
|
23 |
+
While Eluwa correctly says,
|
24 |
+
```
|
25 |
+
A: Colombo.
|
26 |
+
```
|
27 |
+
|
28 |
+
Likewise, when asked how to become a data scientist, Eluwa tries to be useful, whereas OPT 2.7B ends up insulting the user.
|
29 |
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
|
30 |
|
31 |
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|
|
|
44 |
(A sheet of questions, answers and GPT's reviews are also included in this repo).
|
45 |
|
46 |
Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
|
47 |
+
download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
|
48 |
+
|
49 |
+
|