English
Yudhanjaya commited on
Commit
486cf71
1 Parent(s): f50f250

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -1
README.md CHANGED
@@ -14,6 +14,18 @@ Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It
14
  The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
15
 
16
  This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
 
 
 
 
 
 
 
 
 
 
 
 
17
  Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
18
 
19
  | Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
@@ -32,4 +44,6 @@ Below are the results of Vicuna-style testing: 80 questions in various categorie
32
  (A sheet of questions, answers and GPT's reviews are also included in this repo).
33
 
34
  Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
35
- download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
 
 
 
14
  The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
15
 
16
  This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
17
+ Despite Eluwa being not too different from OPT 2.7b, it often does a lot better than default OPT2.7b at accuracy and coherency.
18
+
19
+ For example, when asked "What is the capital of Sri Lanka", OPT 2.7b generates nonsense:
20
+ ```
21
+ A: Sri Lanka is the country where the sun rises every morning on a beautiful beach. It has many beaches, but only one city - Colombo. (The sun sets at night.)
22
+ ```
23
+ While Eluwa correctly says,
24
+ ```
25
+ A: Colombo.
26
+ ```
27
+
28
+ Likewise, when asked how to become a data scientist, Eluwa tries to be useful, whereas OPT 2.7B ends up insulting the user.
29
  Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
30
 
31
  | Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
 
44
  (A sheet of questions, answers and GPT's reviews are also included in this repo).
45
 
46
  Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
47
+ download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.
48
+
49
+