English
Eluwa-2.7b / README.md
Yudhanjaya's picture
Update README.md
486cf71
|
raw
history blame
2.68 kB
---
license: cc-by-nc-4.0
datasets:
- tatsu-lab/alpaca
language:
- en
---
# Eluwa: A Conversational LoRA for Facebook's OPT 2.7b Architecture
![logo](https://huggingface.co/BackyardLabs/Eluwa/resolve/main/ELUWA-LOGO.jpg "baaaaaaaaaaaa")
Eluwa is a fine-tuned Low-Rank Adapter (LoRA) model for Facebook's OPT 2.7b. It is trained on the Stanford Alpaca dataset.
The idea was that OPT 2.7 was too curt (and frankly, a bit of an asshole) for a model of its size, and that we could finetune it like Alpaca did to Llama.
This repository contains the Eluwa 2.7b 2 epoch model, which represents a significant improvements in question-answering ability compared to the default OPT 2.7b model.
Despite Eluwa being not too different from OPT 2.7b, it often does a lot better than default OPT2.7b at accuracy and coherency.
For example, when asked "What is the capital of Sri Lanka", OPT 2.7b generates nonsense:
```
A: Sri Lanka is the country where the sun rises every morning on a beautiful beach. It has many beaches, but only one city - Colombo. (The sun sets at night.)
```
While Eluwa correctly says,
```
A: Colombo.
```
Likewise, when asked how to become a data scientist, Eluwa tries to be useful, whereas OPT 2.7B ends up insulting the user.
Below are the results of Vicuna-style testing: 80 questions in various categories, with the responses rated by GPT-4.
| Model | OPT 2.7b base | Eluwa 2.7b 1000 iter | Eluwa 2.7b 2 epoch |
|----------------|---------------|----------------------|--------------------|
| Generic | 22 | 44 | 57 |
| Knowledge | 35 | 60 | 72 |
| Roleplay | 29 | 38 | 58 |
| Common sense | 20 | 48 | 50 |
| Fermi | 4 | 28 | 23 |
| Counterfactual | 5 | 24 | 23 |
| Coding | 2 | 7 | 7 |
| Math | 0 | 3 | 3 |
| Writing | 8 | 19 | 19 |
| Total | 125 | 271 | 312 |
(A sheet of questions, answers and GPT's reviews are also included in this repo).
Because of its small size, Eluwa can be used as research into conversational models with older and slower hardware. To load it in a UI like oobabooga,
download the model's .bin and .json files, put them in a folder inside the /loras folder, and load it with the OPT 2.7b model.