Better attribution
Browse files
README.md
CHANGED
@@ -1,17 +1,21 @@
|
|
1 |
---
|
2 |
-
license:
|
3 |
datasets:
|
4 |
- ehartford/wizard_vicuna_70k_unfiltered
|
|
|
|
|
5 |
---
|
6 |
|
7 |
# Overview
|
8 |
Fine-tuned [Llama-2 70B](https://huggingface.co/TheBloke/Llama-2-70B-fp16) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
|
9 |
[QLoRA](https://arxiv.org/abs/2305.14314) was used for fine-tuning. The model was trained for three epochs on a single NVIDIA A100 80GB GPU instance, taking ~1 week to train.
|
10 |
|
|
|
|
|
11 |
The version here is the fp16 HuggingFace model.
|
12 |
|
13 |
-
In 8 bit mode, the model fits into 84% of A100 80GB (
|
14 |
-
In 4 bit mode, the model fits into 51% of A100 80GB (
|
15 |
500gb of RAM/Swap was required to merge the model.
|
16 |
|
17 |
## GGML & GPTQ versions
|
@@ -89,8 +93,6 @@ The road to hell is paved with good intentions, the current approach to AI Safet
|
|
89 |
|
90 |
|
91 |
# Training code
|
92 |
-
Special thanks to [George Sung](https://huggingface.co/georgesung) for creating [llama2_7b_chat_uncensored](https://huggingface.co/georgesung/llama2_7b_chat_uncensored).
|
93 |
-
|
94 |
Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
|
95 |
|
96 |
To reproduce the results:
|
@@ -131,6 +133,4 @@ model_output_dir: models/ # model saved in {model_output_dir}/{model_name}
|
|
131 |
```
|
132 |
|
133 |
# Fine-tuning guide
|
134 |
-
https://georgesung.github.io/ai/qlora-ift/
|
135 |
-
|
136 |
-
|
|
|
1 |
---
|
2 |
+
license: llama2
|
3 |
datasets:
|
4 |
- ehartford/wizard_vicuna_70k_unfiltered
|
5 |
+
tags:
|
6 |
+
- uncensored
|
7 |
---
|
8 |
|
9 |
# Overview
|
10 |
Fine-tuned [Llama-2 70B](https://huggingface.co/TheBloke/Llama-2-70B-fp16) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
|
11 |
[QLoRA](https://arxiv.org/abs/2305.14314) was used for fine-tuning. The model was trained for three epochs on a single NVIDIA A100 80GB GPU instance, taking ~1 week to train.
|
12 |
|
13 |
+
Special thanks to [George Sung](https://huggingface.co/georgesung) for creating [llama2_7b_chat_uncensored](https://huggingface.co/georgesung/llama2_7b_chat_uncensored), and to [Eric Hartford](https://huggingface.co/ehartford/) for creating [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered)
|
14 |
+
|
15 |
The version here is the fp16 HuggingFace model.
|
16 |
|
17 |
+
In 8 bit mode, the model fits into 84% of A100 80GB (67.2GB) 68747MiB
|
18 |
+
In 4 bit mode, the model fits into 51% of A100 80GB (40.8GB) 41559MiB
|
19 |
500gb of RAM/Swap was required to merge the model.
|
20 |
|
21 |
## GGML & GPTQ versions
|
|
|
93 |
|
94 |
|
95 |
# Training code
|
|
|
|
|
96 |
Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
|
97 |
|
98 |
To reproduce the results:
|
|
|
133 |
```
|
134 |
|
135 |
# Fine-tuning guide
|
136 |
+
https://georgesung.github.io/ai/qlora-ift/
|
|
|
|