Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ it contains the following quantizations of the original weights from Together's
|
|
43 |
* ~~[Q8_0](https://huggingface.co/rozek/LLaMA-2-7B-32K-Instruct_GGUF/blob/main/LLaMA-2-7B-32K-Instruct-Q8_0.gguf)~~ and
|
44 |
* ~~[F16](https://huggingface.co/rozek/LLaMA-2-7B-32K-Instruct_GGUF/blob/main/LLaMA-2-7B-32K-Instruct-f16.gguf)~~ (unquantized)
|
45 |
|
46 |
-
(strikethrough
|
47 |
|
48 |
> Nota bene: while RoPE makes inferences with large contexts possible, you still need an awful lot of RAM
|
49 |
> when doing so. And since "32K" does not mean that you always have to use a context size of 32768 (only that
|
|
|
43 |
* ~~[Q8_0](https://huggingface.co/rozek/LLaMA-2-7B-32K-Instruct_GGUF/blob/main/LLaMA-2-7B-32K-Instruct-Q8_0.gguf)~~ and
|
44 |
* ~~[F16](https://huggingface.co/rozek/LLaMA-2-7B-32K-Instruct_GGUF/blob/main/LLaMA-2-7B-32K-Instruct-f16.gguf)~~ (unquantized)
|
45 |
|
46 |
+
(strikethrough files are currently being uploaded)
|
47 |
|
48 |
> Nota bene: while RoPE makes inferences with large contexts possible, you still need an awful lot of RAM
|
49 |
> when doing so. And since "32K" does not mean that you always have to use a context size of 32768 (only that
|