Update README.md
Browse files
README.md
CHANGED
@@ -50,7 +50,7 @@ Press the [Open in Colab] button on the link to start Colab
|
|
50 |
|
51 |
### 利用可能なVersion(Available Versions)
|
52 |
|
53 |
-
llama.cpp
|
54 |
Although llama.cpp can be used to reduce the size of the file with various quantization methods, this sample deals with only six types. Smaller models can run faster with less memory, but also reduce the performance of the models. 4 bits (Q4_K_M) is said to be a good balance.
|
55 |
|
56 |
- C3TR-Adapter-IQ3_XXS.gguf 3.6GB
|
|
|
50 |
|
51 |
### 利用可能なVersion(Available Versions)
|
52 |
|
53 |
+
llama.cppを使うと、様々な量子化手法でファイルのサイズを小さくする事が出来ますが、本モデルでは7種類のみを扱います。小さいサイズのモデルは、少ないメモリで高速に動作させることができますが、モデルの性能も低下します。4ビット(Q4_K_M)くらいがバランスが良いと言われています。
|
54 |
Although llama.cpp can be used to reduce the size of the file with various quantization methods, this sample deals with only six types. Smaller models can run faster with less memory, but also reduce the performance of the models. 4 bits (Q4_K_M) is said to be a good balance.
|
55 |
|
56 |
- C3TR-Adapter-IQ3_XXS.gguf 3.6GB
|