sese_ok / docs /Spell-book.md
big-thousand
init
699163a

You have now entered a hidden corner of the internet.

A confusing yet intriguing realm of paradoxes and contradictions.

A place where you will find out that what you thought you knew, you in fact didn't know, and what you didn't know was in front of you all along.

In other words, here I will document little-known facts about this web UI that I could not find another place for in the wiki.

You can train LoRAs in CPU mode

Load the web UI with

python server.py --cpu

and start training the LoRA from the training tab as usual.

8-bit mode works with CPU offloading

python server.py --load-in-8bit --gpu-memory 4000MiB

--pre_layer, and not --gpu-memory, is the right way to do CPU offloading with 4-bit models

python server.py --wbits 4 --groupsize 128 --pre_layer 20

Models can be loaded in 32-bit, 16-bit, 8-bit, and 4-bit modes

python server.py --cpu
python server.py
python server.py --load-in-8bit
python server.py --wbits 4

The web UI works with any version of GPTQ-for-LLaMa

Including the up to date triton and cuda branches. But you have to delete the repositories/GPTQ-for-LLaMa folder and reinstall the new one every time:

cd text-generation-webui/repositories
rm -r GPTQ-for-LLaMa
pip uninstall quant-cuda
git clone https://github.com/oobabooga/GPTQ-for-LLaMa -b cuda # or any other repository and branch
cd GPTQ-for-LLaMa
python setup_cuda.py install

Instruction-following templates are represented as chat characters

https://github.com/oobabooga/text-generation-webui/tree/main/characters/instruction-following

The right way to run Alpaca, Open Assistant, Vicuna, etc is Instruct mode, not normal chat mode

Otherwise the prompt will not be formatted correctly.

  1. Start the web UI with
python server.py --chat
  1. Click on the "instruct" option under "Chat modes"

  2. Select the correct template in the hidden dropdown menu that will become visible.

Notebook mode is best mode

Ascended individuals have realized that notebook mode is the superset of chat mode and can do chats with ultimate flexibility, including group chats, editing replies, starting a new bot reply in a given way, and impersonating.

RWKV is a RNN

Most models are transformers, but not RWKV, which is a RNN. It's a great model.

--gpu-memory is not a hard limit on the GPU memory

It is simply a parameter that is passed to the accelerate library while loading the model. More memory will be allocated during generation. That's why this parameter has to be set to less than your total GPU memory.

Contrastive search perhaps the best preset

But it uses a ton of VRAM.

You can check the sha256sum of downloaded models with the download script

python download-model.py facebook/galactica-125m --check

The download script continues interrupted downloads by default

It doesn't start over.

You can download models with multiple threads

python download-model.py facebook/galactica-125m --threads 8

LoRAs work in 4-bit mode

You need to follow these instructions and then start the web UI with the --monkey-patch flag.