GGUF quants?
#2
by
6346y9uey
- opened
As above
oh ok, I'll do IQ4_XS
llm_load_print_meta: model ftype = IQ4_XS - 4.25 bpw
llm_load_print_meta: model params = 121.89 B
llm_load_print_meta: model size = 61.35 GiB (4.32 BPW)
llm_load_print_meta: general.name = New Volume
llm_load_print_meta: BOS token = 128000 '<|begin_of_text|>'
llm_load_print_meta: EOS token = 128001 '<|end_of_text|>'
llm_load_print_meta: LF token = 128 'Ä'
llm_load_print_meta: EOT token = 128009 '<|eot_id|>'
llm_load_tensors: ggml ctx size = 0.64 MiB