Can you give an example on how to use gguf version or any inference GUI support it now?
#4
by
DrNicefellow
- opened
As title
As title
example,
- https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
- "Deploy" button click
- "HF Inference Endpoints" button click
- "Inference Endpoints" menu, "GPU Server" choice
- 'Deploy' and 'Excute Running'
or
"git clone ..... model or space path...." <- your server
or
- Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
- "Duplicate"
- files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
- commit -> building
As title
example,
- https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
- "Deploy" button click
- "HF Inference Endpoints" button click
- "Inference Endpoints" menu, "GPU Server" choice
- 'Deploy' and 'Excute Running'
or
"git clone ..... model or space path...." <- your server
or
- Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
- "Duplicate"
- files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
- commit -> building
Thank you for the information! I will try them.
DrNicefellow
changed discussion status to
closed