Can you give an example on how to use gguf version or any inference GUI support it now?

#4
by DrNicefellow - opened
VIDraft org

As title

example,

  1. https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
  2. "Deploy" button click
  3. "HF Inference Endpoints" button click
  4. "Inference Endpoints" menu, "GPU Server" choice
  5. 'Deploy' and 'Excute Running'

or

"git clone ..... model or space path...." <- your server

or

  1. Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
  2. "Duplicate"
  3. files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
  4. commit -> building

As title

example,

  1. https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
  2. "Deploy" button click
  3. "HF Inference Endpoints" button click
  4. "Inference Endpoints" menu, "GPU Server" choice
  5. 'Deploy' and 'Excute Running'

or

"git clone ..... model or space path...." <- your server

or

  1. Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
  2. "Duplicate"
  3. files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
  4. commit -> building

Thank you for the information! I will try them.

DrNicefellow changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment