Can you give an example on how to use gguf version or any inference GUI support it now?

#4
by DrNicefellow - opened

As title

VIDraft org

As title

example,

  1. https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
  2. "Deploy" button click
  3. "HF Inference Endpoints" button click
  4. "Inference Endpoints" menu, "GPU Server" choice
  5. 'Deploy' and 'Excute Running'

or

"git clone ..... model or space path...." <- your server

or

  1. Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
  2. "Duplicate"
  3. files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
  4. commit -> building

As title

example,

  1. https://huggingface.co/openfree/Gemma-3-R1984-27B-Q8_0-GGUF
  2. "Deploy" button click
  3. "HF Inference Endpoints" button click
  4. "Inference Endpoints" menu, "GPU Server" choice
  5. 'Deploy' and 'Excute Running'

or

"git clone ..... model or space path...." <- your server

or

  1. Link: https://huggingface.co/spaces/VIDraft/Gemma-3-R1984-27B
  2. "Duplicate"
  3. files -> app.py -> code edit -> Replicated "your choic GGUF Model Path"
  4. commit -> building

Thank you for the information! I will try them.

DrNicefellow changed discussion status to closed

Sign up or log in to comment