Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

TheBloke
/
WizardLM-30B-Uncensored-GPTQ

Text Generation
Transformers
Safetensors
llama
uncensored
text-generation-inference
4-bit precision
gptq
Model card Files Files and versions Community
15
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

Adding Evaluation Results

#15 opened over 1 year ago by
leaderboard-pr-bot

is it possible to use a continuous batching inference server with this model?

#14 opened over 1 year ago by
natserrano

Can the model be sharded over several GPUs?

1
#13 opened almost 2 years ago by
silverfisk

Works well on a 3090 ! Insane good !

1
#11 opened about 2 years ago by
goldrushgames

Anyone has been successful in deploying this to Sagemaker or so?

#9 opened about 2 years ago by
rafa9

Model exits in Windows Booga

8
#8 opened about 2 years ago by
0spr4y

Model Performance Curiosity

3
#7 opened about 2 years ago by
sumuks

Have a problem with rtx4090

17
#6 opened about 2 years ago by
crainto

i got messy code

5
#3 opened about 2 years ago by
Tiankong2023

Will this work with the Local LLMs One-Click UI runpod?

8
#2 opened about 2 years ago by
nichedreams

How much vram+ram 30B needs? I have 3060 12gb + 32gb ram.

21
#1 opened about 2 years ago by
DaveScream
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs