GGUF

QuantFactory Banner

QuantFactory/Promt-generator-GGUF

This is quantized version of UnfilteredAI/Promt-generator created using llama.cpp

Original Model Card

Model Card: UnfilteredAI/Promt-generator

Model Overview

The UnfilteredAI/Promt-generator is a text generation model designed specifically for creating prompts for text-to-image models. It leverages PyTorch and safetensors for optimized performance and storage, ensuring that it can be easily deployed and scaled for prompt generation tasks.

Intended Use

This model is primarily intended for:

  • Prompt generation for text-to-image models.
  • Creative AI applications where generating high-quality, diverse image descriptions is critical.
  • Supporting AI artists and developers working on generative art projects.

How to Use

To generate prompts using this model, follow these steps:

  1. Load the model in your PyTorch environment.
  2. Input your desired parameters for the prompt generation task.
  3. The model will return text descriptions based on the input, which can then be used with text-to-image models.

Example Code:

from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("UnfilteredAI/Promt-generator")
model = AutoModelForCausalLM.from_pretrained("UnfilteredAI/Promt-generator")

prompt = "a red car"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs)
generated_prompt = tokenizer.decode(outputs[0], skip_special_tokens=True)

print(generated_prompt)
Downloads last month
40
GGUF
Model size
0.8B params
Architecture
bloom
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support