wsbagnsv1's picture
Update README.md
2d5f166 verified
|
raw
history blame
2.46 kB
metadata
license: apache-2.0
base_model:
  - Lightricks/LTX-Video
library_name: gguf
tags:
  - video
  - video-generation
pipeline_tag: image-to-video

Comfyui doesnt natively support these Quants yet but there is a workaround:

First of all you need to load the vae, which you can download in this repo too and then

we need to edit this file Comfyui/comfy/ldm/lightricks/model.py

and change the init function for class LTXVModel(torch.nn.Module):

from this:

def __init__(self,
             in_channels=128,
             cross_attention_dim=2048,
             attention_head_dim=64,
             num_attention_heads=32,

             caption_channels=4096,
             num_layers=28,


             positional_embedding_theta=10000.0,
             positional_embedding_max_pos=[20, 2048, 2048],
             causal_temporal_positioning=False,
             vae_scale_factors=(8, 32, 32),
             dtype=None, device=None, operations=None, **kwargs):

to this:

def __init__(self,
             in_channels=128,
             cross_attention_dim=4096,
             attention_head_dim=128, 
             num_attention_heads=32,

             caption_channels=4096,
             num_layers=48,

             
             positional_embedding_theta=10000.0,
             positional_embedding_max_pos=[20, 2048, 2048],
             causal_temporal_positioning=False,
             vae_scale_factors=(8, 32, 32),
             dtype=None, device=None, operations=None, **kwargs):

If you want to use the 2b again just revert it and restart. After a restart the 13b model should work with this and the vae i uploaded.

This is a direct GGUF conversion of Lightricks/ltxv-13b-0.9.7-dev

All quants are created from the FP32 base file, though I only uploaded the Q8_0 and less, if you want the F16 or BF16 one I would upload it per request.

The model files can be used with the ComfyUI-GGUF custom node.

Place model files in ComfyUI/models/unet - see the GitHub readme for further install instructions.

Please refer to this chart for a basic overview of quantization types.

For conversion I used the conversion scripts from city96