hdtrnk commited on
Commit
0c8be5f
·
1 Parent(s): 016d8c0

Add full set of FusioniX checkpoints and README

Browse files
README.md ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # hdtrnk / FusioniX-All
2
+
3
+ **A complete collection of WanGP FusioniX checkpoints**
4
+ – both pure 16-bit dumps and pre-quantized 8-bit versions for image-to-video (I2V) and Phantom text-to-video (T2V) pipelines.
5
+
6
+ ---
7
+
8
+ ## 📂 Files in this repo
9
+
10
+ ### Pure FP16 checkpoints (~28–32 GB each)
11
+ - `Wan14Bi2vFusioniX_pure_fp16.safetensors`
12
+ *Wan2.1 Image→Video FusioniX 14B, pure 16-bit*
13
+
14
+ - `Wan14BT2VFusioniX_Phantom_pure_fp16.safetensors`
15
+ *Wan2.1 Phantom Text→Video FusioniX 14B, pure 16-bit*
16
+
17
+ ### Pre-quantized BF16→Int8 checkpoints (~14–16 GB each)
18
+ - `Wan14Bi2vFusioniX_pure_quanto_bf16_int8.safetensors`
19
+ *I2V FusioniX merged, quantized via Quanto (bf16→int8)*
20
+
21
+ - `Wan14BT2VFusioniX_Phantom_pure_quanto_bf16_int8.safetensors`
22
+ *Phantom T2V FusioniX merged, quantized via Quanto (bf16→int8)*
23
+
24
+ ---
25
+
26
+ ## 🚀 Quick Start (WanGP)
27
+
28
+ 1. **Copy** your desired file into your WanGP `ckpts/` folder.
29
+ 2. **Edit** your finetune JSON (in `app/finetunes/`) so that under `"model"`:
30
+ ```jsonc
31
+ "architecture": "i2v", // or "phantom_14B"
32
+ "URLs": ["ckpts/<filename>.safetensors"],
33
+ "auto_quantize": false
34
+ ```
35
+ 3. **Restart** WanGP (no `--save-quantized` flag).
36
+ 4. In the web UI, set **Transformer Model Quantization → Scaled Int8** (for the int8 files) or leave at BF16/FP16 for pure-16-bit.
37
+ 5. **Generate**!
38
+
39
+ ---
40
+
41
+ ## ⚙️ Recommended Defaults
42
+
43
+ ### I2V FusioniX
44
+ ```yaml
45
+ resolution: "832x480"
46
+ video_length: 81
47
+ num_steps: 8 # sweet-spot: 6–10
48
+ guidance_scale: 6.0
49
+ flow_shift: 7.0
50
+ embed_guide: 6
51
+ prompt_enhancer:on
52
+ slg_switch: 0
53
+ ```
54
+
55
+ ### Phantom T2V FusioniX
56
+ ```yaml
57
+ resolution: "832x480"
58
+ video_length: 81
59
+ num_steps: 8 # sweet-spot: 6–10
60
+ guidance_scale: 7.5
61
+ flow_shift: 5.0
62
+ embed_guide: 6
63
+ prompt_enhancer:on
64
+ slg_switch: 0
65
+ ```
66
+
67
+ ---
68
+
69
+ ## 🔍 Architecture & Compatibility
70
+
71
+ - **WanGP architectures**:
72
+ - `i2v` for image-to-video
73
+ - `phantom_14B` for Phantom text-to-video
74
+ - (See `model_signatures` in `wgp.py` or browse `app/settings/*.json` for all supported keys.)
75
+
76
+ - **WanGP version**: ≥ v6.0 (mmgp 3.4.9)
77
+ - **ComfyUI**: You can load these safetensors in ComfyUI’s Checkpoint Merger or as a Diffusers pipeline, but you may need a custom video pipeline config. They’re _not_ plug-and-play ComfyUI “modelcards.”
78
+
79
+ ---
80
+
81
+ ## 📋 JSON Snippets
82
+
83
+ ### I2V Finetune JSON
84
+ ```json
85
+ {
86
+ "model": {
87
+ "name": "Wan image2video FusioniX 14B",
88
+ "architecture": "i2v",
89
+ "URLs": ["ckpts/Wan14Bi2vFusioniX_pure_quanto_bf16_int8.safetensors"],
90
+ "auto_quantize": false
91
+ },
92
+ "... other settings ..."
93
+ }
94
+ ```
95
+
96
+ ### Phantom T2V Finetune JSON
97
+ ```json
98
+ {
99
+ "model": {
100
+ "name": "Wan text2video Phantom FusioniX 14B",
101
+ "architecture": "phantom_14B",
102
+ "URLs": ["ckpts/Wan14BT2VFusioniX_Phantom_pure_quanto_bf16_int8.safetensors"],
103
+ "auto_quantize": false
104
+ },
105
+ "... other settings ..."
106
+ }
107
+ ```
108
+
109
+ ---
110
+
111
+ ## 🔗 Upstream & Licensing
112
+
113
+ - **Original WanGP repo** by DeepBeepMeep:
114
+ https://github.com/DeepBeepMeep/WanGP
115
+
116
+ - **Original HF models**:
117
+ https://huggingface.co/DeepBeepMeep/Wan2.1
118
+
119
+ - **License**: Inherit from the upstream (check `LICENSE` in the Wan2.1 repo, typically MIT/Apache-2.0).
120
+
121
+ ---
122
+
123
+ *Happy generating!* 🎥✨
Wan14BT2VFusioniX_Phantom_pure_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd250bc6f68da18674a0bdfeee16c9f0f36a80405e5d86f77f9601c00710de08
3
+ size 28577119680
Wan14BT2VFusioniX_Phantom_pure_quanto_bf16_int8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a40335532193c35507af663b3c91a3385fa670cc2f6386418f06e043a7bc771
3
+ size 14532477217
Wan14Bi2vFusioniX_pure_fp16.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab079f8ceb43705d16948c0a34f14ef8f49324fe51e0d5a5f2798b667d5d40eb
3
+ size 32789921392
Wan14Bi2vFusioniX_pure_quanto_bf16_int8.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:822ba84fed8b172cf8e7a4cce9f10634bd839b979e20d1f12615eea791e9a511
3
+ size 16641621899