Upload WanSpeechToVideoPipeline
Browse files
transformer/diffusion_pytorch_model-00001-of-00007.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4972707834
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a982f4454fbc5207dbe5cd417cdd538da08c800001c7d7c68adccb18897fc4d1
|
3 |
size 4972707834
|
transformer/diffusion_pytorch_model.safetensors.index.json
CHANGED
@@ -1227,18 +1227,18 @@
|
|
1227 |
"blocks.9.norm2.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1228 |
"blocks.9.norm2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1229 |
"blocks.9.scale_shift_table": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1230 |
-
"condition_embedder.
|
1231 |
-
"condition_embedder.
|
1232 |
-
"condition_embedder.
|
1233 |
-
"condition_embedder.
|
1234 |
-
"condition_embedder.
|
1235 |
-
"condition_embedder.
|
1236 |
-
"condition_embedder.
|
1237 |
-
"condition_embedder.
|
1238 |
-
"condition_embedder.
|
1239 |
-
"condition_embedder.
|
1240 |
-
"condition_embedder.
|
1241 |
-
"condition_embedder.
|
1242 |
"condition_embedder.pose_embedder.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1243 |
"condition_embedder.pose_embedder.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1244 |
"condition_embedder.text_embedder.linear_1.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
|
|
1227 |
"blocks.9.norm2.bias": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1228 |
"blocks.9.norm2.weight": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1229 |
"blocks.9.scale_shift_table": "diffusion_pytorch_model-00002-of-00007.safetensors",
|
1230 |
+
"condition_embedder.causal_audio_encoder.encoder.conv1_global.conv.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1231 |
+
"condition_embedder.causal_audio_encoder.encoder.conv1_global.conv.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1232 |
+
"condition_embedder.causal_audio_encoder.encoder.conv1_local.conv.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1233 |
+
"condition_embedder.causal_audio_encoder.encoder.conv1_local.conv.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1234 |
+
"condition_embedder.causal_audio_encoder.encoder.conv2.conv.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1235 |
+
"condition_embedder.causal_audio_encoder.encoder.conv2.conv.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1236 |
+
"condition_embedder.causal_audio_encoder.encoder.conv3.conv.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1237 |
+
"condition_embedder.causal_audio_encoder.encoder.conv3.conv.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1238 |
+
"condition_embedder.causal_audio_encoder.encoder.final_linear.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1239 |
+
"condition_embedder.causal_audio_encoder.encoder.final_linear.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1240 |
+
"condition_embedder.causal_audio_encoder.encoder.padding_tokens": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1241 |
+
"condition_embedder.causal_audio_encoder.weights": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1242 |
"condition_embedder.pose_embedder.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1243 |
"condition_embedder.pose_embedder.weight": "diffusion_pytorch_model-00001-of-00007.safetensors",
|
1244 |
"condition_embedder.text_embedder.linear_1.bias": "diffusion_pytorch_model-00001-of-00007.safetensors",
|