Image-to-Video
Diffusers
Safetensors
Spanish
s2v

Improve model card: Add pipeline tag, library name, specific paper, and project page for Wan-S2V

#5
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +16 -3
README.md CHANGED
@@ -1,21 +1,34 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
- # Wan2.2
 
 
 
 
 
 
 
5
 
6
  <p align="center">
7
  <img src="assets/logo.png" width="400"/>
8
  <p>
9
 
10
  <p align="center">
11
- 💜 <a href="https://wan.video"><b>Wan</b></a> &nbsp&nbsp | &nbsp&nbsp 🖥️ <a href="https://github.com/Wan-Video/Wan2.2">GitHub</a> &nbsp&nbsp | &nbsp&nbsp🤗 <a href="https://huggingface.co/Wan-AI/">Hugging Face</a>&nbsp&nbsp | &nbsp&nbsp🤖 <a href="https://modelscope.cn/organization/Wan-AI">ModelScope</a>&nbsp&nbsp | &nbsp&nbsp 📑 <a href="https://arxiv.org/abs/2503.20314">Paper</a> &nbsp&nbsp | &nbsp&nbsp 📑 <a href="https://wan.video/welcome?spm=a2ty_o02.30011076.0.0.6c9ee41eCcluqg">Blog</a> &nbsp&nbsp | &nbsp&nbsp 💬 <a href="https://discord.gg/AKNgpMK4Yj">Discord</a>&nbsp&nbsp
12
  <br>
13
  📕 <a href="https://alidocs.dingtalk.com/i/nodes/jb9Y4gmKWrx9eo4dCql9LlbYJGXn6lpz">使用指南(中文)</a>&nbsp&nbsp | &nbsp&nbsp 📘 <a href="https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y">User Guide(English)</a>&nbsp&nbsp | &nbsp&nbsp💬 <a href="https://gw.alicdn.com/imgextra/i2/O1CN01tqjWFi1ByuyehkTSB_!!6000000000015-0-tps-611-1279.jpg">WeChat(微信)</a>&nbsp&nbsp
14
  <br>
15
 
 
 
 
 
16
  -----
17
 
18
- [**Wan: Open and Advanced Large-Scale Video Generative Models**](https://arxiv.org/abs/2503.20314) <be>
19
 
20
 
21
  We are excited to introduce **Wan2.2**, a major upgrade to our foundational video models. With **Wan2.2**, we have focused on incorporating the following innovations:
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: other
4
+ library_name: diffusers
5
  ---
6
+
7
+ # Wan2.2-S2V-14B: Audio-Driven Cinematic Video Generation
8
+
9
+ This repository features the **Wan2.2-S2V-14B** model, designed for audio-driven cinematic video generation. It was introduced in the paper:
10
+ [**Wan-S2V: Audio-Driven Cinematic Video Generation**](https://huggingface.co/papers/2508.18621)
11
+
12
+ Project Page: [https://humanaigc.github.io/wan-s2v-webpage](https://humanaigc.github.io/wan-s2v-webpage)
13
+ GitHub Repository: [https://github.com/Wan-Video/Wan2.2](https://github.com/Wan-Video/Wan2.2)
14
 
15
  <p align="center">
16
  <img src="assets/logo.png" width="400"/>
17
  <p>
18
 
19
  <p align="center">
20
+ 💜 <a href="https://wan.video"><b>Wan Homepage</b></a> &nbsp&nbsp | &nbsp&nbsp 🖥️ <a href="https://github.com/Wan-Video/Wan2.2">GitHub</a> &nbsp&nbsp | &nbsp&nbsp🤗 <a href="https://huggingface.co/Wan-AI/">Hugging Face Organization</a>&nbsp&nbsp | &nbsp&nbsp🤖 <a href="https://modelscope.cn/organization/Wan-AI">ModelScope Organization</a>&nbsp&nbsp | &nbsp&nbsp 📑 <a href="https://huggingface.co/papers/2508.18621">Wan-S2V Paper</a> &nbsp&nbsp | &nbsp&nbsp 📑 <a href="https://arxiv.org/abs/2503.20314">Wan2.2 Base Paper</a> &nbsp&nbsp | 🌐 <a href="https://humanaigc.github.io/wan-s2v-webpage">Project Page</a> &nbsp&nbsp | &nbsp&nbsp 📑 <a href="https://wan.video/welcome?spm=a2ty_o02.30011076.0.0.6c9ee41eCcluqg">Blog</a> &nbsp&nbsp | &nbsp&nbsp 💬 <a href="https://discord.gg/AKNgpMK4Yj">Discord</a>&nbsp&nbsp
21
  <br>
22
  📕 <a href="https://alidocs.dingtalk.com/i/nodes/jb9Y4gmKWrx9eo4dCql9LlbYJGXn6lpz">使用指南(中文)</a>&nbsp&nbsp | &nbsp&nbsp 📘 <a href="https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y">User Guide(English)</a>&nbsp&nbsp | &nbsp&nbsp💬 <a href="https://gw.alicdn.com/imgextra/i2/O1CN01tqjWFi1ByuyehkTSB_!!6000000000015-0-tps-611-1279.jpg">WeChat(微信)</a>&nbsp&nbsp
23
  <br>
24
 
25
+ ## Abstract (Wan-S2V Paper)
26
+
27
+ Current state-of-the-art (SOTA) methods for audio-driven character animation demonstrate promising performance for scenarios primarily involving speech and singing. However, they often fall short in more complex film and television productions, which demand sophisticated elements such as nuanced character interactions, realistic body movements, and dynamic camera work. To address this long-standing challenge of achieving film-level character animation, we propose an audio-driven model, which we refere to as Wan-S2V, built upon Wan. Our model achieves significantly enhanced expressiveness and fidelity in cinematic contexts compared to existing approaches. We conducted extensive experiments, benchmarking our method against cutting-edge models such as Hunyuan-Avatar and Omnihuman. The experimental results consistently demonstrate that our approach significantly outperforms these existing solutions. Additionally, we explore the versatility of our method through its applications in long-form video generation and precise video lip-sync editing.
28
+
29
  -----
30
 
31
+ [**Wan: Open and Advanced Large-Scale Video Generative Models**](https://arxiv.org/abs/2503.20314) <br>
32
 
33
 
34
  We are excited to introduce **Wan2.2**, a major upgrade to our foundational video models. With **Wan2.2**, we have focused on incorporating the following innovations: