KAD001 commited on
Commit
3ea2ecf
·
verified ·
1 Parent(s): f2373f1

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitignore +47 -0
  2. 1 +0 -0
  3. Custom Resolutions Instructions.txt +16 -0
  4. Dockerfile +92 -0
  5. LICENSE.txt +46 -0
  6. README.md +258 -12
  7. Wan2GP/.gitignore +47 -0
  8. Wan2GP/Custom Resolutions Instructions.txt +16 -0
  9. Wan2GP/Dockerfile +92 -0
  10. Wan2GP/LICENSE.txt +46 -0
  11. Wan2GP/README.md +252 -0
  12. Wan2GP/configs/animate.json +15 -0
  13. Wan2GP/configs/fantasy.json +15 -0
  14. Wan2GP/configs/flf2v_720p.json +15 -0
  15. Wan2GP/configs/i2v.json +14 -0
  16. Wan2GP/configs/i2v_2_2.json +14 -0
  17. Wan2GP/configs/i2v_2_2_multitalk.json +15 -0
  18. Wan2GP/configs/infinitetalk.json +15 -0
  19. Wan2GP/configs/lucy_edit.json +14 -0
  20. Wan2GP/configs/lynx.json +15 -0
  21. Wan2GP/configs/multitalk.json +15 -0
  22. Wan2GP/configs/phantom_1.3B.json +14 -0
  23. Wan2GP/configs/phantom_14B.json +14 -0
  24. Wan2GP/configs/qwen_image_20B.json +18 -0
  25. Wan2GP/configs/sky_df_1.3.json +14 -0
  26. Wan2GP/configs/sky_df_14B.json +14 -0
  27. Wan2GP/configs/standin.json +15 -0
  28. Wan2GP/configs/t2v.json +14 -0
  29. Wan2GP/configs/t2v_1.3B.json +14 -0
  30. Wan2GP/configs/ti2v_2_2.json +14 -0
  31. Wan2GP/configs/vace_1.3B.json +16 -0
  32. Wan2GP/configs/vace_14B.json +16 -0
  33. Wan2GP/configs/vace_lynx_14B.json +17 -0
  34. Wan2GP/configs/vace_multitalk_14B.json +17 -0
  35. Wan2GP/configs/vace_standin_14B.json +17 -0
  36. Wan2GP/defaults/ReadMe.txt +13 -0
  37. Wan2GP/defaults/animate.json +18 -0
  38. Wan2GP/defaults/fantasy.json +11 -0
  39. Wan2GP/defaults/flf2v_720p.json +16 -0
  40. Wan2GP/defaults/flux.json +16 -0
  41. Wan2GP/defaults/flux_chroma.json +18 -0
  42. Wan2GP/defaults/flux_dev_kontext.json +17 -0
  43. Wan2GP/defaults/flux_dev_umo.json +24 -0
  44. Wan2GP/defaults/flux_dev_uso.json +17 -0
  45. Wan2GP/defaults/flux_krea.json +16 -0
  46. Wan2GP/defaults/flux_schnell.json +17 -0
  47. Wan2GP/defaults/flux_srpo.json +15 -0
  48. Wan2GP/defaults/flux_srpo_uso.json +17 -0
  49. Wan2GP/defaults/fun_inp.json +13 -0
  50. Wan2GP/defaults/fun_inp_1.3B.json +11 -0
.gitignore ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ .*
2
+ *.py[cod]
3
+ # *.jpg
4
+ *.jpeg
5
+ # *.png
6
+ *.gif
7
+ *.bmp
8
+ *.mp4
9
+ *.mov
10
+ *.mkv
11
+ *.log
12
+ *.zip
13
+ *.pt
14
+ *.pth
15
+ *.ckpt
16
+ *.safetensors
17
+ #*.json
18
+ # *.txt
19
+ *.backup
20
+ *.pkl
21
+ *.html
22
+ *.pdf
23
+ *.whl
24
+ *.exe
25
+ cache
26
+ __pycache__/
27
+ storage/
28
+ samples/
29
+ !.gitignore
30
+ !requirements.txt
31
+ .DS_Store
32
+ *DS_Store
33
+ google/
34
+ Wan2.1-T2V-14B/
35
+ Wan2.1-T2V-1.3B/
36
+ Wan2.1-I2V-14B-480P/
37
+ Wan2.1-I2V-14B-720P/
38
+ outputs/
39
+ outputs2/
40
+ gradio_outputs/
41
+ ckpts/
42
+ loras/
43
+ loras_i2v/
44
+
45
+ settings/
46
+
47
+ wgp_config.json
1 ADDED
File without changes
Custom Resolutions Instructions.txt ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ You can override the choice of Resolutions offered by WanGP, if you create a file "resolutions.json" in the main WanGP folder.
2
+ This file is composed of a list of 2 elements sublists. Each 2 elements sublist should have the format ["Label", "WxH"] where W, H are respectively the Width and Height of the resolution. Please make sure that W and H are multiples of 16. The letter "x" should be placed inbetween these two dimensions.
3
+
4
+ Here is below a sample "resolutions.json" file :
5
+
6
+ [
7
+ ["1280x720 (16:9, 720p)", "1280x720"],
8
+ ["720x1280 (9:16, 720p)", "720x1280"],
9
+ ["1024x1024 (1:1, 720p)", "1024x1024"],
10
+ ["1280x544 (21:9, 720p)", "1280x544"],
11
+ ["544x1280 (9:21, 720p)", "544x1280"],
12
+ ["1104x832 (4:3, 720p)", "1104x832"],
13
+ ["832x1104 (3:4, 720p)", "832x1104"],
14
+ ["960x960 (1:1, 720p)", "960x960"],
15
+ ["832x480 (16:9, 480p)", "832x480"]
16
+ ]
Dockerfile ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM nvidia/cuda:12.4.1-cudnn-devel-ubuntu22.04
2
+
3
+ # Build arg for GPU architectures - specify which CUDA compute capabilities to compile for
4
+ # Common values:
5
+ # 7.0 - Tesla V100
6
+ # 7.5 - RTX 2060, 2070, 2080, Titan RTX
7
+ # 8.0 - A100, A800 (Ampere data center)
8
+ # 8.6 - RTX 3060, 3070, 3080, 3090 (Ampere consumer)
9
+ # 8.9 - RTX 4070, 4080, 4090 (Ada Lovelace)
10
+ # 9.0 - H100, H800 (Hopper data center)
11
+ # 12.0 - RTX 5070, 5080, 5090 (Blackwell) - Note: sm_120 architecture
12
+ #
13
+ # Examples:
14
+ # RTX 3060: --build-arg CUDA_ARCHITECTURES="8.6"
15
+ # RTX 4090: --build-arg CUDA_ARCHITECTURES="8.9"
16
+ # Multiple: --build-arg CUDA_ARCHITECTURES="8.0;8.6;8.9"
17
+ #
18
+ # Note: Including 8.9 or 9.0 may cause compilation issues on some setups
19
+ # Default includes 8.0 and 8.6 for broad Ampere compatibility
20
+ ARG CUDA_ARCHITECTURES="8.0;8.6"
21
+
22
+ ENV DEBIAN_FRONTEND=noninteractive
23
+
24
+ # Install system dependencies
25
+ RUN apt update && \
26
+ apt install -y \
27
+ python3 python3-pip git wget curl cmake ninja-build \
28
+ libgl1 libglib2.0-0 ffmpeg && \
29
+ apt clean
30
+
31
+ WORKDIR /workspace
32
+
33
+ COPY requirements.txt .
34
+
35
+ # Upgrade pip first
36
+ RUN pip install --upgrade pip setuptools wheel
37
+
38
+ # Install requirements if exists
39
+ RUN pip install -r requirements.txt
40
+
41
+ # Install PyTorch with CUDA support
42
+ RUN pip install --extra-index-url https://download.pytorch.org/whl/cu124 \
43
+ torch==2.6.0+cu124 torchvision==0.21.0+cu124
44
+
45
+ # Install SageAttention from git (patch GPU detection)
46
+ ENV TORCH_CUDA_ARCH_LIST="${CUDA_ARCHITECTURES}"
47
+ ENV FORCE_CUDA="1"
48
+ ENV MAX_JOBS="1"
49
+
50
+ COPY <<EOF /tmp/patch_setup.py
51
+ import os
52
+ with open('setup.py', 'r') as f:
53
+ content = f.read()
54
+
55
+ # Get architectures from environment variable
56
+ arch_list = os.environ.get('TORCH_CUDA_ARCH_LIST')
57
+ arch_set = '{' + ', '.join([f'"{arch}"' for arch in arch_list.split(';')]) + '}'
58
+
59
+ # Replace the GPU detection section
60
+ old_section = '''compute_capabilities = set()
61
+ device_count = torch.cuda.device_count()
62
+ for i in range(device_count):
63
+ major, minor = torch.cuda.get_device_capability(i)
64
+ if major < 8:
65
+ warnings.warn(f"skipping GPU {i} with compute capability {major}.{minor}")
66
+ continue
67
+ compute_capabilities.add(f"{major}.{minor}")'''
68
+
69
+ new_section = 'compute_capabilities = ' + arch_set + '''
70
+ print(f"Manually set compute capabilities: {compute_capabilities}")'''
71
+
72
+ content = content.replace(old_section, new_section)
73
+
74
+ with open('setup.py', 'w') as f:
75
+ f.write(content)
76
+ EOF
77
+
78
+ RUN git clone https://github.com/thu-ml/SageAttention.git /tmp/sageattention && \
79
+ cd /tmp/sageattention && \
80
+ python3 /tmp/patch_setup.py && \
81
+ pip install --no-build-isolation .
82
+
83
+ RUN useradd -u 1000 -ms /bin/bash user
84
+
85
+ RUN chown -R user:user /workspace
86
+
87
+ RUN mkdir /home/user/.cache && \
88
+ chown -R user:user /home/user/.cache
89
+
90
+ COPY entrypoint.sh /workspace/entrypoint.sh
91
+
92
+ ENTRYPOINT ["/workspace/entrypoint.sh"]
LICENSE.txt ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ WanGP NON-COMMERCIAL EVALUATION LICENSE 1.0
2
+
3
+ Definitions
4
+ 1.1 “Software” means the source code, binaries, libraries, utilities and UI released under this license.
5
+ 1.2 “Output” means images, videos or other media produced by running the Software.
6
+ 1.3 “Commercial Use” means:
7
+ a) selling, sublicensing, renting, leasing, or otherwise distributing the Software, in whole or in part, for a fee or other consideration; or
8
+ b) offering the Software (or any derivative) as part of a paid product or hosted service; or
9
+ c) using the Software (or any derivative) to provide cloud-based or backend services, where end users access or pay for those services.
10
+
11
+ License Grant
12
+ Subject to Section 3:
13
+ a) You are granted a worldwide, non-exclusive, royalty-free, revocable license to use, reproduce, modify and distribute the Software for non-commercial purposes only.
14
+ b) You are granted a worldwide, non-exclusive, royalty-free, irrevocable license to use, reproduce, modify and distribute the Output for any purpose, including commercial sale, provided that any commercial distribution of the Output includes a clear notice that the Output was produced (in whole or in part) using WanGP, along with a hyperlink to the WanGP application’s About tab or repository.
15
+
16
+ Restrictions
17
+ 3.1 You MAY NOT distribute, sublicense or otherwise make available the Software (or any derivative) for Commercial Use.
18
+ 3.2 You MAY sell, license or otherwise commercially exploit the Output without restriction.
19
+ 3.3 If you wish to use the Software for Commercial Use, you must obtain a separate commercial license from the Licensor.
20
+
21
+ Third-Party Components 4.1 The Software includes components licensed under various open-source licenses (e.g., Apache 2.0, MIT, BSD). 4.2 You must comply with all applicable terms of those third-party licenses, including preservation of copyright notices, inclusion of required license texts, and patent-grant provisions. 4.3 You can find the full text of each third-party license via the “About” tab in the WanGP application, which provides links to their original GitHub repositories.
22
+
23
+ Attribution
24
+ 5.1 You must give appropriate credit by including:
25
+ • a copy of this license (or a link to it), and
26
+ • a notice that your use is based on “WanGP”.
27
+ 5.2 You may do so in any reasonable manner, but not in any way that suggests the Licensor endorses you or your use.
28
+
29
+ Disclaimer of Warranty & Liability
30
+ THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED.
31
+ IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE.
32
+
33
+ Commercial Licensing The Licensor may offer commercial licenses for the Software, which grant rights to use the Software for Commercial Use. Please contact [[email protected]] for terms and pricing.
34
+
35
+ Effective Date & Previous Versions
36
+ 8.1 This license is effective as of the date the LICENSE file is updated in the WanGP repository.
37
+ 8.2 Any copies of the Software obtained under prior license terms before this Effective Date remain governed by those prior terms; such granted rights are irrevocable.
38
+ 8.3 Use of the Software after the release of any subsequent version by the Licensor is subject to the terms of the then-current license, unless a separate agreement is in place.
39
+
40
+ Acceptable Use / Moral Clause
41
+ 9.1 You MAY NOT use the Software or the Output to facilitate or produce content that is illegal, harmful, violent, harassing, defamatory, fraudulent, or otherwise violates applicable laws or fundamental human rights.
42
+ 9.2 You MAY NOT deploy the Software or Output in contexts that promote hate speech, extremist ideology, human rights abuses, or other actions that could foreseeably cause significant harm to individuals or groups.
43
+ 9.3 The Licensor reserves the right to terminate the rights granted under this license if a licensee materially breaches this Acceptable Use clause.
44
+
45
+ END OF LICENSE
46
+
README.md CHANGED
@@ -1,12 +1,258 @@
1
- ---
2
- title: Wan2GP
3
- emoji:
4
- colorFrom: purple
5
- colorTo: purple
6
- sdk: gradio
7
- sdk_version: 5.49.1
8
- app_file: app.py
9
- pinned: false
10
- ---
11
-
12
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ title: Wan2GP
3
+ app_file: wgp.py
4
+ sdk: gradio
5
+ sdk_version: 5.29.0
6
+ ---
7
+ # WanGP
8
+
9
+ -----
10
+ <p align="center">
11
+ <b>WanGP by DeepBeepMeep : The best Open Source Video Generative Models Accessible to the GPU Poor</b>
12
+ </p>
13
+
14
+ WanGP supports the Wan (and derived models), Hunyuan Video and LTV Video models with:
15
+ - Low VRAM requirements (as low as 6 GB of VRAM is sufficient for certain models)
16
+ - Support for old GPUs (RTX 10XX, 20xx, ...)
17
+ - Very Fast on the latest GPUs
18
+ - Easy to use Full Web based interface
19
+ - Auto download of the required model adapted to your specific architecture
20
+ - Tools integrated to facilitate Video Generation : Mask Editor, Prompt Enhancer, Temporal and Spatial Generation, MMAudio, Video Browser, Pose / Depth / Flow extractor
21
+ - Loras Support to customize each model
22
+ - Queuing system : make your shopping list of videos to generate and come back later
23
+
24
+ **Discord Server to get Help from Other Users and show your Best Videos:** https://discord.gg/g7efUW9jGV
25
+
26
+ **Follow DeepBeepMeep on Twitter/X to get the Latest News**: https://x.com/deepbeepmeep
27
+
28
+ ## 🔥 Latest Updates :
29
+ ### October 3 2025: WanGP v8.992 - One Last Thing before the Big Unknown ...
30
+
31
+ This new version hasn't any new model...
32
+
33
+ ...but temptation to upgrade will be high as it contains a few Loras related features that may change your Life:
34
+ - **Ready to use Loras Accelerators Profiles** per type of model that you can apply on your current *Generation Settings*. Next time I will recommend a *Lora Accelerator*, it will be only one click away. And best of all of the required Loras will be downloaded automatically. When you apply an *Accelerator Profile*, input fields like the *Number of Denoising Steps* *Activated Loras*, *Loras Multipliers* (such as "1;0 0;1" ...) will be automatically filled. However your video specific fields will be preserved, so it will be easy to switch between Profiles to experiment.
35
+
36
+ - **Embedded Loras URL** : WanGP will now try to remember every Lora URLs it sees. For instance if someone sends you some settings that contain Loras URLs or you extract the Settings of Video generated by a friend with Loras URLs, these URLs will be automatically added to *WanGP URL Cache*. Conversely everything you will share (Videos, Settings, Lset files) will contain the download URLs if they are known. You can also download directly a Lora in WanGP by using the *Download Lora* button a the bottom. The Lora will be immediatly available and added to WanGP lora URL cache. This will work with *Hugging Face* as a repository. Support for CivitAi will come as soon as someone will nice enough to post a GitHub PR ...
37
+
38
+ - **.lset file** supports embedded Loras URLs. It has never been easier to share a Lora with a friend. As a reminder a .lset file can be created directly from *WanGP Web Interface* and it contains a list of Loras and their multipliers, a Prompt and Instructions how to use these loras (like the Lora's *Trigger*). So with embedded Loras URL, you can send an .lset file by email or share it on discord: it is just a 1 KB tiny text, but with it other people will be able to use Gigabytes Loras as these will be automatically downloaded.
39
+
40
+ I have created the new Discord Channel **share-your-settings** where you can post your *Settings* or *Lset files*. I will be pleased to add new Loras Accelerators in the list of WanGP *Accelerators Profiles if you post some good ones there.
41
+
42
+ Last but not least the Lora's documentation has been updated.
43
+
44
+ *update 8.991*: full power of *Vace Lynx* unleashed with new combinations such as Landscape + Face / Clothes + Face / Injectd Frame (Start/End frames/...) + Face
45
+ *update 8.992*: optimized gen with Lora, should be 10% if many loras
46
+
47
+ ### September 30 2025: WanGP v8.9 - Combinatorics
48
+
49
+ This new version of WanGP introduces **Wan 2.1 Lynx** the best Control Net so far to transfer *Facial Identity*. You will be amazed to recognize your friends even with a completely different hair style. Congrats to the *Byte Dance team* for this achievement. Lynx works quite with well *Fusionix t2v* 10 steps.
50
+
51
+ *WanGP 8.9* also illustrate how existing WanGP features can be easily combined with new models. For instance with *Lynx* you will get out of the box *Video to Video* and *Image/Text to Image*.
52
+
53
+ Another fun combination is *Vace* + *Lynx*, which works much better than *Vace StandIn*. I have added sliders to change the weight of Vace & Lynx to allow you to tune the effects.
54
+
55
+
56
+ ### September 28 2025: WanGP v8.76 - ~~Here Are Two Three New Contenders in the Vace Arena !~~ The Never Ending Release
57
+
58
+ So in ~~today's~~ this release you will find two Wannabe Vace that covers each only a subset of Vace features but offers some interesting advantages:
59
+ - **Wan 2.2 Animate**: this model is specialized in *Body Motion* and *Facial Motion transfers*. It does that very well. You can use this model to either *Replace* a person in an in Video or *Animate* the person of your choice using an existing *Pose Video* (remember *Animate Anyone* ?). By default it will keep the original soundtrack. *Wan 2.2 Animate* seems to be under the hood a derived i2v model and should support the corresponding Loras Accelerators (for instance *FusioniX t2v*). Also as a WanGP exclusivity, you will find support for *Outpainting*.
60
+
61
+ In order to use Wan 2.2 Animate you will need first to stop by the *Mat Anyone* embedded tool, to extract the *Video Mask* of the person from which you want to extract the motion.
62
+
63
+ With version WanGP 8.74, there is an extra option that allows you to apply *Relighting* when Replacing a person. Also, you can now Animate a person without providing a Video Mask to target the source of the motion (with the risk it will be less precise)
64
+
65
+ For those of you who have a mask halo effect when Animating a character I recommend trying *SDPA attention* and to use the *FusioniX i2v* lora. If this issue persists (this will depend on the control video) you have now a choice of the two *Animate Mask Options* in *WanGP 8.76*. The old masking option which was a WanGP exclusive has been renamed *See Through Mask* because the background behind the animated character was preserved but this creates sometime visual artifacts. The new option which has the shorter name is what you may find elsewhere online. As it uses internally a much larger mask, there is no halo. However the immediate background behind the character is not preserved and may end completely different.
66
+
67
+ - **Lucy Edit**: this one claims to be a *Nano Banana* for Videos. Give it a video and asks it to change it (it is specialized in clothes changing) and voila ! The nice thing about it is that is it based on the *Wan 2.2 5B* model and therefore is very fast especially if you the *FastWan* finetune that is also part of the package.
68
+
69
+ Also because I wanted to spoil you:
70
+ - **Qwen Edit Plus**: also known as the *Qwen Edit 25th September Update* which is specialized in combining multiple Objects / People. There is also a new support for *Pose transfer* & *Recolorisation*. All of this made easy to use in WanGP. You will find right now only the quantized version since HF crashes when uploading the unquantized version.
71
+
72
+ - **T2V Video 2 Video Masking**: ever wanted to apply a Lora, a process (for instance Upsampling) or a Text Prompt on only a (moving) part of a Source Video. Look no further, I have added *Masked Video 2 Video* (which works also in image2image) in the *Text 2 Video* models. As usual you just need to use *Matanyone* to creatre the mask.
73
+
74
+
75
+ *Update 8.71*: fixed Fast Lucy Edit that didnt contain the lora
76
+ *Update 8.72*: shadow drop of Qwen Edit Plus
77
+ *Update 8.73*: Qwen Preview & InfiniteTalk Start image
78
+ *Update 8.74*: Animate Relighting / Nomask mode , t2v Masked Video to Video
79
+ *Update 8.75*: REDACTED
80
+ *Update 8.76*: Alternate Animate masking that fixes the mask halo effect that some users have
81
+
82
+ ### September 15 2025: WanGP v8.6 - Attack of the Clones
83
+
84
+ - The long awaited **Vace for Wan 2.2** is at last here or maybe not: it has been released by the *Fun Team* of *Alibaba* and it is not official. You can play with the vanilla version (**Vace Fun**) or with the one accelerated with Loras (**Vace Fan Cocktail**)
85
+
86
+ - **First Frame / Last Frame for Vace** : Vace models are so powerful that they could do *First frame / Last frame* since day one using the *Injected Frames* feature. However this required to compute by hand the locations of each end frame since this feature expects frames positions. I made it easier to compute these locations by using the "L" alias :
87
+
88
+ For a video Gen from scratch *"1 L L L"* means the 4 Injected Frames will be injected like this: frame no 1 at the first position, the next frame at the end of the first window, then the following frame at the end of the next window, and so on ....
89
+ If you *Continue a Video* , you just need *"L L L"* since the first frame is the last frame of the *Source Video*. In any case remember that numeral frames positions (like "1") are aligned by default to the beginning of the source window, so low values such as 1 will be considered in the past unless you change this behaviour in *Sliding Window Tab/ Control Video, Injected Frames aligment*.
90
+
91
+ - **Qwen Edit Inpainting** exists now in two versions: the original version of the previous release and a Lora based version. Each version has its pros and cons. For instance the Lora version supports also **Outpainting** ! However it tends to change slightly the original image even outside the outpainted area.
92
+
93
+ - **Better Lipsync with all the Audio to Video models**: you probably noticed that *Multitalk*, *InfiniteTalk* or *Hunyuan Avatar* had so so lipsync when the audio provided contained some background music. The problem should be solved now thanks to an automated background music removal all done by IA. Don't worry you will still hear the music as it is added back in the generated Video.
94
+
95
+ ### September 11 2025: WanGP v8.5/8.55 - Wanna be a Cropper or a Painter ?
96
+
97
+ I have done some intensive internal refactoring of the generation pipeline to ease support of existing models or add new models. Nothing really visible but this makes WanGP is little more future proof.
98
+
99
+ Otherwise in the news:
100
+ - **Cropped Input Image Prompts**: as quite often most *Image Prompts* provided (*Start Image, Input Video, Reference Image, Control Video, ...*) rarely matched your requested *Output Resolution*. In that case I used the resolution you gave either as a *Pixels Budget* or as an *Outer Canvas* for the Generated Video. However in some occasion you really want the requested Output Resolution and nothing else. Besides some models deliver much better Generations if you stick to one of their supported resolutions. In order to address this need I have added a new Output Resolution choice in the *Configuration Tab*: **Dimensions Correspond to the Ouput Weight & Height as the Prompt Images will be Cropped to fit Exactly these dimensins**. In short if needed the *Input Prompt Images* will be cropped (centered cropped for the moment). You will see this can make quite a difference for some models
101
+
102
+ - *Qwen Edit* has now a new sub Tab called **Inpainting**, that lets you target with a brush which part of the *Image Prompt* you want to modify. This is quite convenient if you find that Qwen Edit modifies usually too many things. Of course, as there are more constraints for Qwen Edit don't be surprised if sometime it will return the original image unchanged. A piece of advise: describe in your *Text Prompt* where (for instance *left to the man*, *top*, ...) the parts that you want to modify are located.
103
+
104
+ The mask inpainting is fully compatible with *Matanyone Mask generator*: generate first an *Image Mask* with Matanyone, transfer it to the current Image Generator and modify the mask with the *Paint Brush*. Talking about matanyone I have fixed a bug that caused a mask degradation with long videos (now WanGP Matanyone is as good as the original app and still requires 3 times less VRAM)
105
+
106
+ - This **Inpainting Mask Editor** has been added also to *Vace Image Mode*. Vace is probably still one of best Image Editor today. Here is a very simple & efficient workflow that do marvels with Vace:
107
+ Select *Vace Cocktail > Control Image Process = Perform Inpainting & Area Processed = Masked Area > Upload a Control Image, then draw your mask directly on top of the image & enter a text Prompt that describes the expected change > Generate > Below the Video Gallery click 'To Control Image' > Keep on doing more changes*.
108
+
109
+ Doing more sophisticated thing Vace Image Editor works very well too: try Image Outpainting, Pose transfer, ...
110
+
111
+ For the best quality I recommend to set in *Quality Tab* the option: "*Generate a 9 Frames Long video...*"
112
+
113
+ **update 8.55**: Flux Festival
114
+ - **Inpainting Mode** also added for *Flux Kontext*
115
+ - **Flux SRPO** : new finetune with x3 better quality vs Flux Dev according to its authors. I have also created a *Flux SRPO USO* finetune which is certainly the best open source *Style Transfer* tool available
116
+ - **Flux UMO**: model specialized in combining multiple reference objects / people together. Works quite well at 768x768
117
+
118
+ Good luck with finding your way through all the Flux models names !
119
+
120
+ ### September 5 2025: WanGP v8.4 - Take me to Outer Space
121
+ You have probably seen these short AI generated movies created using *Nano Banana* and the *First Frame - Last Frame* feature of *Kling 2.0*. The idea is to generate an image, modify a part of it with Nano Banana and give the these two images to Kling that will generate the Video between these two images, use now the previous Last Frame as the new First Frame, rinse and repeat and you get a full movie.
122
+
123
+ I have made it easier to do just that with *Qwen Edit* and *Wan*:
124
+ - **End Frames can now be combined with Continue a Video** (and not just a Start Frame)
125
+ - **Multiple End Frames can be inputed**, each End Frame will be used for a different Sliding Window
126
+
127
+ You can plan in advance all your shots (one shot = one Sliding Window) : I recommend using Wan 2.2 Image to Image with multiple End Frames (one for each shot / Sliding Window), and a different Text Prompt for each shot / Sliding Winow (remember to enable *Sliding Windows/Text Prompts Will be used for a new Sliding Window of the same Video Generation*)
128
+
129
+ The results can quite be impressive. However, Wan 2.1 & 2.2 Image 2 Image are restricted to a single overlap frame when using Slide Windows, which means only one frame is reeused for the motion. This may be unsufficient if you are trying to connect two shots with fast movement.
130
+
131
+ This is where *InfinitTalk* comes into play. Beside being one best models to generate animated audio driven avatars, InfiniteTalk uses internally more one than motion frames. It is quite good to maintain the motions between two shots. I have tweaked InfinitTalk so that **its motion engine can be used even if no audio is provided**.
132
+ So here is how to use InfiniteTalk: enable *Sliding Windows/Text Prompts Will be used for a new Sliding Window of the same Video Generation*), and if you continue an existing Video *Misc/Override Frames per Second" should be set to "Source Video*. Each Reference Frame inputed will play the same role as the End Frame except it wont be exactly an End Frame (it will correspond more to a middle frame, the actual End Frame will differ but will be close)
133
+
134
+
135
+ You will find below a 33s movie I have created using these two methods. Quality could be much better as I havent tuned at all the settings (I couldn't bother, I used 10 steps generation without Loras Accelerators for most of the gens).
136
+
137
+ ### September 2 2025: WanGP v8.31 - At last the pain stops
138
+
139
+ - This single new feature should give you the strength to face all the potential bugs of this new release:
140
+ **Images Management (multiple additions or deletions, reordering) for Start Images / End Images / Images References.**
141
+
142
+ - Unofficial **Video to Video (Non Sparse this time) for InfinitTalk**. Use the Strength Noise slider to decide how much motion of the original window you want to keep. I have also *greatly reduced the VRAM requirements for Multitalk / Infinitalk* (especially the multispeakers version & when generating at 1080p).
143
+
144
+ - **Experimental Sage 3 Attention support**: you will need to deserve this one, first you need a Blackwell GPU (RTX50xx) and request an access to Sage 3 Github repo, then you will have to compile Sage 3, install it and cross your fingers ...
145
+
146
+
147
+ *update 8.31: one shouldnt talk about bugs if one doesn't want to attract bugs*
148
+
149
+
150
+ See full changelog: **[Changelog](docs/CHANGELOG.md)**
151
+
152
+ ## 📋 Table of Contents
153
+
154
+ - [🚀 Quick Start](#-quick-start)
155
+ - [📦 Installation](#-installation)
156
+ - [🎯 Usage](#-usage)
157
+ - [📚 Documentation](#-documentation)
158
+ - [🔗 Related Projects](#-related-projects)
159
+
160
+ ## 🚀 Quick Start
161
+
162
+ **One-click installation:**
163
+ - Get started instantly with [Pinokio App](https://pinokio.computer/)
164
+ - Use Redtash1 [One Click Install with Sage](https://github.com/Redtash1/Wan2GP-Windows-One-Click-Install-With-Sage)
165
+
166
+ **Manual installation:**
167
+ ```bash
168
+ git clone https://github.com/deepbeepmeep/Wan2GP.git
169
+ cd Wan2GP
170
+ conda create -n wan2gp python=3.10.9
171
+ conda activate wan2gp
172
+ pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
173
+ pip install -r requirements.txt
174
+ ```
175
+
176
+ **Run the application:**
177
+ ```bash
178
+ python wgp.py
179
+ ```
180
+
181
+ **Update the application:**
182
+ If using Pinokio use Pinokio to update otherwise:
183
+ Get in the directory where WanGP is installed and:
184
+ ```bash
185
+ git pull
186
+ pip install -r requirements.txt
187
+ ```
188
+
189
+ ## 🐳 Docker:
190
+
191
+ **For Debian-based systems (Ubuntu, Debian, etc.):**
192
+
193
+ ```bash
194
+ ./run-docker-cuda-deb.sh
195
+ ```
196
+
197
+ This automated script will:
198
+
199
+ - Detect your GPU model and VRAM automatically
200
+ - Select optimal CUDA architecture for your GPU
201
+ - Install NVIDIA Docker runtime if needed
202
+ - Build a Docker image with all dependencies
203
+ - Run WanGP with optimal settings for your hardware
204
+
205
+ **Docker environment includes:**
206
+
207
+ - NVIDIA CUDA 12.4.1 with cuDNN support
208
+ - PyTorch 2.6.0 with CUDA 12.4 support
209
+ - SageAttention compiled for your specific GPU architecture
210
+ - Optimized environment variables for performance (TF32, threading, etc.)
211
+ - Automatic cache directory mounting for faster subsequent runs
212
+ - Current directory mounted in container - all downloaded models, loras, generated videos and files are saved locally
213
+
214
+ **Supported GPUs:** RTX 40XX, RTX 30XX, RTX 20XX, GTX 16XX, GTX 10XX, Tesla V100, A100, H100, and more.
215
+
216
+ ## 📦 Installation
217
+
218
+ For detailed installation instructions for different GPU generations:
219
+ - **[Installation Guide](docs/INSTALLATION.md)** - Complete setup instructions for RTX 10XX to RTX 50XX
220
+
221
+ ## 🎯 Usage
222
+
223
+ ### Basic Usage
224
+ - **[Getting Started Guide](docs/GETTING_STARTED.md)** - First steps and basic usage
225
+ - **[Models Overview](docs/MODELS.md)** - Available models and their capabilities
226
+
227
+ ### Advanced Features
228
+ - **[Loras Guide](docs/LORAS.md)** - Using and managing Loras for customization
229
+ - **[Finetunes](docs/FINETUNES.md)** - Add manually new models to WanGP
230
+ - **[VACE ControlNet](docs/VACE.md)** - Advanced video control and manipulation
231
+ - **[Command Line Reference](docs/CLI.md)** - All available command line options
232
+
233
+ ## 📚 Documentation
234
+
235
+ - **[Changelog](docs/CHANGELOG.md)** - Latest updates and version history
236
+ - **[Troubleshooting](docs/TROUBLESHOOTING.md)** - Common issues and solutions
237
+
238
+ ## 📚 Video Guides
239
+ - Nice Video that explain how to use Vace:\
240
+ https://www.youtube.com/watch?v=FMo9oN2EAvE
241
+ - Another Vace guide:\
242
+ https://www.youtube.com/watch?v=T5jNiEhf9xk
243
+
244
+ ## 🔗 Related Projects
245
+
246
+ ### Other Models for the GPU Poor
247
+ - **[HuanyuanVideoGP](https://github.com/deepbeepmeep/HunyuanVideoGP)** - One of the best open source Text to Video generators
248
+ - **[Hunyuan3D-2GP](https://github.com/deepbeepmeep/Hunyuan3D-2GP)** - Image to 3D and text to 3D tool
249
+ - **[FluxFillGP](https://github.com/deepbeepmeep/FluxFillGP)** - Inpainting/outpainting tools based on Flux
250
+ - **[Cosmos1GP](https://github.com/deepbeepmeep/Cosmos1GP)** - Text to world generator and image/video to world
251
+ - **[OminiControlGP](https://github.com/deepbeepmeep/OminiControlGP)** - Flux-derived application for object transfer
252
+ - **[YuE GP](https://github.com/deepbeepmeep/YuEGP)** - Song generator with instruments and singer's voice
253
+
254
+ ---
255
+
256
+ <p align="center">
257
+ Made with ❤️ by DeepBeepMeep
258
+ </p>
Wan2GP/.gitignore ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ .*
2
+ *.py[cod]
3
+ # *.jpg
4
+ *.jpeg
5
+ # *.png
6
+ *.gif
7
+ *.bmp
8
+ *.mp4
9
+ *.mov
10
+ *.mkv
11
+ *.log
12
+ *.zip
13
+ *.pt
14
+ *.pth
15
+ *.ckpt
16
+ *.safetensors
17
+ #*.json
18
+ # *.txt
19
+ *.backup
20
+ *.pkl
21
+ *.html
22
+ *.pdf
23
+ *.whl
24
+ *.exe
25
+ cache
26
+ __pycache__/
27
+ storage/
28
+ samples/
29
+ !.gitignore
30
+ !requirements.txt
31
+ .DS_Store
32
+ *DS_Store
33
+ google/
34
+ Wan2.1-T2V-14B/
35
+ Wan2.1-T2V-1.3B/
36
+ Wan2.1-I2V-14B-480P/
37
+ Wan2.1-I2V-14B-720P/
38
+ outputs/
39
+ outputs2/
40
+ gradio_outputs/
41
+ ckpts/
42
+ loras/
43
+ loras_i2v/
44
+
45
+ settings/
46
+
47
+ wgp_config.json
Wan2GP/Custom Resolutions Instructions.txt ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ You can override the choice of Resolutions offered by WanGP, if you create a file "resolutions.json" in the main WanGP folder.
2
+ This file is composed of a list of 2 elements sublists. Each 2 elements sublist should have the format ["Label", "WxH"] where W, H are respectively the Width and Height of the resolution. Please make sure that W and H are multiples of 16. The letter "x" should be placed inbetween these two dimensions.
3
+
4
+ Here is below a sample "resolutions.json" file :
5
+
6
+ [
7
+ ["1280x720 (16:9, 720p)", "1280x720"],
8
+ ["720x1280 (9:16, 720p)", "720x1280"],
9
+ ["1024x1024 (1:1, 720p)", "1024x1024"],
10
+ ["1280x544 (21:9, 720p)", "1280x544"],
11
+ ["544x1280 (9:21, 720p)", "544x1280"],
12
+ ["1104x832 (4:3, 720p)", "1104x832"],
13
+ ["832x1104 (3:4, 720p)", "832x1104"],
14
+ ["960x960 (1:1, 720p)", "960x960"],
15
+ ["832x480 (16:9, 480p)", "832x480"]
16
+ ]
Wan2GP/Dockerfile ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ FROM nvidia/cuda:12.4.1-cudnn-devel-ubuntu22.04
2
+
3
+ # Build arg for GPU architectures - specify which CUDA compute capabilities to compile for
4
+ # Common values:
5
+ # 7.0 - Tesla V100
6
+ # 7.5 - RTX 2060, 2070, 2080, Titan RTX
7
+ # 8.0 - A100, A800 (Ampere data center)
8
+ # 8.6 - RTX 3060, 3070, 3080, 3090 (Ampere consumer)
9
+ # 8.9 - RTX 4070, 4080, 4090 (Ada Lovelace)
10
+ # 9.0 - H100, H800 (Hopper data center)
11
+ # 12.0 - RTX 5070, 5080, 5090 (Blackwell) - Note: sm_120 architecture
12
+ #
13
+ # Examples:
14
+ # RTX 3060: --build-arg CUDA_ARCHITECTURES="8.6"
15
+ # RTX 4090: --build-arg CUDA_ARCHITECTURES="8.9"
16
+ # Multiple: --build-arg CUDA_ARCHITECTURES="8.0;8.6;8.9"
17
+ #
18
+ # Note: Including 8.9 or 9.0 may cause compilation issues on some setups
19
+ # Default includes 8.0 and 8.6 for broad Ampere compatibility
20
+ ARG CUDA_ARCHITECTURES="8.0;8.6"
21
+
22
+ ENV DEBIAN_FRONTEND=noninteractive
23
+
24
+ # Install system dependencies
25
+ RUN apt update && \
26
+ apt install -y \
27
+ python3 python3-pip git wget curl cmake ninja-build \
28
+ libgl1 libglib2.0-0 ffmpeg && \
29
+ apt clean
30
+
31
+ WORKDIR /workspace
32
+
33
+ COPY requirements.txt .
34
+
35
+ # Upgrade pip first
36
+ RUN pip install --upgrade pip setuptools wheel
37
+
38
+ # Install requirements if exists
39
+ RUN pip install -r requirements.txt
40
+
41
+ # Install PyTorch with CUDA support
42
+ RUN pip install --extra-index-url https://download.pytorch.org/whl/cu124 \
43
+ torch==2.6.0+cu124 torchvision==0.21.0+cu124
44
+
45
+ # Install SageAttention from git (patch GPU detection)
46
+ ENV TORCH_CUDA_ARCH_LIST="${CUDA_ARCHITECTURES}"
47
+ ENV FORCE_CUDA="1"
48
+ ENV MAX_JOBS="1"
49
+
50
+ COPY <<EOF /tmp/patch_setup.py
51
+ import os
52
+ with open('setup.py', 'r') as f:
53
+ content = f.read()
54
+
55
+ # Get architectures from environment variable
56
+ arch_list = os.environ.get('TORCH_CUDA_ARCH_LIST')
57
+ arch_set = '{' + ', '.join([f'"{arch}"' for arch in arch_list.split(';')]) + '}'
58
+
59
+ # Replace the GPU detection section
60
+ old_section = '''compute_capabilities = set()
61
+ device_count = torch.cuda.device_count()
62
+ for i in range(device_count):
63
+ major, minor = torch.cuda.get_device_capability(i)
64
+ if major < 8:
65
+ warnings.warn(f"skipping GPU {i} with compute capability {major}.{minor}")
66
+ continue
67
+ compute_capabilities.add(f"{major}.{minor}")'''
68
+
69
+ new_section = 'compute_capabilities = ' + arch_set + '''
70
+ print(f"Manually set compute capabilities: {compute_capabilities}")'''
71
+
72
+ content = content.replace(old_section, new_section)
73
+
74
+ with open('setup.py', 'w') as f:
75
+ f.write(content)
76
+ EOF
77
+
78
+ RUN git clone https://github.com/thu-ml/SageAttention.git /tmp/sageattention && \
79
+ cd /tmp/sageattention && \
80
+ python3 /tmp/patch_setup.py && \
81
+ pip install --no-build-isolation .
82
+
83
+ RUN useradd -u 1000 -ms /bin/bash user
84
+
85
+ RUN chown -R user:user /workspace
86
+
87
+ RUN mkdir /home/user/.cache && \
88
+ chown -R user:user /home/user/.cache
89
+
90
+ COPY entrypoint.sh /workspace/entrypoint.sh
91
+
92
+ ENTRYPOINT ["/workspace/entrypoint.sh"]
Wan2GP/LICENSE.txt ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ WanGP NON-COMMERCIAL EVALUATION LICENSE 1.0
2
+
3
+ Definitions
4
+ 1.1 “Software” means the source code, binaries, libraries, utilities and UI released under this license.
5
+ 1.2 “Output” means images, videos or other media produced by running the Software.
6
+ 1.3 “Commercial Use” means:
7
+ a) selling, sublicensing, renting, leasing, or otherwise distributing the Software, in whole or in part, for a fee or other consideration; or
8
+ b) offering the Software (or any derivative) as part of a paid product or hosted service; or
9
+ c) using the Software (or any derivative) to provide cloud-based or backend services, where end users access or pay for those services.
10
+
11
+ License Grant
12
+ Subject to Section 3:
13
+ a) You are granted a worldwide, non-exclusive, royalty-free, revocable license to use, reproduce, modify and distribute the Software for non-commercial purposes only.
14
+ b) You are granted a worldwide, non-exclusive, royalty-free, irrevocable license to use, reproduce, modify and distribute the Output for any purpose, including commercial sale, provided that any commercial distribution of the Output includes a clear notice that the Output was produced (in whole or in part) using WanGP, along with a hyperlink to the WanGP application’s About tab or repository.
15
+
16
+ Restrictions
17
+ 3.1 You MAY NOT distribute, sublicense or otherwise make available the Software (or any derivative) for Commercial Use.
18
+ 3.2 You MAY sell, license or otherwise commercially exploit the Output without restriction.
19
+ 3.3 If you wish to use the Software for Commercial Use, you must obtain a separate commercial license from the Licensor.
20
+
21
+ Third-Party Components 4.1 The Software includes components licensed under various open-source licenses (e.g., Apache 2.0, MIT, BSD). 4.2 You must comply with all applicable terms of those third-party licenses, including preservation of copyright notices, inclusion of required license texts, and patent-grant provisions. 4.3 You can find the full text of each third-party license via the “About” tab in the WanGP application, which provides links to their original GitHub repositories.
22
+
23
+ Attribution
24
+ 5.1 You must give appropriate credit by including:
25
+ • a copy of this license (or a link to it), and
26
+ • a notice that your use is based on “WanGP”.
27
+ 5.2 You may do so in any reasonable manner, but not in any way that suggests the Licensor endorses you or your use.
28
+
29
+ Disclaimer of Warranty & Liability
30
+ THE SOFTWARE IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED.
31
+ IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE.
32
+
33
+ Commercial Licensing The Licensor may offer commercial licenses for the Software, which grant rights to use the Software for Commercial Use. Please contact [[email protected]] for terms and pricing.
34
+
35
+ Effective Date & Previous Versions
36
+ 8.1 This license is effective as of the date the LICENSE file is updated in the WanGP repository.
37
+ 8.2 Any copies of the Software obtained under prior license terms before this Effective Date remain governed by those prior terms; such granted rights are irrevocable.
38
+ 8.3 Use of the Software after the release of any subsequent version by the Licensor is subject to the terms of the then-current license, unless a separate agreement is in place.
39
+
40
+ Acceptable Use / Moral Clause
41
+ 9.1 You MAY NOT use the Software or the Output to facilitate or produce content that is illegal, harmful, violent, harassing, defamatory, fraudulent, or otherwise violates applicable laws or fundamental human rights.
42
+ 9.2 You MAY NOT deploy the Software or Output in contexts that promote hate speech, extremist ideology, human rights abuses, or other actions that could foreseeably cause significant harm to individuals or groups.
43
+ 9.3 The Licensor reserves the right to terminate the rights granted under this license if a licensee materially breaches this Acceptable Use clause.
44
+
45
+ END OF LICENSE
46
+
Wan2GP/README.md ADDED
@@ -0,0 +1,252 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # WanGP
2
+
3
+ -----
4
+ <p align="center">
5
+ <b>WanGP by DeepBeepMeep : The best Open Source Video Generative Models Accessible to the GPU Poor</b>
6
+ </p>
7
+
8
+ WanGP supports the Wan (and derived models), Hunyuan Video and LTV Video models with:
9
+ - Low VRAM requirements (as low as 6 GB of VRAM is sufficient for certain models)
10
+ - Support for old GPUs (RTX 10XX, 20xx, ...)
11
+ - Very Fast on the latest GPUs
12
+ - Easy to use Full Web based interface
13
+ - Auto download of the required model adapted to your specific architecture
14
+ - Tools integrated to facilitate Video Generation : Mask Editor, Prompt Enhancer, Temporal and Spatial Generation, MMAudio, Video Browser, Pose / Depth / Flow extractor
15
+ - Loras Support to customize each model
16
+ - Queuing system : make your shopping list of videos to generate and come back later
17
+
18
+ **Discord Server to get Help from Other Users and show your Best Videos:** https://discord.gg/g7efUW9jGV
19
+
20
+ **Follow DeepBeepMeep on Twitter/X to get the Latest News**: https://x.com/deepbeepmeep
21
+
22
+ ## 🔥 Latest Updates :
23
+ ### October 3 2025: WanGP v8.992 - One Last Thing before the Big Unknown ...
24
+
25
+ This new version hasn't any new model...
26
+
27
+ ...but temptation to upgrade will be high as it contains a few Loras related features that may change your Life:
28
+ - **Ready to use Loras Accelerators Profiles** per type of model that you can apply on your current *Generation Settings*. Next time I will recommend a *Lora Accelerator*, it will be only one click away. And best of all of the required Loras will be downloaded automatically. When you apply an *Accelerator Profile*, input fields like the *Number of Denoising Steps* *Activated Loras*, *Loras Multipliers* (such as "1;0 0;1" ...) will be automatically filled. However your video specific fields will be preserved, so it will be easy to switch between Profiles to experiment.
29
+
30
+ - **Embedded Loras URL** : WanGP will now try to remember every Lora URLs it sees. For instance if someone sends you some settings that contain Loras URLs or you extract the Settings of Video generated by a friend with Loras URLs, these URLs will be automatically added to *WanGP URL Cache*. Conversely everything you will share (Videos, Settings, Lset files) will contain the download URLs if they are known. You can also download directly a Lora in WanGP by using the *Download Lora* button a the bottom. The Lora will be immediatly available and added to WanGP lora URL cache. This will work with *Hugging Face* as a repository. Support for CivitAi will come as soon as someone will nice enough to post a GitHub PR ...
31
+
32
+ - **.lset file** supports embedded Loras URLs. It has never been easier to share a Lora with a friend. As a reminder a .lset file can be created directly from *WanGP Web Interface* and it contains a list of Loras and their multipliers, a Prompt and Instructions how to use these loras (like the Lora's *Trigger*). So with embedded Loras URL, you can send an .lset file by email or share it on discord: it is just a 1 KB tiny text, but with it other people will be able to use Gigabytes Loras as these will be automatically downloaded.
33
+
34
+ I have created the new Discord Channel **share-your-settings** where you can post your *Settings* or *Lset files*. I will be pleased to add new Loras Accelerators in the list of WanGP *Accelerators Profiles if you post some good ones there.
35
+
36
+ Last but not least the Lora's documentation has been updated.
37
+
38
+ *update 8.991*: full power of *Vace Lynx* unleashed with new combinations such as Landscape + Face / Clothes + Face / Injectd Frame (Start/End frames/...) + Face
39
+ *update 8.992*: optimized gen with Lora, should be 10% if many loras
40
+
41
+ ### September 30 2025: WanGP v8.9 - Combinatorics
42
+
43
+ This new version of WanGP introduces **Wan 2.1 Lynx** the best Control Net so far to transfer *Facial Identity*. You will be amazed to recognize your friends even with a completely different hair style. Congrats to the *Byte Dance team* for this achievement. Lynx works quite with well *Fusionix t2v* 10 steps.
44
+
45
+ *WanGP 8.9* also illustrate how existing WanGP features can be easily combined with new models. For instance with *Lynx* you will get out of the box *Video to Video* and *Image/Text to Image*.
46
+
47
+ Another fun combination is *Vace* + *Lynx*, which works much better than *Vace StandIn*. I have added sliders to change the weight of Vace & Lynx to allow you to tune the effects.
48
+
49
+
50
+ ### September 28 2025: WanGP v8.76 - ~~Here Are Two Three New Contenders in the Vace Arena !~~ The Never Ending Release
51
+
52
+ So in ~~today's~~ this release you will find two Wannabe Vace that covers each only a subset of Vace features but offers some interesting advantages:
53
+ - **Wan 2.2 Animate**: this model is specialized in *Body Motion* and *Facial Motion transfers*. It does that very well. You can use this model to either *Replace* a person in an in Video or *Animate* the person of your choice using an existing *Pose Video* (remember *Animate Anyone* ?). By default it will keep the original soundtrack. *Wan 2.2 Animate* seems to be under the hood a derived i2v model and should support the corresponding Loras Accelerators (for instance *FusioniX t2v*). Also as a WanGP exclusivity, you will find support for *Outpainting*.
54
+
55
+ In order to use Wan 2.2 Animate you will need first to stop by the *Mat Anyone* embedded tool, to extract the *Video Mask* of the person from which you want to extract the motion.
56
+
57
+ With version WanGP 8.74, there is an extra option that allows you to apply *Relighting* when Replacing a person. Also, you can now Animate a person without providing a Video Mask to target the source of the motion (with the risk it will be less precise)
58
+
59
+ For those of you who have a mask halo effect when Animating a character I recommend trying *SDPA attention* and to use the *FusioniX i2v* lora. If this issue persists (this will depend on the control video) you have now a choice of the two *Animate Mask Options* in *WanGP 8.76*. The old masking option which was a WanGP exclusive has been renamed *See Through Mask* because the background behind the animated character was preserved but this creates sometime visual artifacts. The new option which has the shorter name is what you may find elsewhere online. As it uses internally a much larger mask, there is no halo. However the immediate background behind the character is not preserved and may end completely different.
60
+
61
+ - **Lucy Edit**: this one claims to be a *Nano Banana* for Videos. Give it a video and asks it to change it (it is specialized in clothes changing) and voila ! The nice thing about it is that is it based on the *Wan 2.2 5B* model and therefore is very fast especially if you the *FastWan* finetune that is also part of the package.
62
+
63
+ Also because I wanted to spoil you:
64
+ - **Qwen Edit Plus**: also known as the *Qwen Edit 25th September Update* which is specialized in combining multiple Objects / People. There is also a new support for *Pose transfer* & *Recolorisation*. All of this made easy to use in WanGP. You will find right now only the quantized version since HF crashes when uploading the unquantized version.
65
+
66
+ - **T2V Video 2 Video Masking**: ever wanted to apply a Lora, a process (for instance Upsampling) or a Text Prompt on only a (moving) part of a Source Video. Look no further, I have added *Masked Video 2 Video* (which works also in image2image) in the *Text 2 Video* models. As usual you just need to use *Matanyone* to creatre the mask.
67
+
68
+
69
+ *Update 8.71*: fixed Fast Lucy Edit that didnt contain the lora
70
+ *Update 8.72*: shadow drop of Qwen Edit Plus
71
+ *Update 8.73*: Qwen Preview & InfiniteTalk Start image
72
+ *Update 8.74*: Animate Relighting / Nomask mode , t2v Masked Video to Video
73
+ *Update 8.75*: REDACTED
74
+ *Update 8.76*: Alternate Animate masking that fixes the mask halo effect that some users have
75
+
76
+ ### September 15 2025: WanGP v8.6 - Attack of the Clones
77
+
78
+ - The long awaited **Vace for Wan 2.2** is at last here or maybe not: it has been released by the *Fun Team* of *Alibaba* and it is not official. You can play with the vanilla version (**Vace Fun**) or with the one accelerated with Loras (**Vace Fan Cocktail**)
79
+
80
+ - **First Frame / Last Frame for Vace** : Vace models are so powerful that they could do *First frame / Last frame* since day one using the *Injected Frames* feature. However this required to compute by hand the locations of each end frame since this feature expects frames positions. I made it easier to compute these locations by using the "L" alias :
81
+
82
+ For a video Gen from scratch *"1 L L L"* means the 4 Injected Frames will be injected like this: frame no 1 at the first position, the next frame at the end of the first window, then the following frame at the end of the next window, and so on ....
83
+ If you *Continue a Video* , you just need *"L L L"* since the first frame is the last frame of the *Source Video*. In any case remember that numeral frames positions (like "1") are aligned by default to the beginning of the source window, so low values such as 1 will be considered in the past unless you change this behaviour in *Sliding Window Tab/ Control Video, Injected Frames aligment*.
84
+
85
+ - **Qwen Edit Inpainting** exists now in two versions: the original version of the previous release and a Lora based version. Each version has its pros and cons. For instance the Lora version supports also **Outpainting** ! However it tends to change slightly the original image even outside the outpainted area.
86
+
87
+ - **Better Lipsync with all the Audio to Video models**: you probably noticed that *Multitalk*, *InfiniteTalk* or *Hunyuan Avatar* had so so lipsync when the audio provided contained some background music. The problem should be solved now thanks to an automated background music removal all done by IA. Don't worry you will still hear the music as it is added back in the generated Video.
88
+
89
+ ### September 11 2025: WanGP v8.5/8.55 - Wanna be a Cropper or a Painter ?
90
+
91
+ I have done some intensive internal refactoring of the generation pipeline to ease support of existing models or add new models. Nothing really visible but this makes WanGP is little more future proof.
92
+
93
+ Otherwise in the news:
94
+ - **Cropped Input Image Prompts**: as quite often most *Image Prompts* provided (*Start Image, Input Video, Reference Image, Control Video, ...*) rarely matched your requested *Output Resolution*. In that case I used the resolution you gave either as a *Pixels Budget* or as an *Outer Canvas* for the Generated Video. However in some occasion you really want the requested Output Resolution and nothing else. Besides some models deliver much better Generations if you stick to one of their supported resolutions. In order to address this need I have added a new Output Resolution choice in the *Configuration Tab*: **Dimensions Correspond to the Ouput Weight & Height as the Prompt Images will be Cropped to fit Exactly these dimensins**. In short if needed the *Input Prompt Images* will be cropped (centered cropped for the moment). You will see this can make quite a difference for some models
95
+
96
+ - *Qwen Edit* has now a new sub Tab called **Inpainting**, that lets you target with a brush which part of the *Image Prompt* you want to modify. This is quite convenient if you find that Qwen Edit modifies usually too many things. Of course, as there are more constraints for Qwen Edit don't be surprised if sometime it will return the original image unchanged. A piece of advise: describe in your *Text Prompt* where (for instance *left to the man*, *top*, ...) the parts that you want to modify are located.
97
+
98
+ The mask inpainting is fully compatible with *Matanyone Mask generator*: generate first an *Image Mask* with Matanyone, transfer it to the current Image Generator and modify the mask with the *Paint Brush*. Talking about matanyone I have fixed a bug that caused a mask degradation with long videos (now WanGP Matanyone is as good as the original app and still requires 3 times less VRAM)
99
+
100
+ - This **Inpainting Mask Editor** has been added also to *Vace Image Mode*. Vace is probably still one of best Image Editor today. Here is a very simple & efficient workflow that do marvels with Vace:
101
+ Select *Vace Cocktail > Control Image Process = Perform Inpainting & Area Processed = Masked Area > Upload a Control Image, then draw your mask directly on top of the image & enter a text Prompt that describes the expected change > Generate > Below the Video Gallery click 'To Control Image' > Keep on doing more changes*.
102
+
103
+ Doing more sophisticated thing Vace Image Editor works very well too: try Image Outpainting, Pose transfer, ...
104
+
105
+ For the best quality I recommend to set in *Quality Tab* the option: "*Generate a 9 Frames Long video...*"
106
+
107
+ **update 8.55**: Flux Festival
108
+ - **Inpainting Mode** also added for *Flux Kontext*
109
+ - **Flux SRPO** : new finetune with x3 better quality vs Flux Dev according to its authors. I have also created a *Flux SRPO USO* finetune which is certainly the best open source *Style Transfer* tool available
110
+ - **Flux UMO**: model specialized in combining multiple reference objects / people together. Works quite well at 768x768
111
+
112
+ Good luck with finding your way through all the Flux models names !
113
+
114
+ ### September 5 2025: WanGP v8.4 - Take me to Outer Space
115
+ You have probably seen these short AI generated movies created using *Nano Banana* and the *First Frame - Last Frame* feature of *Kling 2.0*. The idea is to generate an image, modify a part of it with Nano Banana and give the these two images to Kling that will generate the Video between these two images, use now the previous Last Frame as the new First Frame, rinse and repeat and you get a full movie.
116
+
117
+ I have made it easier to do just that with *Qwen Edit* and *Wan*:
118
+ - **End Frames can now be combined with Continue a Video** (and not just a Start Frame)
119
+ - **Multiple End Frames can be inputed**, each End Frame will be used for a different Sliding Window
120
+
121
+ You can plan in advance all your shots (one shot = one Sliding Window) : I recommend using Wan 2.2 Image to Image with multiple End Frames (one for each shot / Sliding Window), and a different Text Prompt for each shot / Sliding Winow (remember to enable *Sliding Windows/Text Prompts Will be used for a new Sliding Window of the same Video Generation*)
122
+
123
+ The results can quite be impressive. However, Wan 2.1 & 2.2 Image 2 Image are restricted to a single overlap frame when using Slide Windows, which means only one frame is reeused for the motion. This may be unsufficient if you are trying to connect two shots with fast movement.
124
+
125
+ This is where *InfinitTalk* comes into play. Beside being one best models to generate animated audio driven avatars, InfiniteTalk uses internally more one than motion frames. It is quite good to maintain the motions between two shots. I have tweaked InfinitTalk so that **its motion engine can be used even if no audio is provided**.
126
+ So here is how to use InfiniteTalk: enable *Sliding Windows/Text Prompts Will be used for a new Sliding Window of the same Video Generation*), and if you continue an existing Video *Misc/Override Frames per Second" should be set to "Source Video*. Each Reference Frame inputed will play the same role as the End Frame except it wont be exactly an End Frame (it will correspond more to a middle frame, the actual End Frame will differ but will be close)
127
+
128
+
129
+ You will find below a 33s movie I have created using these two methods. Quality could be much better as I havent tuned at all the settings (I couldn't bother, I used 10 steps generation without Loras Accelerators for most of the gens).
130
+
131
+ ### September 2 2025: WanGP v8.31 - At last the pain stops
132
+
133
+ - This single new feature should give you the strength to face all the potential bugs of this new release:
134
+ **Images Management (multiple additions or deletions, reordering) for Start Images / End Images / Images References.**
135
+
136
+ - Unofficial **Video to Video (Non Sparse this time) for InfinitTalk**. Use the Strength Noise slider to decide how much motion of the original window you want to keep. I have also *greatly reduced the VRAM requirements for Multitalk / Infinitalk* (especially the multispeakers version & when generating at 1080p).
137
+
138
+ - **Experimental Sage 3 Attention support**: you will need to deserve this one, first you need a Blackwell GPU (RTX50xx) and request an access to Sage 3 Github repo, then you will have to compile Sage 3, install it and cross your fingers ...
139
+
140
+
141
+ *update 8.31: one shouldnt talk about bugs if one doesn't want to attract bugs*
142
+
143
+
144
+ See full changelog: **[Changelog](docs/CHANGELOG.md)**
145
+
146
+ ## 📋 Table of Contents
147
+
148
+ - [🚀 Quick Start](#-quick-start)
149
+ - [📦 Installation](#-installation)
150
+ - [🎯 Usage](#-usage)
151
+ - [📚 Documentation](#-documentation)
152
+ - [🔗 Related Projects](#-related-projects)
153
+
154
+ ## 🚀 Quick Start
155
+
156
+ **One-click installation:**
157
+ - Get started instantly with [Pinokio App](https://pinokio.computer/)
158
+ - Use Redtash1 [One Click Install with Sage](https://github.com/Redtash1/Wan2GP-Windows-One-Click-Install-With-Sage)
159
+
160
+ **Manual installation:**
161
+ ```bash
162
+ git clone https://github.com/deepbeepmeep/Wan2GP.git
163
+ cd Wan2GP
164
+ conda create -n wan2gp python=3.10.9
165
+ conda activate wan2gp
166
+ pip install torch==2.7.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/test/cu128
167
+ pip install -r requirements.txt
168
+ ```
169
+
170
+ **Run the application:**
171
+ ```bash
172
+ python wgp.py
173
+ ```
174
+
175
+ **Update the application:**
176
+ If using Pinokio use Pinokio to update otherwise:
177
+ Get in the directory where WanGP is installed and:
178
+ ```bash
179
+ git pull
180
+ pip install -r requirements.txt
181
+ ```
182
+
183
+ ## 🐳 Docker:
184
+
185
+ **For Debian-based systems (Ubuntu, Debian, etc.):**
186
+
187
+ ```bash
188
+ ./run-docker-cuda-deb.sh
189
+ ```
190
+
191
+ This automated script will:
192
+
193
+ - Detect your GPU model and VRAM automatically
194
+ - Select optimal CUDA architecture for your GPU
195
+ - Install NVIDIA Docker runtime if needed
196
+ - Build a Docker image with all dependencies
197
+ - Run WanGP with optimal settings for your hardware
198
+
199
+ **Docker environment includes:**
200
+
201
+ - NVIDIA CUDA 12.4.1 with cuDNN support
202
+ - PyTorch 2.6.0 with CUDA 12.4 support
203
+ - SageAttention compiled for your specific GPU architecture
204
+ - Optimized environment variables for performance (TF32, threading, etc.)
205
+ - Automatic cache directory mounting for faster subsequent runs
206
+ - Current directory mounted in container - all downloaded models, loras, generated videos and files are saved locally
207
+
208
+ **Supported GPUs:** RTX 40XX, RTX 30XX, RTX 20XX, GTX 16XX, GTX 10XX, Tesla V100, A100, H100, and more.
209
+
210
+ ## 📦 Installation
211
+
212
+ For detailed installation instructions for different GPU generations:
213
+ - **[Installation Guide](docs/INSTALLATION.md)** - Complete setup instructions for RTX 10XX to RTX 50XX
214
+
215
+ ## 🎯 Usage
216
+
217
+ ### Basic Usage
218
+ - **[Getting Started Guide](docs/GETTING_STARTED.md)** - First steps and basic usage
219
+ - **[Models Overview](docs/MODELS.md)** - Available models and their capabilities
220
+
221
+ ### Advanced Features
222
+ - **[Loras Guide](docs/LORAS.md)** - Using and managing Loras for customization
223
+ - **[Finetunes](docs/FINETUNES.md)** - Add manually new models to WanGP
224
+ - **[VACE ControlNet](docs/VACE.md)** - Advanced video control and manipulation
225
+ - **[Command Line Reference](docs/CLI.md)** - All available command line options
226
+
227
+ ## 📚 Documentation
228
+
229
+ - **[Changelog](docs/CHANGELOG.md)** - Latest updates and version history
230
+ - **[Troubleshooting](docs/TROUBLESHOOTING.md)** - Common issues and solutions
231
+
232
+ ## 📚 Video Guides
233
+ - Nice Video that explain how to use Vace:\
234
+ https://www.youtube.com/watch?v=FMo9oN2EAvE
235
+ - Another Vace guide:\
236
+ https://www.youtube.com/watch?v=T5jNiEhf9xk
237
+
238
+ ## 🔗 Related Projects
239
+
240
+ ### Other Models for the GPU Poor
241
+ - **[HuanyuanVideoGP](https://github.com/deepbeepmeep/HunyuanVideoGP)** - One of the best open source Text to Video generators
242
+ - **[Hunyuan3D-2GP](https://github.com/deepbeepmeep/Hunyuan3D-2GP)** - Image to 3D and text to 3D tool
243
+ - **[FluxFillGP](https://github.com/deepbeepmeep/FluxFillGP)** - Inpainting/outpainting tools based on Flux
244
+ - **[Cosmos1GP](https://github.com/deepbeepmeep/Cosmos1GP)** - Text to world generator and image/video to world
245
+ - **[OminiControlGP](https://github.com/deepbeepmeep/OminiControlGP)** - Flux-derived application for object transfer
246
+ - **[YuE GP](https://github.com/deepbeepmeep/YuEGP)** - Song generator with instruments and singer's voice
247
+
248
+ ---
249
+
250
+ <p align="center">
251
+ Made with ❤️ by DeepBeepMeep
252
+ </p>
Wan2GP/configs/animate.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "motion_encoder_dim": 512
15
+ }
Wan2GP/configs/fantasy.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "fantasytalking_dim": 2048
15
+ }
Wan2GP/configs/flf2v_720p.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "flf": true
15
+ }
Wan2GP/configs/i2v.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/i2v_2_2.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.33.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v2_2",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/i2v_2_2_multitalk.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.33.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v2_2",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "multitalk_output_dim": 768
15
+ }
Wan2GP/configs/infinitetalk.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "multitalk_output_dim": 768
15
+ }
Wan2GP/configs/lucy_edit.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.33.0",
4
+ "dim": 3072,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 14336,
7
+ "freq_dim": 256,
8
+ "in_dim": 96,
9
+ "model_type": "ti2v2_2",
10
+ "num_heads": 24,
11
+ "num_layers": 30,
12
+ "out_dim": 48,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/lynx.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "lynx": "full"
15
+ }
Wan2GP/configs/multitalk.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 36,
9
+ "model_type": "i2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "multitalk_output_dim": 768
15
+ }
Wan2GP/configs/phantom_1.3B.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 1536,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 8960,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 12,
11
+ "num_layers": 30,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/phantom_14B.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/qwen_image_20B.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "QwenImageTransformer2DModel",
3
+ "_diffusers_version": "0.34.0.dev0",
4
+ "attention_head_dim": 128,
5
+ "axes_dims_rope": [
6
+ 16,
7
+ 56,
8
+ 56
9
+ ],
10
+ "guidance_embeds": false,
11
+ "in_channels": 64,
12
+ "joint_attention_dim": 3584,
13
+ "num_attention_heads": 24,
14
+ "num_layers": 60,
15
+ "out_channels": 16,
16
+ "patch_size": 2,
17
+ "pooled_projection_dim": 768
18
+ }
Wan2GP/configs/sky_df_1.3.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 1536,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 8960,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 12,
11
+ "num_layers": 30,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/sky_df_14B.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/standin.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "standin": true
15
+ }
Wan2GP/configs/t2v.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/t2v_1.3B.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 1536,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 8960,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 12,
11
+ "num_layers": 30,
12
+ "out_dim": 16,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/ti2v_2_2.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.33.0",
4
+ "dim": 3072,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 14336,
7
+ "freq_dim": 256,
8
+ "in_dim": 48,
9
+ "model_type": "ti2v2_2",
10
+ "num_heads": 24,
11
+ "num_layers": 30,
12
+ "out_dim": 48,
13
+ "text_len": 512
14
+ }
Wan2GP/configs/vace_1.3B.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "VaceWanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 1536,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 8960,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 12,
11
+ "num_layers": 30,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "vace_layers": [0, 2, 4, 6, 8, 10, 12, 14, 16, 18, 20, 22, 24, 26, 28],
15
+ "vace_in_dim": 96
16
+ }
Wan2GP/configs/vace_14B.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "VaceWanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "vace_layers": [0, 5, 10, 15, 20, 25, 30, 35],
15
+ "vace_in_dim": 96
16
+ }
Wan2GP/configs/vace_lynx_14B.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "WanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "vace_layers": [0, 5, 10, 15, 20, 25, 30, 35],
15
+ "vace_in_dim": 96,
16
+ "lynx": "full"
17
+ }
Wan2GP/configs/vace_multitalk_14B.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "VaceWanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "vace_layers": [0, 5, 10, 15, 20, 25, 30, 35],
15
+ "vace_in_dim": 96,
16
+ "multitalk_output_dim": 768
17
+ }
Wan2GP/configs/vace_standin_14B.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "VaceWanModel",
3
+ "_diffusers_version": "0.30.0",
4
+ "dim": 5120,
5
+ "eps": 1e-06,
6
+ "ffn_dim": 13824,
7
+ "freq_dim": 256,
8
+ "in_dim": 16,
9
+ "model_type": "t2v",
10
+ "num_heads": 40,
11
+ "num_layers": 40,
12
+ "out_dim": 16,
13
+ "text_len": 512,
14
+ "vace_layers": [0, 5, 10, 15, 20, 25, 30, 35],
15
+ "vace_in_dim": 96,
16
+ "standin": true
17
+ }
Wan2GP/defaults/ReadMe.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Please dot not modify any file in this Folder.
2
+
3
+ If you want to change a property of a default model, copy the corrresponding model file in the ./finetunes folder and modify the properties you want to change in the new file.
4
+ If a property is not in the new file, it will be inherited automatically from the default file that matches the same name file.
5
+
6
+ For instance to hide a model:
7
+
8
+ {
9
+ "model":
10
+ {
11
+ "visible": false
12
+ }
13
+ }
Wan2GP/defaults/animate.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Wan2.2 Animate",
4
+ "architecture": "animate",
5
+ "description": "Wan-Animate takes a video and a character image as input, and generates a video in either 'Animation' or 'Replacement' mode. Sliding Window of 81 frames at least are recommeded to obtain the best Style continuity.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_quanto_fp16_int8.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_14B_quanto_bf16_int8.safetensors"
10
+ ],
11
+ "preload_URLs" :
12
+ [
13
+ "https://huggingface.co/DeepBeepMeep/Wan2.2/resolve/main/wan2.2_animate_relighting_lora.safetensors"
14
+ ],
15
+ "settings_dir": [ "wan" ],
16
+ "group": "wan2_2"
17
+ }
18
+ }
Wan2GP/defaults/fantasy.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fantasy Talking 720p",
5
+ "architecture" : "fantasy",
6
+ "modules": [ ["https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_fantasy_speaking_14B_bf16.safetensors"]],
7
+ "description": "The Fantasy Talking model corresponds to the original Wan image 2 video model combined with the Fantasy Speaking module to process an audio Input.",
8
+ "URLs": "i2v_720p"
9
+ },
10
+ "resolution": "1280x720"
11
+ }
Wan2GP/defaults/flf2v_720p.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "First Last Frame to Video 720p (FLF2V) 14B",
5
+ "architecture" : "flf2v_720p",
6
+ "visible" : true,
7
+ "description": "The First Last Frame 2 Video model is the official model Image 2 Video model that supports Start and End frames.",
8
+ "URLs": [
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_mbf16.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_quanto_mbf16_int8.safetensors",
11
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_FLF2V_720p_14B_quanto_mfp16_int8.safetensors"
12
+ ],
13
+ "auto_quantize": true
14
+ },
15
+ "resolution": "1280x720"
16
+ }
Wan2GP/defaults/flux.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Dev is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true,
11
+ "flux-model": "flux-dev"
12
+ },
13
+ "prompt": "draw a hat",
14
+ "resolution": "1280x720",
15
+ "batch_size": 1
16
+ }
Wan2GP/defaults/flux_chroma.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Chroma 1 HD 8.9B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Chroma is a 8.9 billion parameters model. As a base model, Chroma1 is intentionally designed to be an excellent starting point for finetuning. It provides a strong, neutral foundation for developers, researchers, and artists to create specialized models..",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-chroma_hd_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-chroma_hd_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true,
11
+ "flux-model": "flux-chroma"
12
+ },
13
+ "prompt": "draw a hat",
14
+ "resolution": "1280x720",
15
+ "guidance_scale": 3.0,
16
+ "num_inference_steps": 20,
17
+ "batch_size": 1
18
+ }
Wan2GP/defaults/flux_dev_kontext.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev Kontext 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Kontext is a 12 billion parameter rectified flow transformer capable of editing images based on instructions stored in the Prompt. Please be aware that Flux Kontext is picky on the resolution of the input image and the output dimensions may not match the dimensions of the input image.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1_kontext_dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "flux-model": "flux-dev-kontext"
11
+ },
12
+ "prompt": "add a hat",
13
+ "resolution": "1280x720",
14
+ "batch_size": 1
15
+ }
16
+
17
+
Wan2GP/defaults/flux_dev_umo.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev UMO 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Dev UMO is a model that can Edit Images with a specialization in combining multiple image references (resized internally at 512x512 max) to produce an Image output. Best Image preservation at 768x768 Resolution Output.",
6
+ "URLs": "flux",
7
+ "flux-model": "flux-dev-umo",
8
+ "loras": ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-UMO_dit_lora_bf16.safetensors"],
9
+ "resolutions": [ ["1024x1024 (1:1)", "1024x1024"],
10
+ ["768x1024 (3:4)", "768x1024"],
11
+ ["1024x768 (4:3)", "1024x768"],
12
+ ["512x1024 (1:2)", "512x1024"],
13
+ ["1024x512 (2:1)", "1024x512"],
14
+ ["768x768 (1:1)", "768x768"],
15
+ ["768x512 (3:2)", "768x512"],
16
+ ["512x768 (2:3)", "512x768"]]
17
+ },
18
+ "prompt": "the man is wearing a hat",
19
+ "embedded_guidance_scale": 4,
20
+ "resolution": "768x768",
21
+ "batch_size": 1
22
+ }
23
+
24
+
Wan2GP/defaults/flux_dev_uso.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Dev USO 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Dev USO is a model that can Edit Images with a specialization in Style Transfers (up to two).",
6
+ "modules": [ ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-USO_projector_bf16.safetensors"]],
7
+ "URLs": "flux",
8
+ "loras": ["https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-dev-USO_dit_lora_bf16.safetensors"],
9
+ "flux-model": "flux-dev-uso"
10
+ },
11
+ "prompt": "the man is wearing a hat",
12
+ "embedded_guidance_scale": 4,
13
+ "resolution": "1024x1024",
14
+ "batch_size": 1
15
+ }
16
+
17
+
Wan2GP/defaults/flux_krea.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Krea Dev 12B",
4
+ "architecture": "flux",
5
+ "description": "Cutting-edge output quality, with a focus on aesthetic photography..",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-krea-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-krea-dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true,
11
+ "flux-model": "flux-dev"
12
+ },
13
+ "prompt": "draw a hat",
14
+ "resolution": "1280x720",
15
+ "batch_size": 1
16
+ }
Wan2GP/defaults/flux_schnell.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 Schnell 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 Schnell is a 12 billion parameter rectified flow transformer capable of generating images from text descriptions. As a distilled model it requires fewer denoising steps.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-schnell_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-schnell_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "image_outputs": true,
11
+ "flux-model": "flux-schnell"
12
+ },
13
+ "prompt": "draw a hat",
14
+ "resolution": "1280x720",
15
+ "num_inference_steps": 10,
16
+ "batch_size": 1
17
+ }
Wan2GP/defaults/flux_srpo.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 SRPO Dev 12B",
4
+ "architecture": "flux",
5
+ "description": "By fine-tuning the FLUX.1.dev model with optimized denoising and online reward adjustment, SRPO improves its human-evaluated realism and aesthetic quality by over 3x.",
6
+ "URLs": [
7
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-srpo-dev_bf16.safetensors",
8
+ "https://huggingface.co/DeepBeepMeep/Flux/resolve/main/flux1-srpo-dev_quanto_bf16_int8.safetensors"
9
+ ],
10
+ "flux-model": "flux-dev"
11
+ },
12
+ "prompt": "draw a hat",
13
+ "resolution": "1024x1024",
14
+ "batch_size": 1
15
+ }
Wan2GP/defaults/flux_srpo_uso.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": {
3
+ "name": "Flux 1 SRPO USO 12B",
4
+ "architecture": "flux",
5
+ "description": "FLUX.1 SRPO USO is a model that can Edit Images with a specialization in Style Transfers (up to two). It leverages the improved Image quality brought by the SRPO process",
6
+ "modules": [ "flux_dev_uso"],
7
+ "URLs": "flux_srpo",
8
+ "loras": "flux_dev_uso",
9
+ "flux-model": "flux-dev-uso"
10
+ },
11
+ "prompt": "the man is wearing a hat",
12
+ "embedded_guidance_scale": 4,
13
+ "resolution": "1024x1024",
14
+ "batch_size": 1
15
+ }
16
+
17
+
Wan2GP/defaults/fun_inp.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fun InP image2video 14B",
5
+ "architecture" : "fun_inp",
6
+ "description": "The Fun model is an alternative image 2 video that supports out the box End Image fixing (contrary to the original Wan image 2 video model).",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_bf16.safetensors",
9
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_quanto_int8.safetensors",
10
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_14B_quanto_fp16_int8.safetensors"
11
+ ]
12
+ }
13
+ }
Wan2GP/defaults/fun_inp_1.3B.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model":
3
+ {
4
+ "name": "Fun InP image2video 1.3B",
5
+ "architecture" : "fun_inp_1.3B",
6
+ "description": "The Fun model is an alternative image 2 video that supports out the box End Image fixing (contrary to the original Wan image 2 video model). The 1.3B adds also image 2 to video capability to the 1.3B model.",
7
+ "URLs": [
8
+ "https://huggingface.co/DeepBeepMeep/Wan2.1/resolve/main/wan2.1_Fun_InP_1.3B_bf16.safetensors"
9
+ ]
10
+ }
11
+ }