Hugging Face
Models
Datasets
Spaces
Community
Docs
Enterprise
Pricing
Log In
Sign Up
6.4
TFLOPS
56
140
134
Tolga Cangöz
tolgacangoz
Follow
umutphp's profile picture
a-r-r-o-w's profile picture
ShitpostofAI's profile picture
27 followers
·
67 following
standard_ai
AI & ML interests
AIGC
Recent Activity
reacted
to
sayakpaul
's
post
with 🚀
2 days ago
Diffusers supports a good variety of quantization backends. It can be challenging to navigate through them, given the complex nature of diffusion pipelines in general. So, @derekl35 set out to write a comprehensive guide that puts users in the front seat. Explore the different backends we support, learn the trade-offs they offer, and finally, check out the cool space we built that lets you compare quantization results. Give it a go here: https://lnkd.in/gf8Pi4-2
replied
to
sayakpaul
's
post
2 days ago
Diffusers supports a good variety of quantization backends. It can be challenging to navigate through them, given the complex nature of diffusion pipelines in general. So, @derekl35 set out to write a comprehensive guide that puts users in the front seat. Explore the different backends we support, learn the trade-offs they offer, and finally, check out the cool space we built that lets you compare quantization results. Give it a go here: https://lnkd.in/gf8Pi4-2
reacted
to
sayakpaul
's
post
with 🔥
2 days ago
Fast LoRA inference for Flux with Diffusers and PEFT 🚨 There are great materials that demonstrate how to optimize inference for popular image generation models, such as Flux. However, very few cover how to serve LoRAs fast, despite LoRAs being an inseparable part of their adoption. In our latest post, @BenjaminB and I show different techniques to optimize LoRA inference for the Flux family of models for image generation. Our recipe includes the use of: 1. `torch.compile` 2. Flash Attention 3 (when compatible) 3. Dynamic FP8 weight quantization (when compatible) 4. Hotswapping for avoiding recompilation during swapping new LoRAs 🤯 We have tested our recipe with Flux.1-Dev on both H100 and RTX 4090. We achieve at least a *2x speedup* in either of the GPUs. We believe our recipe is grounded in the reality of how LoRA-based use cases are generally served. So, we hope this will be beneficial to the community 🤗 Even though our recipe was tested primarily with NVIDIA GPUs, it should also work with AMD GPUs. Learn the details and the full code here: https://huggingface.co/blog/lora-fast
View all activity
Organizations
tolgacangoz
's models
16
Sort: Recently updated
tolgacangoz/Wan2.2-Animate-14B-Diffusers
Updated
6 days ago
•
77
tolgacangoz/Wan2.2-S2V-14B-Diffusers
Updated
Sep 7
•
1.47k
•
2
tolgacangoz/MAGI-1-T2V-4.5B-Diffusers
Updated
Aug 24
•
3
tolgacangoz/SkyReels-V2-FLF2V-1.3B-540P-Diffusers
Updated
Jul 8
•
1
tolgacangoz/SkyReels-V2-I2V-14B-720P-Diffusers
Updated
Jul 8
•
2
tolgacangoz/SkyReels-V2-I2V-14B-540P-Diffusers
Updated
Jul 8
•
2
tolgacangoz/SkyReels-V2-I2V-1.3B-540P-Diffusers
Updated
Jul 8
tolgacangoz/SkyReels-V2-T2V-14B-720P-Diffusers
Updated
Jul 8
•
1
tolgacangoz/SkyReels-V2-T2V-14B-540P-Diffusers
Updated
Jul 8
•
4
tolgacangoz/SkyReels-V2-DF-14B-720P-Diffusers
Updated
Jul 8
•
3
tolgacangoz/SkyReels-V2-DF-14B-540P-Diffusers
Updated
Jul 8
•
3
tolgacangoz/SkyReels-V2-DF-1.3B-540P-Diffusers
Updated
Jul 8
•
11
tolgacangoz/MAGI-1-I2V-4.5B-Diffusers
Updated
Jun 28
•
2
tolgacangoz/anytext
Text-to-Image
•
Updated
Feb 27
•
61
tolgacangoz/anytext-controlnet
Text-to-Image
•
Updated
Feb 26
•
32
tolgacangoz/matryoshka-diffusion-models
Text-to-Image
•
Updated
Oct 20, 2024
•
22
•
4