Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[enhancement]: Feature Suggestion: Implementing WaveSpeed for Accelerating Flux by 10x #7592

Open
1 task done
Aittor opened this issue Jan 25, 2025 · 2 comments
Open
1 task done
Labels
enhancement New feature or request

Comments

@Aittor
Copy link

Aittor commented Jan 25, 2025

Is there an existing issue for this?

  • I have searched the existing issues

Contact Details

[email protected]

What should this feature add?

Dear Invoke AI Team,

I hope this message finds you well. I am writing to propose the integration of WaveSpeed into the Flux workflow within Invoke AI.

WaveSpeed is a cutting-edge optimization tool designed to significantly accelerate computational workflows, boasting a remarkable potential to achieve up to a 10x speed increase. Leveraging its capabilities within Flux could substantially enhance user experience by reducing generation times and increasing operational efficiency.

Key Benefits of Integration:
1. Performance Optimization: Accelerate Flux by up to 10x, enabling faster image generation and smoother workflows.
2. Improved User Experience: Shortened waiting times translate to enhanced satisfaction for users relying on real-time outputs.
3. Scalability: WaveSpeed’s optimization capabilities align with the needs of large-scale operations, particularly for demanding use cases in AI content generation.

WaveSpeed Highlights:
• Open-source and actively maintained repository.
• Compatible with existing AI frameworks.
• Proven benchmarks showcasing significant acceleration across various tasks.

I believe this integration aligns perfectly with Invoke AI’s commitment to innovation and efficiency. I am happy to provide additional details, collaborate on testing, or assist with the initial setup to evaluate feasibility.

Thank you for considering this suggestion. I look forward to hearing your thoughts and exploring the potential of this enhancement together.

Best regards,
William

Alternatives

No response

Additional Content

No response

@Aittor Aittor added the enhancement New feature or request label Jan 25, 2025
@iwr-redmond
Copy link

iwr-redmond commented Jan 25, 2025

The associated diffusers code is ParaAttention. Both versions provide two speed-enhancing mechanisms: FBCache and simplified torch.compile.

Some notes:

  • While this is very new code, the same author was responsible for developing stable-fast
  • The Comfy version linked has recently been updated for better compatibility with SDXL (and SD1.5, see Comfy-WaveSpeed#84) in addition to the original Flux and SD3.5 support
  • However, ParaAttention only supports Flux (ParaAttention#20)
  • torch.compile does not work on Windows due to a missing Triton dependency (see Comfy-WaveSpeed#38); however there is a somewhat credible fork that is already in use by the installable ComfyUI
  • There does not appear to be a Triton fork for Mac OS X
  • Unlike the MIT-licensed Comfy code, ParaAttention has a restrictive license barring use on hosted services (a.k.a. Invoke Professional)

This means that only FBCache for Flux could be supported by Invoke for now, and even then only if ParaAttention is re-licensed.

@CommanderJ
Copy link

Even without torch.compile, the speedup from WaveSpeed is significant.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants