Papers
arxiv:2412.18653

1.58-bit FLUX

Published on Dec 24, 2024
· Submitted by akhaliq on Dec 30, 2024
#2 Paper of the day
Authors:
,
,

Abstract

We present 1.58-bit FLUX, the first successful approach to quantizing the state-of-the-art text-to-image generation model, FLUX.1-dev, using 1.58-bit weights (i.e., values in {-1, 0, +1}) while maintaining comparable performance for generating 1024 x 1024 images. Notably, our quantization method operates without access to image data, relying solely on self-supervision from the FLUX.1-dev model. Additionally, we develop a custom kernel optimized for 1.58-bit operations, achieving a 7.7x reduction in model storage, a 5.1x reduction in inference memory, and improved inference latency. Extensive evaluations on the GenEval and T2I Compbench benchmarks demonstrate the effectiveness of 1.58-bit FLUX in maintaining generation quality while significantly enhancing computational efficiency.

Community

Paper submitter

Screenshot 2024-12-30 at 12.24.45 PM.png

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Am I missing something or is this paper very light on details, aka completely lacking even a hint of what they're actually doing?

Where is the model?

Out of curiosity - how all this quantized approaches behave with control nets? Are base control nets supported out of the box or CNs need to be retrained too? Will CNs suffer from quantization simplifications or this is orthogonal things?

Without control nets AI imaging is not that useful in real-life tasks, huge crowd still using SDXL (and even SD1.5) purely because of how control nets are effective on UNet archs

give us WEIGHTS
or be GONE!
shoooo
shooo

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2412.18653 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2412.18653 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2412.18653 in a Space README.md to link it from this page.

Collections including this paper 16