This repo contains bitsandbytes 4bit-NF4 float16 model weights for OmniGen-v1. These are intended for Google Colab users or those with a GPU that does not support bfloat16. Other 4-bit seekers should prefer the bf16-bnb-4bit model as it produces higher quality images. For info about OmniGen see the original model card.

Usage

Set up your environment by following the original Quick Start Guide before getting started.

NOTE: This feature is not officially supported yet. You'll need to install the repo from this pull request.

from OmniGen import OmniGenPipeline, OmniGen

# pass the quantized model in the pipeline
model = OmniGen.from_pretrained('gryan/OmniGen-v1-fp16-bnb-4bit', dtype=torch.float16)
pipe = OmniGenPipeline.from_pretrained("Shitao/OmniGen-v1", model=model)

# proceed as normal!

## Text to Image
images = pipe(
    prompt="A curly-haired man in a red shirt is drinking tea.", 
    height=1024, 
    width=1024, 
    guidance_scale=2.5,
    seed=0,
)
images[0].save("example_t2i.png")  # save output PIL Image

## Multi-modal to Image
# In the prompt, we use the placeholder to represent the image. The image placeholder should be in the format of <img><|image_*|></img>
# You can add multiple images in the input_images. Please ensure that each image has its placeholder. For example, for the list input_images [img1_path, img2_path], the prompt needs to have two placeholders: <img><|image_1|></img>, <img><|image_2|></img>.
images = pipe(
    prompt="A man in a black shirt is reading a book. The man is the right man in <img><|image_1|></img>.",
    input_images=["./imgs/test_cases/two_man.jpg"],
    height=1024, 
    width=1024,
    guidance_scale=2.5, 
    img_guidance_scale=1.6,
    seed=0
)
images[0].save("example_ti2i.png")  # save output PIL image

Image Samples

Text Only FP16 4bit Single Image FP16 4bit Double Image FP16 4bit
Downloads last month
25
Safetensors
Model size
2.1B params
Tensor type
F32
FP16
U8
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for gryan/OmniGen-v1-fp16-bnb-4bit

Base model

Shitao/OmniGen-v1
Quantized
(3)
this model