license: apache-2.0
language:
- zh
- en
tags:
- Kolors
Kolors: Effective Training of Diffusion Model for Photorealistic Text-to-Image Synthesis
๐ Introduction
Kolors is a large-scale text-to-image generation model based on latent diffusion, developed by the Kuaishou Kolors team. Trained on billions of text-image pairs, Kolors exhibits significant advantages over both open-source and proprietary models in visual quality, complex semantic accuracy, and text rendering for both Chinese and English characters. Furthermore, Kolors supports both Chinese and English inputs, demonstrating strong performance in understanding and generating Chinese-specific content. For more details, please refer to this technical report.
๐ Quick Start
Using with Diffusers
Make sure you upgrade to the latest version of diffusers: pip install -U diffusers
. And then you can run:
import torch
from diffusers import KolorsPipeline
pipe = KolorsPipeline.from_pretrained(
"Kwai-Kolors/Kolors-diffusers",
torch_dtype=torch.float16,
variant='fp16',
trust_remote_code=True,
force_zeros_for_empty_prompt=False,)
pipe = pipe.to("cuda")
image = pipe(
"ไธๅผ ็ข่ซ็็
ง็๏ผๅพฎ่ท๏ผๅ็ฆ๏ผ้ซ่ดจ้๏ผ็ตๅฝฑ๏ผๆฟ็ไธไธช็ๅญ๏ผๅ็โๅฏๅพโ",
height=1024,
width=1024,
num_inference_steps=50,
guidance_scale=5.0,
generator=torch.Generator(pipe.device).manual_seed(66),
).images[0]
image.show()
๐ License&Citation
License
Kolors are fully open-sourced for academic research. For commercial use, please fill out this questionnaire and sent it to [email protected] for registration.
We open-source Kolors to promote the development of large text-to-image models in collaboration with the open-source community. The code of this project is open-sourced under the Apache-2.0 license. We sincerely urge all developers and users to strictly adhere to the open-source license, avoiding the use of the open-source model, code, and its derivatives for any purposes that may harm the country and society or for any services not evaluated and registered for safety. Note that despite our best efforts to ensure the compliance, accuracy, and safety of the data during training, due to the diversity and combinability of generated content and the probabilistic randomness affecting the model, we cannot guarantee the accuracy and safety of the output content, and the model is susceptible to misleading. This project does not assume any legal responsibility for any data security issues, public opinion risks, or risks and liabilities arising from the model being misled, abused, misused, or improperly utilized due to the use of the open-source model and code.
Citation
If you find our work helpful, please cite it!
@article{kolors,
title={Kolors: Effective Training of Diffusion Model for Photorealistic Text-to-Image Synthesis},
author={Kolors Team},
journal={arXiv preprint},
year={2024}
}
Acknowledgments
- Thanks to Diffusers for providing the codebase.
- Thanks to ChatGLM3 for providing the powerful Chinese language model.
Contact Us
If you want to leave a message for our R&D team and product team, feel free to join our WeChat group. You can also contact us via email ([email protected]).