Divyasreepat commited on
Commit
4bce176
·
verified ·
1 Parent(s): 7dffeb4

Update README.md with new model card content

Browse files
Files changed (1) hide show
  1. README.md +24 -13
README.md CHANGED
@@ -6,34 +6,41 @@ tags:
6
  - keras
7
  ---
8
  ### Model Overview
9
- # Stable Diffusion 3 Medium
10
- ![demo](https://huggingface.co/stabilityai/stable-diffusion-3-medium/resolve/main/sd3demo.jpg)
11
-
12
- ## Model
13
-
14
- ![mmdit](https://huggingface.co/stabilityai/stable-diffusion-3-medium/resolve/main/mmdit.png)
15
-
16
  [Stable Diffusion 3 Medium](https://stability.ai/news/stable-diffusion-3-medium) is a Multimodal Diffusion Transformer (MMDiT) text-to-image model that features greatly improved performance in image quality, typography, complex prompt understanding, and resource-efficiency.
17
 
18
  For more technical details, please refer to the [Research paper](https://stability.ai/news/stable-diffusion-3-research-paper).
19
 
20
  Please note: this model is released under the Stability Community License. For Enterprise License visit Stability.ai or [contact us](https://stability.ai/enterprise) for commercial licensing details.
21
 
 
 
 
 
 
 
 
 
 
 
 
 
22
 
 
 
 
23
 
24
- ### Model Description
25
 
26
  - **Developed by:** Stability AI
27
  - **Model type:** MMDiT text-to-image generative model
28
- - **Model Description:** This is a model that can be used to generate images based on text prompts. It is a Multimodal Diffusion Transformer
29
- (https://arxiv.org/abs/2403.03206) that uses three fixed, pretrained text encoders
30
  ([OpenCLIP-ViT/G](https://github.com/mlfoundations/open_clip), [CLIP-ViT/L](https://github.com/openai/CLIP/tree/main) and [T5-xxl](https://huggingface.co/google/t5-v1_1-xxl))
31
 
32
- ### Model card
33
- https://huggingface.co/stabilityai/stable-diffusion-3-medium
34
-
35
  ## Example Usage
36
  ```python
 
 
 
 
37
  # Pretrained Stable Diffusion 3 model.
38
  model = keras_hub.models.StableDiffusion3Backbone.from_preset(
39
  "stable_diffusion_3_medium"
@@ -156,6 +163,10 @@ text_to_image.generate(
156
  ## Example Usage with Hugging Face URI
157
 
158
  ```python
 
 
 
 
159
  # Pretrained Stable Diffusion 3 model.
160
  model = keras_hub.models.StableDiffusion3Backbone.from_preset(
161
  "hf://keras/stable_diffusion_3_medium"
 
6
  - keras
7
  ---
8
  ### Model Overview
 
 
 
 
 
 
 
9
  [Stable Diffusion 3 Medium](https://stability.ai/news/stable-diffusion-3-medium) is a Multimodal Diffusion Transformer (MMDiT) text-to-image model that features greatly improved performance in image quality, typography, complex prompt understanding, and resource-efficiency.
10
 
11
  For more technical details, please refer to the [Research paper](https://stability.ai/news/stable-diffusion-3-research-paper).
12
 
13
  Please note: this model is released under the Stability Community License. For Enterprise License visit Stability.ai or [contact us](https://stability.ai/enterprise) for commercial licensing details.
14
 
15
+ ## Links
16
+
17
+ * [SD3 Quickstart Notebook Text-to-image](https://www.kaggle.com/code/laxmareddypatlolla/stablediffusion3-quickstart-notebook)
18
+ * [SD3 Quickstart Notebook Image-to-image](https://colab.sandbox.google.com/gist/laxmareddyp/46de6fbb274b12e8515c7bc55dfc5c57/stable-diffusion-3.ipynb)
19
+ * [SD3 API Documentation](https://keras.io/keras_hub/api/models/stable_diffusion_3/)
20
+ * [SD3 Model Card](https://huggingface.co/stabilityai/stable-diffusion-3-medium)
21
+ * [KerasHub Beginner Guide](https://keras.io/guides/keras_hub/getting_started/)
22
+ * [KerasHub Model Publishing Guide](https://keras.io/guides/keras_hub/upload/)
23
+
24
+ ## Presets
25
+
26
+ The following model checkpoints are provided by the Keras team. Full code examples for each are available below.
27
 
28
+ | Preset name | Parameters | Description |
29
+ |---------------------------------------|------------|--------------------------------------------------------------------------------------------------------------|
30
+ | stable_diffusion_3_medium | 2.99B | 3 billion parameter, including CLIP L and CLIP G text encoders, MMDiT generative model, and VAE autoencoder. Developed by Stability AI. |
31
 
 
32
 
33
  - **Developed by:** Stability AI
34
  - **Model type:** MMDiT text-to-image generative model
35
+ - **Model Description:** This is a model that can be used to generate images based on text prompts. It is a [Multimodal Diffusion Transformer](https://arxiv.org/abs/2403.03206) that uses three fixed, pretrained text encoders
 
36
  ([OpenCLIP-ViT/G](https://github.com/mlfoundations/open_clip), [CLIP-ViT/L](https://github.com/openai/CLIP/tree/main) and [T5-xxl](https://huggingface.co/google/t5-v1_1-xxl))
37
 
 
 
 
38
  ## Example Usage
39
  ```python
40
+ !pip install -U keras-hub
41
+ !pip install -U keras
42
+ ```
43
+ ```
44
  # Pretrained Stable Diffusion 3 model.
45
  model = keras_hub.models.StableDiffusion3Backbone.from_preset(
46
  "stable_diffusion_3_medium"
 
163
  ## Example Usage with Hugging Face URI
164
 
165
  ```python
166
+ !pip install -U keras-hub
167
+ !pip install -U keras
168
+ ```
169
+ ```
170
  # Pretrained Stable Diffusion 3 model.
171
  model = keras_hub.models.StableDiffusion3Backbone.from_preset(
172
  "hf://keras/stable_diffusion_3_medium"