Text-to-Image
Diffusers
Safetensors
English
StableDiffusionPipeline
common-canvas
stable-diffusion
Skylion007 commited on
Commit
abebcb4
·
verified ·
1 Parent(s): ac29fe9

Add Model Card Properly

Browse files
Files changed (1) hide show
  1. README.md +48 -2
README.md CHANGED
@@ -2,10 +2,56 @@
2
  license: cc-by-sa-4.0
3
  tags:
4
  - common-canvas
5
- - text-to-image
6
  datasets:
7
  - common-canvas/commoncatalog-cc-by-sa
8
  - common-canvas/commoncatalog-cc-by
9
  language:
10
  - en
11
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  license: cc-by-sa-4.0
3
  tags:
4
  - common-canvas
 
5
  datasets:
6
  - common-canvas/commoncatalog-cc-by-sa
7
  - common-canvas/commoncatalog-cc-by
8
  language:
9
  - en
10
+ ---
11
+ # CommonCanvas-SNC
12
+
13
+ ## Summary
14
+ CommonCanvas is a family of latent diffusion models capable of generating images from a given text prompt. The architecture is based off of Stable Diffusion 2. Different CommonCanvas models are trained exclusively on subsets of the CommonCatalog Dataset (See Data Card), a large dataset of Creative Commons licensed images with synthetic captions produced using a pre-trained BLIP-2 captioning model.
15
+
16
+ Input: CommonCatalog Text Captions
17
+ Output: CommonCatalog Images
18
+ Architecture: Stable Diffusion 2
19
+ Version Number: 0.1
20
+
21
+ The goal of this purpose is to produce a model that is competitive with Stable Diffusion 2, but to do so using an easily accessible dataset of known provenance. Doing so makes replicating the model significantly easier and provides proper attribution to all the creative commons work used to train the model. The exact training recipe of the model can be found in the paper hosted at this link. https://arxiv.org/abs/2310.16825
22
+
23
+ ## Performance Limitations
24
+
25
+ CommonCanvas under-performs in several categories, including faces, general photography, and paintings (see paper, Figure 8). These datasets all originated from the Conceptual Captions dataset, which relies on web-scraped data. These web-sourced captions, while abundant, may not always align with human-generated language nuances. Transitioning to synthetic captions introduces certain performance challenges, however, the drop in performance is not as dramatic as one might assume.
26
+
27
+ ## Training Dataset Limitations
28
+ The model is trained on 10 year old YFCC data and may not have modern concepts or recent events in its training corpus. Performance on this model will be worse on certain proper nouns or specific celebrities, but this is a feature not a bug. The model may not generate known artwork, individual celebrities, or specific locations due to the autogenerated nature of the caption data.
29
+
30
+
31
+ Note: The non-commercial variants of this model are explicitly not intended to be use
32
+ * It is trained on data derived from the Flickr100M dataset. The information is dated and known to have a bias towards internet connected Western countries. Some areas such as the global south lack representation.
33
+
34
+ ## Associated Risks
35
+ * Text in images produced by the model will likely be difficult to read.
36
+ * The model struggles with more complex tasks that require compositional understanding
37
+ * It may not accurately generate faces or representations of specific people.
38
+ * The model primarily learned from English descriptions and may not perform as effectively in other languages.
39
+ * The autoencoder aspect of the model introduces some information loss.
40
+ * It may be possible to guide the model to generate objectionable content, i.e. nudity or other NSFW material.
41
+
42
+ ## Intended Uses
43
+ * Using the model for generative AI research
44
+ * Safe deployment of models which have the potential to generate harmful content.
45
+ * Probing and understanding the limitations and biases of generative models.
46
+ * Generation of artworks and use in design and other artistic processes.
47
+ * Applications in educational or creative tools.
48
+ * Research on generative models.
49
+
50
+ ## Usage
51
+ We recommend using the MosaicML Diffusion Repo to finetune / train the model: https://github.com/mosaicml/diffusion . Example finetuning code coming soon. See the Gradio Demo [Here](https://github.com/mosaicml/diffusion/blob/main/scripts/gradio_demo.py)
52
+
53
+ ## Evaluation/Validation
54
+ We validated the model against Stability AI’s SD2 model and compared human user study
55
+
56
+ ## Acknowledgements
57
+ We thank @multimodalart, @Wauplin, and @lhoestq at Hugging Face for helping us host the dataset, and model weights.