Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -1,18 +1,94 @@
|
|
1 |
-
|
2 |
|
3 |
-
|
4 |
-
|
5 |
-
because that is the one most used by AI generative models at the moment
|
6 |
|
7 |
-
|
8 |
-
This file is filtered from the full list of tokens in this model
|
9 |
-
(generated by dumptokens.py),
|
10 |
-
and has only the tokens that are full standalone words
|
11 |
-
(and are also plain English ascii. Sorry, fancy languages)
|
12 |
|
13 |
-
*
|
14 |
-
|
15 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
-
* showtokens.py
|
18 |
-
Given a word, or string of words, shows how T5 tokenizes it
|
|
|
1 |
+
# tokenspace directory
|
2 |
|
3 |
+
This directory contains utilities for the purpose of browsing the
|
4 |
+
"token space" of CLIP ViT-L/14
|
|
|
5 |
|
6 |
+
Primary tools are:
|
|
|
|
|
|
|
|
|
7 |
|
8 |
+
* "calculate-distances.py": allows command-line browsing of words and their neighbours
|
9 |
+
* "graph-embeddings.py": plots graph of full values of two embeddings
|
10 |
+
|
11 |
+
|
12 |
+
## (clipmodel,cliptextmodel)-calculate-distances.py
|
13 |
+
|
14 |
+
Loads the generated embeddings, reads in a word, calculates "distance" to every
|
15 |
+
embedding, and then shows the closest "neighbours".
|
16 |
+
|
17 |
+
To run this requires the files "embeddings.safetensors" and "dictionary",
|
18 |
+
in matching format
|
19 |
+
|
20 |
+
You will need to rename or copy appropriate files for this as mentioned below.
|
21 |
+
|
22 |
+
Note that SD models use cliptextmodel, NOT clipmodel
|
23 |
+
|
24 |
+
## graph-textmodels.py
|
25 |
+
|
26 |
+
Shows the difference between the same word, embedded by CLIPTextModel
|
27 |
+
vs CLIPModel
|
28 |
+
|
29 |
+
## graph-embeddings.py
|
30 |
+
|
31 |
+
Run the script. It will ask you for two text strings.
|
32 |
+
Once you enter both, it will plot the graph and display it for you
|
33 |
+
|
34 |
+
Note that this tool does not require any of the other files; just that you
|
35 |
+
have the requisite python modules installed. (pip install -r requirements.txt)
|
36 |
+
|
37 |
+
### embeddings.safetensors
|
38 |
+
|
39 |
+
You can either copy one of the provided files, or generate your own.
|
40 |
+
See generate-embeddings.py for that.
|
41 |
+
|
42 |
+
Note that you muist always use the "dictionary" file that matchnes your embeddings file
|
43 |
+
|
44 |
+
### embeddings.allids.safetensors
|
45 |
+
|
46 |
+
DO NOT USE THIS ONE for programs that expect a matching dictionary.
|
47 |
+
This one is purely numeric based.
|
48 |
+
Its intention is more for research datamining, but it does have a matching
|
49 |
+
graph front end, graph-byid.py
|
50 |
+
|
51 |
+
|
52 |
+
### dictionary
|
53 |
+
|
54 |
+
Make sure to always use the dictionary file that matches your embeddings file.
|
55 |
+
|
56 |
+
The "dictionary.fullword" file is pulled from fullword.json, which is distilled from "full words"
|
57 |
+
present in the ViT-L/14 CLIP model's provided token dictionary, called "vocab.json".
|
58 |
+
Thus there are only around 30,000 words in it
|
59 |
+
|
60 |
+
If you want to use the provided "embeddings.safetensors.huge" file, you will want to use the matching
|
61 |
+
"dictionary.huge" file, which has over 300,000 words
|
62 |
+
|
63 |
+
This huge file comes from the linux "wamerican-huge" package, which delivers it under
|
64 |
+
/usr/share/dict/american-english-huge
|
65 |
+
|
66 |
+
There also exists a "american-insane" package
|
67 |
+
|
68 |
+
|
69 |
+
## generate-embeddings.py
|
70 |
+
|
71 |
+
Generates the "embeddings.safetensor" file, based on the "dictionary" file present.
|
72 |
+
Takes a few minutes to run, depending on size of the dictionary
|
73 |
+
|
74 |
+
The shape of the embeddings tensor, is
|
75 |
+
[number-of-words][768]
|
76 |
+
|
77 |
+
Note that yes, it is possible to directly pull a tensor from the CLIP model,
|
78 |
+
using keyname of text_model.embeddings.token_embedding.weight
|
79 |
+
|
80 |
+
This will NOT GIVE YOU THE RIGHT DISTANCES!
|
81 |
+
Hence why we are calculating and then storing the embedding weights actually
|
82 |
+
generated by the CLIP process
|
83 |
+
|
84 |
+
|
85 |
+
## fullword.json
|
86 |
+
|
87 |
+
This file contains a collection of "one word, one CLIP token id" pairings.
|
88 |
+
The file was taken from vocab.json, which is part of multiple SD models in huggingface.co
|
89 |
+
|
90 |
+
The file was optimized for what people are actually going to type as words.
|
91 |
+
First all the non-(/w) entries were stripped out.
|
92 |
+
Then all the garbage punctuation and foreign characters were stripped out.
|
93 |
+
Finally, the actual (/w) was stripped out, for ease of use.
|
94 |
|
|
|
|