Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ tags:
|
|
16 |
|
17 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
18 |
|
19 |
-
This model is an implementation of
|
20 |
|
21 |
|
22 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|
|
|
16 |
|
17 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
18 |
|
19 |
+
This model is an implementation of OpenAI-Clip found [here](https://github.com/openai/CLIP/).
|
20 |
|
21 |
|
22 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|