VictorSanh
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,8 @@ This is the model implementation of the CVPR 2024 'Sequential Modeling Enables S
|
|
11 |
|
12 |
LVM is a vision pretraining model that converts various kinds of visual data into visual sentences and performs next-token prediction autoregressively. It is compatible with both GPU and TPU.
|
13 |
|
|
|
|
|
14 |
LVM is built on top of [OpenLLaMA](https://github.com/openlm-research/open_llama) (an autoregressive model) and [OpenMuse](https://github.com/huggingface/open-muse) (a VQGAN that converts images into visual tokens).
|
15 |
|
16 |
This was trained in collaboration with HuggingFace. Thanks [Victor Sanh](https://huggingface.co/VictorSanh) for the support in this project.
|
|
|
11 |
|
12 |
LVM is a vision pretraining model that converts various kinds of visual data into visual sentences and performs next-token prediction autoregressively. It is compatible with both GPU and TPU.
|
13 |
|
14 |
+
You can try out the demo [here](https://huggingface.co/spaces/Emma02/LVM).
|
15 |
+
|
16 |
LVM is built on top of [OpenLLaMA](https://github.com/openlm-research/open_llama) (an autoregressive model) and [OpenMuse](https://github.com/huggingface/open-muse) (a VQGAN that converts images into visual tokens).
|
17 |
|
18 |
This was trained in collaboration with HuggingFace. Thanks [Victor Sanh](https://huggingface.co/VictorSanh) for the support in this project.
|