Update README.md
Browse files
README.md
CHANGED
@@ -5,7 +5,7 @@ license: apache-2.0
|
|
5 |
|
6 |
# Model Card: LlavaOLMoBitnet1B
|
7 |
|
8 |
-
Multimodal Large Language Models (MM-LLMs) have seen significant advancements in the last year, demonstrating impressive performance across tasks.
|
9 |
|
10 |
## Paper Abstract
|
11 |
|
|
|
5 |
|
6 |
# Model Card: LlavaOLMoBitnet1B
|
7 |
|
8 |
+
Multimodal Large Language Models (MM-LLMs) have seen significant advancements in the last year, demonstrating impressive performance across tasks. However, to truly democratize AI, models must exhibit strong capabilities and be able to run efficiently on small compute footprints accessible by most. Part of this quest, we introduce LLaVaOLMoBitnet1B - the first Ternary Multimodal LLM capable of accepting Image(s)+Text inputs to produce coherent textual responses. The model is open-sourced along with weights and training scripts to encourage future research into ternary models. We also release a technical report highlighting the training proecss, challenges associated with ternary models and future oppurtunities.
|
9 |
|
10 |
## Paper Abstract
|
11 |
|