Update README.md
Browse files
README.md
CHANGED
@@ -18,9 +18,11 @@ tags:
|
|
18 |
|
19 |
# PaliGemma-3B-Chat-v0.1
|
20 |
|
21 |
-
This model is fine-tuned from [google/paligemma-3b-mix-448](https://huggingface.co/google/paligemma-3b-mix-448)
|
22 |
|
23 |
-
![
|
|
|
|
|
24 |
|
25 |
## Usage
|
26 |
|
@@ -54,6 +56,8 @@ generate_ids = model.generate(input_ids, pixel_values=pixel_values, streamer=str
|
|
54 |
|
55 |
## Training procedure
|
56 |
|
|
|
|
|
57 |
### Training hyperparameters
|
58 |
|
59 |
The following hyperparameters were used during training:
|
|
|
18 |
|
19 |
# PaliGemma-3B-Chat-v0.1
|
20 |
|
21 |
+
This model is fine-tuned from [google/paligemma-3b-mix-448](https://huggingface.co/google/paligemma-3b-mix-448) for multiturn chat completions.
|
22 |
|
23 |
+
![example_en](assets/example_en.png)
|
24 |
+
![example_zh](assets/example_zh.png)
|
25 |
+
![example_ja](assets/example_ja.png)
|
26 |
|
27 |
## Usage
|
28 |
|
|
|
56 |
|
57 |
## Training procedure
|
58 |
|
59 |
+
We used [LLaMA Factory](https://github.com/hiyouga/LLaMA-Factory) to fine-tune this model. During fine-tuning, we freezed the vision tower and adjusted the parameters in the language model and projector layer.
|
60 |
+
|
61 |
### Training hyperparameters
|
62 |
|
63 |
The following hyperparameters were used during training:
|