Update README.md
#18
by
nielsr
HF staff
- opened
README.md
CHANGED
@@ -59,6 +59,19 @@ BLIP2 has not been tested in real world applications. It should not be directly
|
|
59 |
|
60 |
For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
|
61 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
62 |
#### Running the model on CPU
|
63 |
|
64 |
<details>
|
|
|
59 |
|
60 |
For code examples, we refer to the [documentation](https://huggingface.co/docs/transformers/main/en/model_doc/blip-2#transformers.Blip2ForConditionalGeneration.forward.example).
|
61 |
|
62 |
+
### Memory requirements
|
63 |
+
|
64 |
+
The memory requirements differ based on the precision one uses. One can use 4-bit inference using [Bitsandbytes](https://huggingface.co/blog/4bit-transformers-bitsandbytes), which greatly reduce the memory requirements.
|
65 |
+
|
66 |
+
Training requires 4 times the
|
67 |
+
|
68 |
+
| dtype | Largest Layer or Residual Group | Total Size | Training using Adam |
|
69 |
+
|-------------------|---------------------------------|------------|----------------------|
|
70 |
+
| float32 | 490.94 MB | 14.43 GB | 57.72 GB |
|
71 |
+
| float16/bfloat16 | 245.47 MB | 7.21 GB | 28.86 GB |
|
72 |
+
| int8 | 122.73 MB | 3.61 GB | 14.43 GB |
|
73 |
+
| int4 | 61.37 MB | 1.8 GB | 7.21 GB |
|
74 |
+
|
75 |
#### Running the model on CPU
|
76 |
|
77 |
<details>
|