pankajmathur commited on
Commit
a0c83d9
·
verified ·
1 Parent(s): 19f0211

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -38,7 +38,7 @@ Hello Orca Mini, what can you do for me?<|eot_id|>
38
  <|start_header_id|>assistant<|end_header_id|>
39
  ```
40
 
41
- Below shows a code example on how to use this model in default half precision (bfloat16) format
42
 
43
  ```python
44
  import torch
@@ -58,7 +58,7 @@ outputs = pipeline(messages, max_new_tokens=128, do_sample=True, temperature=0.0
58
  print(outputs[0]["generated_text"][-1])
59
  ```
60
 
61
- Below shows a code example on how to use this model in 4-bit format via bitsandbytes library
62
 
63
  ```python
64
  import torch
@@ -86,7 +86,7 @@ print(outputs[0]["generated_text"][-1])
86
 
87
  ```
88
 
89
- Below shows a code example on how to use this model in 8-bit format via bitsandbytes library
90
 
91
  ```python
92
  import torch
 
38
  <|start_header_id|>assistant<|end_header_id|>
39
  ```
40
 
41
+ Below shows a code example on how to use this model in default half precision (bfloat16) format, it requires around ~133GB VRAM
42
 
43
  ```python
44
  import torch
 
58
  print(outputs[0]["generated_text"][-1])
59
  ```
60
 
61
+ Below shows a code example on how to use this model in 4-bit format via bitsandbytes library, it requires around ~39GB VRAM
62
 
63
  ```python
64
  import torch
 
86
 
87
  ```
88
 
89
+ Below shows a code example on how to use this model in 8-bit format via bitsandbytes library, it requires around ~69GB VRAM
90
 
91
  ```python
92
  import torch