Update README.md
Browse files
README.md
CHANGED
@@ -36,9 +36,6 @@ For test run results (and good indicator of target use cases), please see the fi
|
|
36 |
- **License:** Apache 2.0
|
37 |
- **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
|
38 |
|
39 |
-
## Uses
|
40 |
-
|
41 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
42 |
|
43 |
### Direct Use
|
44 |
|
@@ -60,7 +57,7 @@ Any model can provide inaccurate or incomplete information, and should be used i
|
|
60 |
|
61 |
## How to Get Started with the Model
|
62 |
|
63 |
-
The fastest way to get started with
|
64 |
|
65 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
66 |
tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
|
@@ -68,7 +65,7 @@ The fastest way to get started with dRAGon is through direct import in transform
|
|
68 |
|
69 |
Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
|
70 |
|
71 |
-
The
|
72 |
|
73 |
full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
|
74 |
|
|
|
36 |
- **License:** Apache 2.0
|
37 |
- **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
|
38 |
|
|
|
|
|
|
|
39 |
|
40 |
### Direct Use
|
41 |
|
|
|
57 |
|
58 |
## How to Get Started with the Model
|
59 |
|
60 |
+
The fastest way to get started with BLING is through direct import in transformers:
|
61 |
|
62 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
63 |
tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
|
|
|
65 |
|
66 |
Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
|
67 |
|
68 |
+
The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
|
69 |
|
70 |
full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
|
71 |
|