doberst commited on
Commit
439a804
·
1 Parent(s): d2e2d23

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -36,9 +36,6 @@ For test run results (and good indicator of target use cases), please see the fi
36
  - **License:** Apache 2.0
37
  - **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
38
 
39
- ## Uses
40
-
41
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
 
43
  ### Direct Use
44
 
@@ -60,7 +57,7 @@ Any model can provide inaccurate or incomplete information, and should be used i
60
 
61
  ## How to Get Started with the Model
62
 
63
- The fastest way to get started with dRAGon is through direct import in transformers:
64
 
65
  from transformers import AutoTokenizer, AutoModelForCausalLM
66
  tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
@@ -68,7 +65,7 @@ The fastest way to get started with dRAGon is through direct import in transform
68
 
69
  Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
70
 
71
- The dRAGon model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
72
 
73
  full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
74
 
 
36
  - **License:** Apache 2.0
37
  - **Finetuned from model:** TinyLlama-1.1b - 2.5T checkpoint
38
 
 
 
 
39
 
40
  ### Direct Use
41
 
 
57
 
58
  ## How to Get Started with the Model
59
 
60
+ The fastest way to get started with BLING is through direct import in transformers:
61
 
62
  from transformers import AutoTokenizer, AutoModelForCausalLM
63
  tokenizer = AutoTokenizer.from_pretrained("bling-tiny-llama-v0", trust_remote_code=True)
 
65
 
66
  Please refer to the generation_test .py files in the Files repository, which includes 200 samples and script to test the model. The **generation_test_llmware_script.py** includes built-in llmware capabilities for fact-checking, as well as easy integration with document parsing and actual retrieval to swap out the test set for RAG workflow consisting of business documents.
67
 
68
+ The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
69
 
70
  full_prompt = "<human>: " + my_prompt + "\n" + "<bot>:"
71