emrgnt-cmplxty
commited on
Commit
•
c70a2a5
1
Parent(s):
8f60d8e
Update README.md
Browse files
README.md
CHANGED
@@ -1,14 +1,14 @@
|
|
1 |
---
|
2 |
-
license:
|
3 |
---
|
4 |
|
5 |
# SciPhi-SearchAgent-Alpha-7B Model Card
|
6 |
|
7 |
-
The SciPhi-SearchAgent-Alpha-7B is a Large Language Model (LLM) fine-tuned from Mistral-7B-v0.1. This model
|
8 |
|
9 |
-
|
10 |
|
11 |
-
|
12 |
|
13 |
## Model Architecture
|
14 |
|
@@ -25,19 +25,33 @@ Base Model: Mistral-7B-v0.1
|
|
25 |
|
26 |
It is recommended to use a single search query. The model will return an answer using search results as context.
|
27 |
|
28 |
-
|
29 |
-
|
30 |
```
|
31 |
export SCIPHI_API_KEY=MY_SCIPHI_API_KEY
|
32 |
# Use the SciPhi `SearchAgent` for LLM RAG w/ AgentSearch
|
33 |
python -m agent_search.scripts.run_rag run --query="What is Fermat's last theorem?"
|
34 |
```
|
35 |
|
36 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
|
38 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
39 |
|
40 |
## References
|
41 |
|
42 |
-
1. Mistral AI. (2023). Model Card for Mistral-7B-v0.1. The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks tested. For full details, please refer to the paper and release blog post. Model Architecture: Transformer with Grouped-Query Attention, Sliding-Window Attention, and Byte-fallback BPE tokenizer. [Link](https://huggingface.co/mistralai/Mistral-7B-v0.1)
|
43 |
-
|
|
|
1 |
---
|
2 |
+
license: apache-2.0
|
3 |
---
|
4 |
|
5 |
# SciPhi-SearchAgent-Alpha-7B Model Card
|
6 |
|
7 |
+
The SciPhi-SearchAgent-Alpha-7B is a Large Language Model (LLM) fine-tuned from Mistral-7B-v0.1. This model was fine tuned with a fully synthetic dataset to specialize at performing retrieval-augmented generation (RAG) over detailed web search results. This work aims to train an agent which specializes in using search engines such as [AgentSearch](https://huggingface.co/datasets/SciPhi/AgentSearch-V1) to generate accurate and well-cited summaries from a range of search results, providing more accurate answers to user queries. Please refer to the [docs here](https://agent-search.readthedocs.io/en/latest/) for more information on how to run the agent in practice.
|
8 |
|
9 |
+
Currently, SciPhi-SearchAgent-Alpha-7B is available via hosted api at https://www.sciphi.ai.
|
10 |
|
11 |
+
You can try a demonstration of SearchAgent [here](https://search.sciphi.ai/).
|
12 |
|
13 |
## Model Architecture
|
14 |
|
|
|
25 |
|
26 |
It is recommended to use a single search query. The model will return an answer using search results as context.
|
27 |
|
28 |
+
Using the AgentSearch package an example is shown below.
|
|
|
29 |
```
|
30 |
export SCIPHI_API_KEY=MY_SCIPHI_API_KEY
|
31 |
# Use the SciPhi `SearchAgent` for LLM RAG w/ AgentSearch
|
32 |
python -m agent_search.scripts.run_rag run --query="What is Fermat's last theorem?"
|
33 |
```
|
34 |
|
35 |
+
Alternatively, you may provide your own search context directly to the model by adhereing to the following format:
|
36 |
+
|
37 |
+
```
|
38 |
+
### Instruction:
|
39 |
+
Your task is to perform retrieval augmented generation (RAG) over the given query and search results. Return your answer with three sections `My Work`, `My Answer`, and `My Further Considerations`.
|
40 |
+
|
41 |
+
Query:
|
42 |
+
{query}
|
43 |
+
|
44 |
+
Search Results:
|
45 |
+
{search_results}
|
46 |
+
|
47 |
+
Query:
|
48 |
+
{query}
|
49 |
+
|
50 |
+
### Response:
|
51 |
+
```
|
52 |
|
53 |
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
54 |
|
55 |
## References
|
56 |
|
57 |
+
1. Mistral AI. (2023). Model Card for Mistral-7B-v0.1. The Mistral-7B-v0.1 Large Language Model (LLM) is a pretrained generative text model with 7 billion parameters. Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks tested. For full details, please refer to the paper and release blog post. Model Architecture: Transformer with Grouped-Query Attention, Sliding-Window Attention, and Byte-fallback BPE tokenizer. [Link](https://huggingface.co/mistralai/Mistral-7B-v0.1)
|
|