morriszms commited on
Commit
500d111
1 Parent(s): c6faf02

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ falcon-7b-instruct-sharded-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ falcon-7b-instruct-sharded-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ falcon-7b-instruct-sharded-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ falcon-7b-instruct-sharded-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ falcon-7b-instruct-sharded-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ falcon-7b-instruct-sharded-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ falcon-7b-instruct-sharded-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ falcon-7b-instruct-sharded-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ falcon-7b-instruct-sharded-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ falcon-7b-instruct-sharded-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ falcon-7b-instruct-sharded-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ falcon-7b-instruct-sharded-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,89 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - tiiuae/falcon-refinedweb
4
+ language:
5
+ - en
6
+ inference: true
7
+ widget:
8
+ - text: Hey Falcon! Any recommendations for my holidays in Abu Dhabi?
9
+ example_title: Abu Dhabi Trip
10
+ - text: What's the Everett interpretation of quantum mechanics?
11
+ example_title: 'Q/A: Quantum & Answers'
12
+ - text: Give me a list of the top 10 dive sites you would recommend around the world.
13
+ example_title: Diving Top 10
14
+ - text: Can you tell me more about deep-water soloing?
15
+ example_title: Extreme sports
16
+ - text: Can you write a short tweet about the Apache 2.0 release of our latest AI
17
+ model, Falcon LLM?
18
+ example_title: Twitter Helper
19
+ - text: What are the responsabilities of a Chief Llama Officer?
20
+ example_title: Trendy Jobs
21
+ license: apache-2.0
22
+ base_model: vilsonrodrigues/falcon-7b-instruct-sharded
23
+ tags:
24
+ - TensorBlock
25
+ - GGUF
26
+ ---
27
+
28
+ <div style="width: auto; margin-left: auto; margin-right: auto">
29
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
30
+ </div>
31
+ <div style="display: flex; justify-content: space-between; width: 100%;">
32
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
33
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
34
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
35
+ </p>
36
+ </div>
37
+ </div>
38
+
39
+ ## vilsonrodrigues/falcon-7b-instruct-sharded - GGUF
40
+
41
+ This repo contains GGUF format model files for [vilsonrodrigues/falcon-7b-instruct-sharded](https://huggingface.co/vilsonrodrigues/falcon-7b-instruct-sharded).
42
+
43
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
44
+
45
+ ## Prompt template
46
+
47
+ ```
48
+
49
+ ```
50
+
51
+ ## Model file specification
52
+
53
+ | Filename | Quant type | File Size | Description |
54
+ | -------- | ---------- | --------- | ----------- |
55
+ | [falcon-7b-instruct-sharded-Q2_K.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q2_K.gguf) | Q2_K | 3.440 GB | smallest, significant quality loss - not recommended for most purposes |
56
+ | [falcon-7b-instruct-sharded-Q3_K_S.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q3_K_S.gguf) | Q3_K_S | 3.440 GB | very small, high quality loss |
57
+ | [falcon-7b-instruct-sharded-Q3_K_M.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q3_K_M.gguf) | Q3_K_M | 3.702 GB | very small, high quality loss |
58
+ | [falcon-7b-instruct-sharded-Q3_K_L.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q3_K_L.gguf) | Q3_K_L | 3.923 GB | small, substantial quality loss |
59
+ | [falcon-7b-instruct-sharded-Q4_0.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q4_0.gguf) | Q4_0 | 3.767 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
60
+ | [falcon-7b-instruct-sharded-Q4_K_S.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q4_K_S.gguf) | Q4_K_S | 4.230 GB | small, greater quality loss |
61
+ | [falcon-7b-instruct-sharded-Q4_K_M.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q4_K_M.gguf) | Q4_K_M | 4.444 GB | medium, balanced quality - recommended |
62
+ | [falcon-7b-instruct-sharded-Q5_0.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q5_0.gguf) | Q5_0 | 4.538 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
63
+ | [falcon-7b-instruct-sharded-Q5_K_S.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q5_K_S.gguf) | Q5_K_S | 4.770 GB | large, low quality loss - recommended |
64
+ | [falcon-7b-instruct-sharded-Q5_K_M.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q5_K_M.gguf) | Q5_K_M | 5.131 GB | large, very low quality loss - recommended |
65
+ | [falcon-7b-instruct-sharded-Q6_K.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q6_K.gguf) | Q6_K | 6.256 GB | very large, extremely low quality loss |
66
+ | [falcon-7b-instruct-sharded-Q8_0.gguf](https://huggingface.co/tensorblock/falcon-7b-instruct-sharded-GGUF/tree/main/falcon-7b-instruct-sharded-Q8_0.gguf) | Q8_0 | 6.852 GB | very large, extremely low quality loss - not recommended |
67
+
68
+
69
+ ## Downloading instruction
70
+
71
+ ### Command line
72
+
73
+ Firstly, install Huggingface Client
74
+
75
+ ```shell
76
+ pip install -U "huggingface_hub[cli]"
77
+ ```
78
+
79
+ Then, downoad the individual model file the a local directory
80
+
81
+ ```shell
82
+ huggingface-cli download tensorblock/falcon-7b-instruct-sharded-GGUF --include "falcon-7b-instruct-sharded-Q2_K.gguf" --local-dir MY_LOCAL_DIR
83
+ ```
84
+
85
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
86
+
87
+ ```shell
88
+ huggingface-cli download tensorblock/falcon-7b-instruct-sharded-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
89
+ ```
falcon-7b-instruct-sharded-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eab68e11c517da939fad376af61512000c7b41e2b308d44a200cbd38faa538cd
3
+ size 3693518144
falcon-7b-instruct-sharded-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e85ad37ace1fef92d6c937880b74fd8fb08a25e27443983bcd8481e259bc6c6
3
+ size 4212043072
falcon-7b-instruct-sharded-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f6e2616980a19af711e0f11733fafc90d159341d458ad4202764039173580a2
3
+ size 3974591808
falcon-7b-instruct-sharded-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d1ef9f10dd6627d203819d1c1f3483e2f16af62409dd2ff5bfc87128cfa9ad5
3
+ size 3693518144
falcon-7b-instruct-sharded-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c3a6328eaa49c20dd0ead46542ec7c214b00361477860d3351ad914a9c3c614
3
+ size 4044533056
falcon-7b-instruct-sharded-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c845ea4ce7c493ba6562cfaf18fe5d20a1f913097f6dc66e8ff18dd45dfb3a3
3
+ size 4771991104
falcon-7b-instruct-sharded-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aef524788a4cbfb97b1bce23950d0b6d12fb5ebf2f91843ff2e0df2d35de867e
3
+ size 4542410048
falcon-7b-instruct-sharded-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3437fdffe283dafa100a913a6f80037a1f0341b9f8ad7d9770fe8fa982908a60
3
+ size 4872777024
falcon-7b-instruct-sharded-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99a2698bd6640e264b1ddb1f4b257a7d768cde50c5a4fc0f4a63e2768de02056
3
+ size 5509518656
falcon-7b-instruct-sharded-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c225ca261f3ab7a0f962a20e34548fff0267b0b399f4a4102b77255f640502d
3
+ size 5121715520
falcon-7b-instruct-sharded-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e591a73393bfca0f0a28939cdc3cd8e0ce17e25bd581205aa7eaa57996253d0
3
+ size 6717422912
falcon-7b-instruct-sharded-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a0a18ee1d37b5e785d7c9afb7b0519ddfe2347c5cc0f29a2e87ed37f31373cd
3
+ size 7357508928