ego-hf commited on
Commit
046d990
·
verified ·
1 Parent(s): 7576c88

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -4
README.md CHANGED
@@ -1,20 +1,48 @@
1
  ---
2
- license: other
3
  tags:
4
  - LLM
5
  - CodeLlama
 
6
  - lama.cpp
 
7
  ---
8
 
9
- # Model Card: CodeLlama-7b-Python gguf
10
 
11
- ## Model Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
 
13
  Original Meta `CodeLlama-7b-Python` model converted with [python3 convert.py](https://github.com/ggerganov/llama.cpp) to `gguf` and
14
  `CodeLlama-7b-Python/ggml-model-f32.gguf` and splitted with [gguf-split](https://github.com/ggerganov/llama.cpp) to smaller size chunks up to `split-max-tensors 32`.
15
 
16
  ```bash
17
  python3 convert.py ../codellama/CodeLlama-7b-Python
18
-
19
  ./gguf-split --split --split-max-tensors 32 ./models/CodeLlama-7b-Python/ggml-model-f32.gguf ./models/CodeLlama-7b-Python/ggml-model-f32
20
  ```
 
 
 
 
 
 
 
 
1
  ---
2
+ license: llama2
3
  tags:
4
  - LLM
5
  - CodeLlama
6
+ - CodeLlama-7b-Python
7
  - lama.cpp
8
+ - Python
9
  ---
10
 
11
+ # Model Card: Meta CodeLlama-7b-Python gguf
12
 
13
+ Origin Meta model [CodeLlama-7b-Python](https://llama.meta.com/llama-downloads/), [code llama large language model coding](https://ai.meta.com/blog/code-llama-large-language-model-coding/), [codellama](https://github.com/meta-llama/codellama) converted into gguf format with [llama.cpp](https://github.com/ggerganov/llama.cpp)
14
+
15
+ *Licen*: "Llama 2 is licensed under the LLAMA 2 Community License, Copyright © Meta Platforms, Inc. All Rights Reserved."
16
+
17
+ [Policy](https://llama.meta.com/use-policy/)
18
+
19
+
20
+ # Run model
21
+
22
+ ```bash
23
+ ./main -m ggml-model-f32-00001-of-00010.gguf -p "def fibonacci("
24
+ ```
25
+
26
+
27
+ ## Convert to gguf
28
+ ```bash
29
+ python3 convert.py ../codellama/CodeLlama-7b-Python
30
+ ```
31
+
32
+
33
+ ## Split Model
34
 
35
  Original Meta `CodeLlama-7b-Python` model converted with [python3 convert.py](https://github.com/ggerganov/llama.cpp) to `gguf` and
36
  `CodeLlama-7b-Python/ggml-model-f32.gguf` and splitted with [gguf-split](https://github.com/ggerganov/llama.cpp) to smaller size chunks up to `split-max-tensors 32`.
37
 
38
  ```bash
39
  python3 convert.py ../codellama/CodeLlama-7b-Python
 
40
  ./gguf-split --split --split-max-tensors 32 ./models/CodeLlama-7b-Python/ggml-model-f32.gguf ./models/CodeLlama-7b-Python/ggml-model-f32
41
  ```
42
+
43
+
44
+ ## Merge-back model use
45
+
46
+ ```bash
47
+ ./gguf-split --merge ggml-model-f32-00001-of-00010.gguf ggml-model-f32.gguf
48
+ ```