krish-emissary commited on
Commit
6b88e5f
·
verified ·
1 Parent(s): 7a82ba9

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -1,2 +1,35 @@
1
- *.safetensors filter=lfs diff=lfs merge=lfs -text
 
2
  *.bin filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
  *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,85 +1,87 @@
 
1
  ---
2
- license: llama2
3
- base_model: meta-llama/CodeLlama-70b-Python-hf
4
- tags:
5
- - code
6
- - code-generation
7
- - tab-completion
8
- - python
9
- - llama
10
- - finetuned
11
  language:
12
- - code
 
 
 
 
13
  ---
 
 
14
 
15
- # Python Tab Completion CodeLlama 70B
 
16
 
17
- ## Model Description
 
 
 
 
 
18
 
19
- This is a finetuned version of Code-Llama-70B specifically optimized for Python tab completion tasks. The model excels at predicting the next tokens in Python code, making it ideal for IDE autocomplete features and code assistance tools.
20
 
21
- ## Intended Use
22
 
23
- - **Primary use case**: Python code tab completion in IDEs and code editors
24
- - **Secondary uses**:
25
- - Code generation
26
- - Code explanation
27
- - Python snippet completion
28
 
29
- ## Usage
30
 
31
- ### Quick Start
 
 
 
32
 
33
- ```python
34
- from transformers import AutoModelForCausalLM, AutoTokenizer
35
- import torch
36
 
37
- model_id = "emissary-ai/Python-Tab-Completion-CodeLlama-70b"
38
- tokenizer = AutoTokenizer.from_pretrained(model_id)
39
- model = AutoModelForCausalLM.from_pretrained(
40
- model_id,
41
- torch_dtype=torch.float16,
42
- device_map="auto"
43
- )
44
- ```
45
 
46
- ### Example: Complete Python code
47
 
48
- ```python
49
- prompt = "def calculate_average(numbers):\n "
50
- inputs = tokenizer(prompt, return_tensors="pt")
51
- outputs = model.generate(**inputs, max_length=100, temperature=0.7)
52
- completion = tokenizer.decode(outputs[0], skip_special_tokens=True)
53
- print(completion)
54
- ```
55
 
56
- ### Limitations
57
 
58
- - Optimized specifically for Python; performance on other languages may vary
59
- - Best suited for short to medium-length completions
60
- - May require significant computational resources due to model size (70B parameters)
61
 
62
- ### Ethical Considerations
63
 
64
- - Should not be used as the sole tool for production code without human review
65
- - May reflect biases present in the training data
66
- - Generated code should be tested and validated before deployment
67
 
68
- # License
 
69
 
70
- This model is subject to the Meta Llama 2 Community License Agreement. By using this model, you agree to the terms of the Llama 2 license.
71
 
72
- # Citation
73
 
74
- If you use this model in your research or applications, please cite:
75
- ```bibtex
76
- @misc{python-tab-completion-codellama-70b,
77
- author = {Emissary AI},
78
- title = {Python Tab Completion CodeLlama 70B},
79
- year = {2024},
80
- publisher = {Hugging Face},
81
- howpublished = {\url{https://huggingface.co/emissary-ai/Python-Tab-Completion-CodeLlama-70b}}
82
- }
83
- ```
84
- # Contact
85
- For questions or issues, please open an issue in the Community tab.
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
 
 
 
 
 
 
 
 
 
3
  language:
4
+ - code
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - llama-2
8
+ license: llama2
9
  ---
10
+ # **Code Llama**
11
+ Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 70B Python specialist version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom.
12
 
13
+ > [!NOTE]
14
+ > This is a non-official Code Llama repo. You can find the official Meta repository in the [Meta Llama organization](https://huggingface.co/meta-llama/CodeLlama-70b-Python-hf).
15
 
16
+ | | Base Model | Python | Instruct |
17
+ | --- | ----------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------- |
18
+ | 7B | [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) | [codellama/CodeLlama-7b-Python-hf](https://huggingface.co/codellama/CodeLlama-7b-Python-hf) | [codellama/CodeLlama-7b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-7b-Instruct-hf) |
19
+ | 13B | [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) | [codellama/CodeLlama-13b-Python-hf](https://huggingface.co/codellama/CodeLlama-13b-Python-hf) | [codellama/CodeLlama-13b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-13b-Instruct-hf) |
20
+ | 34B | [codellama/CodeLlama-34b-hf](https://huggingface.co/codellama/CodeLlama-34b-hf) | [codellama/CodeLlama-34b-Python-hf](https://huggingface.co/codellama/CodeLlama-34b-Python-hf) | [codellama/CodeLlama-34b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-34b-Instruct-hf) |
21
+ | 70B | [codellama/CodeLlama-70b-hf](https://huggingface.co/codellama/CodeLlama-70b-hf) | [codellama/CodeLlama-70b-Python-hf](https://huggingface.co/codellama/CodeLlama-70b-Python-hf) | [codellama/CodeLlama-70b-Instruct-hf](https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf) |
22
 
23
+ ## Model Use
24
 
25
+ To use this model, please make sure to install `transformers`.
26
 
27
+ ```bash
28
+ pip install transformers accelerate
29
+ ```
 
 
30
 
31
+ Model capabilities:
32
 
33
+ - [x] Code completion.
34
+ - [ ] Infilling.
35
+ - [ ] Instructions / chat.
36
+ - [x] Python specialist.
37
 
38
+ ## Model Details
39
+ *Note: Use of this model is governed by the Meta license. Meta developed and publicly released the Code Llama family of large language models (LLMs).
 
40
 
41
+ **Model Developers** Meta
 
 
 
 
 
 
 
42
 
43
+ **Variations** Code Llama comes in four model sizes, and three variants:
44
 
45
+ * Code Llama: base models designed for general code synthesis and understanding
46
+ * Code Llama - Python: designed specifically for Python
47
+ * Code Llama - Instruct: for instruction following and safer deployment
 
 
 
 
48
 
49
+ All variants are available in sizes of 7B, 13B, 34B, and 70B parameters.
50
 
51
+ **This repository contains the Python version of the 70B parameters model.**
 
 
52
 
53
+ **Input** Models input text only.
54
 
55
+ **Output** Models generate text only.
 
 
56
 
57
+ **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture.
58
+ **Model Architecture** Code Llama is an auto-regressive language model that uses an optimized transformer architecture. It was fine-tuned with up to 16k tokens. This variant **does not** support long context of up to 100k tokens.
59
 
60
+ **Model Dates** Code Llama and its variants have been trained between January 2023 and January 2024.
61
 
62
+ **Status** This is a static model trained on an offline dataset. Future versions of Code Llama - Instruct will be released as we improve model safety with community feedback.
63
 
64
+ **License** A custom commercial license is available at: [https://ai.meta.com/resources/models-and-libraries/llama-downloads/](https://ai.meta.com/resources/models-and-libraries/llama-downloads/)
65
+
66
+ **Research Paper** More information can be found in the paper "[Code Llama: Open Foundation Models for Code](https://ai.meta.com/research/publications/code-llama-open-foundation-models-for-code/)" or its [arXiv page](https://arxiv.org/abs/2308.12950).
67
+
68
+ ## Intended Use
69
+ **Intended Use Cases** Code Llama and its variants are intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama - Python is designed specifically to handle the Python programming language, and Code Llama - Instruct is intended to be safer to use for code assistant and generation applications.
70
+
71
+ **Out-of-Scope Uses** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. Use in any other way that is prohibited by the Acceptable Use Policy and Licensing Agreement for Code Llama and its variants.
72
+
73
+ ## Hardware and Software
74
+ **Training Factors** We used custom training libraries. The training and fine-tuning of the released models have been performed Meta’s Research Super Cluster.
75
+
76
+ **Carbon Footprint** In aggregate, training all 12 Code Llama models required 1400K GPU hours of computation on hardware of type A100-80GB (TDP of 350-400W). Estimated total emissions were 228.55 tCO2eq, 100% of which were offset by Meta’s sustainability program.
77
+
78
+ ## Evaluation Results
79
+
80
+ See evaluations for the main models and detailed ablations in Section 3 and safety evaluations in Section 4 of the research paper.
81
+
82
+
83
+ ## Ethical Considerations and Limitations
84
+
85
+ Code Llama and its variants are a new technology that carries risks with use. Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, Code Llama’s potential outputs cannot be predicted in advance, and the model may in some instances produce inaccurate or objectionable responses to user prompts. Therefore, before deploying any applications of Code Llama, developers should perform safety testing and tuning tailored to their specific applications of the model.
86
+
87
+ Please see the Responsible Use Guide available available at [https://ai.meta.com/llama/responsible-use-guide](https://ai.meta.com/llama/responsible-use-guide).
config.json CHANGED
@@ -6,13 +6,11 @@
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
- "head_dim": 128,
10
  "hidden_act": "silu",
11
  "hidden_size": 8192,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 28672,
14
  "max_position_embeddings": 4096,
15
- "mlp_bias": false,
16
  "model_type": "llama",
17
  "num_attention_heads": 64,
18
  "num_hidden_layers": 80,
 
6
  "attention_dropout": 0.0,
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
 
9
  "hidden_act": "silu",
10
  "hidden_size": 8192,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 28672,
13
  "max_position_embeddings": 4096,
 
14
  "model_type": "llama",
15
  "num_attention_heads": 64,
16
  "num_hidden_layers": 80,
model-00001-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4fc914f0686202109415fe8acbd52592873800c6e3aa8c8f748464f13d0bb1f3
3
- size 4718922456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dc39667e9e9bb99b37e164514e2517c859a21b69083d1e391c95454538b8fc0
3
+ size 4718922480
model-00002-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd8714a02fcab84f418162d065db13cccd737f314cd6264af5a6ece8702c1d37
3
- size 4664167352
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93cb9a3be35a63540022de4c9e0d19f48310ea491ebd11f077575c0907b523d7
3
+ size 4664167376
model-00003-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42c6245a0d19a35d6be65a8326c52831448d074ed6a7fedf0262ef2e323e3f7a
3
- size 4999711672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:804d76bf0780390092e0ad15338b40b82221f5a89df019e58c7ef71ddf953433
3
+ size 4999711704
model-00006-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbe52e66965bca75113dcc123cc58f5316da5c00e03cbff7bc436e08becffd9a
3
- size 4664167384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f325631196d211f66f4b614d7ef7733869da094044732b48399da5f14dc6992
3
+ size 4664167408
model-00007-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:006bfeaff222b529cac7d0e6a915b5a9848932f8c77f27d1473ed4bc8f8a35f4
3
- size 4664167376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49e617ff57141a9cd574fb68738d8cc7700d8732f3f793b6e6b5ffd99967f856
3
+ size 4664167408
model-00009-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc1afc067015b19212ac1d73b99028c7c39baa763af38595fe0b3554f4ff6ff1
3
- size 4966157032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10339c063a9708cadde61d3a690d1a2b65cbdb66a3af02f762fa6220379e9494
3
+ size 4966157056
model-00010-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed320d184e72cce9838e1c4db01b4be0147aa4fd16bf91b7538171b23674d01c
3
- size 4664134384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44cf1bc764a3183a1a30478399dd14810da918dc861f1f3f8ca5722e3051aa5c
3
+ size 4664134408
model-00011-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7b2fb09972a6553c8e1967d6caa10f7b2b2217c7a916ca3e68ea8e79dcdaa0d
3
- size 4664167384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:850147b678c3950665c18042ed185022d8c18b9e7706a08f5b56a3e165ea1d08
3
+ size 4664167408
model-00012-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0778a3f85527b9cb3e228cf7591837e451f2bd262b0bc70b2664b6751c63e946
3
- size 4664167376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:054899a12a20c0014e304b1f69260676d4ee7b6437789d3a06eabd48588ffe0b
3
+ size 4664167408
model-00013-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f24829683f7fbccfeb53f053349e19b7d8fb57eed155c8b713386d8b361646c7
3
- size 4999711704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebf60aa2ee9a142b54d164c207340aa6fc2633e9986d54fe820b3fff7abfc75a
3
+ size 4999711728
model-00014-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0ab4b7eab285245caaca0b39abb636d33fb6e9d05f61b4cf2b0638bbae556cc
3
- size 4966157032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99d9ec085297c0e9771261376e128487b7e7d8cb8ff48ca68cd1004d2f99ceba
3
+ size 4966157056
model-00018-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:391419edfb8639479997da77bbcb29d52be7c8fa4d0cd454bb129e4e41a51c89
3
- size 4999711704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5309faee1ab036dfbe8103f4aa8c2dba77988e9a2c5c92fd5f00580f650f486
3
+ size 4999711728
model-00019-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c540ad1ad3859989476660ca36e27f4f218c64fd79f284b8443e549208971ec
3
- size 4966157032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:140288ba636559cef56179e649253997f8fa72437717b21bc060262bc3416468
3
+ size 4966157056
model-00020-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9b5ef3fcdef762e4cdf719e9bf7b96f43e62e2e93c5cc5c9c96a761814fbacb
3
- size 4664134384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7466b2e9e116445857d71019cef746dbc63176b5568ad600da48a08294cb63e
3
+ size 4664134408
model-00021-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7dceac750e81d0767fb28409668739120ba790213fd09caede4319cdc1ba81f1
3
- size 4664167384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d28bcf68247f93c5b70626b917aa0349103feaedd267714b6e4e9bc5ac65b689
3
+ size 4664167408
model-00022-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6dd8bcc76452faeb4b236509fc99d1329b990c1bfc4898fbcb3356dbb1cba1b6
3
- size 4664167376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d9419b0dbd0fd6a8784180d8aca472fc922cecefb03083b9f8371a86c708097
3
+ size 4664167408
model-00023-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cad07c41268fb2bbc61df903ecdaff5eb1647d4195b71d4d99a377a3d872f37
3
- size 4999711704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b0b88e28c69c3f2245d5b68f11832e2161f578881ed7f6a8dfe5f01d53051d3
3
+ size 4999711728
model-00024-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d3bb1e96bce1b7c1b89c51fc9ba7a8ce57ad2edbd2fe0cae2d30b93f2d0d22e
3
- size 4966157032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed64d23c1c38899c2ff00e1d3bf1f73239ca06b7d691878668dbe31cbee115a8
3
+ size 4966157056
model-00025-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6775f3c0015b01f8e91dcbc614cf078a91b4328f292e1acae6565ed3d2bc0bc
3
- size 4664134384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b93038be5716e03c6304a7574dd0cb90f65ebb54fb519669c5046ae1249060b5
3
+ size 4664134408
model-00026-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd432e358119b853a72fb923c64ac61f6ed4e8b4a322f90d9d29458ad1a5d43d
3
- size 4664167384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf37ac601e61c9d476cd2edda9b2f95d7a3c89e9d90a6226f455758b891c8dd1
3
+ size 4664167408
model-00027-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb7c71616906b667a611bc32bd4ad86a945d77f64dcaaf6bc83465a4464e0af5
3
- size 4664167376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f561fea5218a2cb49f210748bf766dfe63603912e8950f0a7f5d4ddaa5d0512a
3
+ size 4664167408
model-00028-of-00029.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55895c4b3500bfacf39547577a0149fc7b85f58a7a8f152438c3de9f619ff11b
3
- size 4999711704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1001859efc68cdefe27f45f9910af9ff8ea7c864124270e2fcfc6038c22b531b
3
+ size 4999711728
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -13,7 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "</s>",
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
16
  "unk_token": {
17
  "content": "<unk>",
18
  "lstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,7 +1,6 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
- "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
@@ -31,10 +30,9 @@
31
  "bos_token": "<s>",
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
- "extra_special_tokens": {},
35
  "legacy": true,
36
  "model_max_length": 1000000000000000019884624838656,
37
- "pad_token": "</s>",
38
  "sp_model_kwargs": {},
39
  "spaces_between_special_tokens": false,
40
  "tokenizer_class": "LlamaTokenizer",
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
 
30
  "bos_token": "<s>",
31
  "clean_up_tokenization_spaces": false,
32
  "eos_token": "</s>",
 
33
  "legacy": true,
34
  "model_max_length": 1000000000000000019884624838656,
35
+ "pad_token": null,
36
  "sp_model_kwargs": {},
37
  "spaces_between_special_tokens": false,
38
  "tokenizer_class": "LlamaTokenizer",