README.md CHANGED
@@ -1,13 +1,14 @@
1
  ---
2
  extra_gated_heading: Access Llama 2 on Hugging Face
3
- extra_gated_description: >-
4
- This is a form to enable access to Llama 2 on Hugging Face after you have been
5
- granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads) and accept our
6
- license terms and acceptable use policy before submitting this form. Requests
7
- will be processed in 1-2 days.
8
  extra_gated_button_content: Submit
9
  extra_gated_fields:
10
- I agree to share my name, email address and username with Meta and confirm that I have already been granted download access on the Meta website: checkbox
 
 
11
  language:
12
  - en
13
  pipeline_tag: text-generation
@@ -18,6 +19,8 @@ tags:
18
  - pytorch
19
  - llama
20
  - llama-2
 
 
21
  ---
22
  # **Llama 2**
23
  Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.
 
1
  ---
2
  extra_gated_heading: Access Llama 2 on Hugging Face
3
+ extra_gated_description: This is a form to enable access to Llama 2 on Hugging Face
4
+ after you have been granted access from Meta. Please visit the [Meta website](https://ai.meta.com/resources/models-and-libraries/llama-downloads)
5
+ and accept our license terms and acceptable use policy before submitting this form.
6
+ Requests will be processed in 1-2 days.
 
7
  extra_gated_button_content: Submit
8
  extra_gated_fields:
9
+ ? I agree to share my name, email address and username with Meta and confirm that
10
+ I have already been granted download access on the Meta website
11
+ : checkbox
12
  language:
13
  - en
14
  pipeline_tag: text-generation
 
19
  - pytorch
20
  - llama
21
  - llama-2
22
+ - trl
23
+ - sft
24
  ---
25
  # **Llama 2**
26
  Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters. This is the repository for the 7B fine-tuned model, optimized for dialogue use cases and converted for the Hugging Face Transformers format. Links to other models can be found in the index at the bottom.
config.json CHANGED
@@ -1,25 +1,46 @@
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
 
 
5
  "bos_token_id": 1,
6
  "eos_token_id": 2,
 
7
  "hidden_act": "silu",
8
  "hidden_size": 4096,
9
  "initializer_range": 0.02,
10
  "intermediate_size": 11008,
11
  "max_position_embeddings": 4096,
 
12
  "model_type": "llama",
13
  "num_attention_heads": 32,
14
  "num_hidden_layers": 32,
15
  "num_key_value_heads": 32,
16
  "pad_token_id": 0,
17
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
  "rms_norm_eps": 1e-05,
19
  "rope_scaling": null,
 
20
  "tie_word_embeddings": false,
21
- "torch_dtype": "float16",
22
- "transformers_version": "4.31.0.dev0",
23
- "use_cache": true,
24
  "vocab_size": 32000
25
  }
 
1
  {
2
+ "_name_or_path": "NousResearch/Llama-2-7b-chat-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
  "bos_token_id": 1,
9
  "eos_token_id": 2,
10
+ "head_dim": 128,
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
13
  "initializer_range": 0.02,
14
  "intermediate_size": 11008,
15
  "max_position_embeddings": 4096,
16
+ "mlp_bias": false,
17
  "model_type": "llama",
18
  "num_attention_heads": 32,
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 32,
21
  "pad_token_id": 0,
22
  "pretraining_tp": 1,
23
+ "quantization_config": {
24
+ "_load_in_4bit": true,
25
+ "_load_in_8bit": false,
26
+ "bnb_4bit_compute_dtype": "float16",
27
+ "bnb_4bit_quant_storage": "uint8",
28
+ "bnb_4bit_quant_type": "nf4",
29
+ "bnb_4bit_use_double_quant": false,
30
+ "llm_int8_enable_fp32_cpu_offload": false,
31
+ "llm_int8_has_fp16_weight": false,
32
+ "llm_int8_skip_modules": null,
33
+ "llm_int8_threshold": 6.0,
34
+ "load_in_4bit": true,
35
+ "load_in_8bit": false,
36
+ "quant_method": "bitsandbytes"
37
+ },
38
  "rms_norm_eps": 1e-05,
39
  "rope_scaling": null,
40
+ "rope_theta": 10000.0,
41
  "tie_word_embeddings": false,
42
+ "torch_dtype": "float32",
43
+ "transformers_version": "4.46.3",
44
+ "use_cache": false,
45
  "vocab_size": 32000
46
  }
generation_config.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
 
4
  "eos_token_id": 2,
5
  "pad_token_id": 32000,
6
- "do_sample": true,
7
  "temperature": 0.9,
8
  "top_p": 0.6,
9
- "transformers_version": "4.31.0.dev0"
10
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "do_sample": true,
5
  "eos_token_id": 2,
6
  "pad_token_id": 32000,
 
7
  "temperature": 0.9,
8
  "top_p": 0.6,
9
+ "transformers_version": "4.46.3"
10
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66dec18c9f1705b9387d62f8485f4e7d871ca388718786737ed3c72dbfaac9fb
3
- size 9976576152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a1de4c58437e56a92c27a0a9333685a550b4bdf453481bebb3afda814eaf02c
3
+ size 4807953512
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fd6895090da1b2ccffdb93964847709a3b31e6b69fe7dc5a480dce37c811b1d
3
- size 3500296424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5d447419f1baed870ad522be6a83e9801e4a06f51b7aef0e82a1f6cf9e8a79e
3
+ size 524288128
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff