Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
TinyLamda
/
llama-3.2-1b-4bit-128g
like
0
Text Generation
Transformers
PyTorch
8 languages
llama
facebook
meta
llama-3
conversational
Inference Endpoints
4-bit precision
gptq
arxiv:
2204.05149
License:
llama3.2
Model card
Files
Files and versions
Community
Train
Deploy
Use this model
main
llama-3.2-1b-4bit-128g
1 contributor
History:
4 commits
TinyLamda
AutoGPTQ model for meta-llama/Llama-3.2-1B: 4bits, gr128, desc_act=False
8dc579c
verified
5 months ago
.gitattributes
Safe
1.52 kB
initial commit
5 months ago
LICENSE.txt
Safe
7.71 kB
Upload 7 files
5 months ago
README.md
Safe
35.9 kB
Upload 7 files
5 months ago
USE_POLICY.md
Safe
6.02 kB
Upload 7 files
5 months ago
config.json
Safe
1.24 kB
AutoGPTQ model for meta-llama/Llama-3.2-1B: 4bits, gr128, desc_act=False
5 months ago
generation_config.json
Safe
189 Bytes
Upload 7 files
5 months ago
gptq_model-4bit-128g.safetensors
Safe
1.56 GB
LFS
AutoGPTQ model for meta-llama/Llama-3.2-1B: 4bits, gr128, desc_act=False
5 months ago
quantize_config.json
Safe
306 Bytes
AutoGPTQ model for meta-llama/Llama-3.2-1B: 4bits, gr128, desc_act=False
5 months ago
special_tokens_map.json
Safe
296 Bytes
Upload 7 files
5 months ago
tokenizer.json
Safe
9.09 MB
Upload 7 files
5 months ago
tokenizer_config.json
Safe
54.5 kB
Upload 7 files
5 months ago