This is a Llama 2 architecture model series trained on the FineWeb dataset. This is ~500 Million model uses lamma tokenizer. trained using code from Karpathy lamma2
- Downloads last month
- 4
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support