license: cc-by-4.0 datasets: - santoshtyss/uk_legislation language: - en base_model: - unsloth/Llama-3.2-3B tags: - legal
Llama 3.2 UK Legislation 3B
This model is a fine-tuned version of the Llama 3.2 UK Legislation 3B base. It was instruction-tuned for Q and A on UK legislation. It was trained as part of a blog series, see the article here
Model Details
Model Description
- Developed by: GPT-LABS.AI
- Model type: Transformer-based language model
- Language: English
- License: CC BY 4.0
- Base model: llama-3.2-uk-legislation-3b
Model Sources
- Repository: EryriLabs/llama-3.2-uk-legislation-3b
- Blog Post: Making a Domain-Specific UK Legislation LLM: Part 1 - Pretraining
Uses
Intended Use
This model is designed to serve as Q and A for UK legislation and for further development for tasks such as:
- Domain-specific applications in law or other fields
- Research and experimentation in natural language processing
- General-purpose natural language understanding and generation
Out-of-Scope Use
This model is not suitable for:
- Providing domain-specific expertise
- Applications requiring high accuracy or nuanced understanding of UK legislation
- Tasks involving sensitive or critical real-world applications without rigorous evaluation
Bias, Risks, and Limitations
- Bias: The model may reflect biases inherent in the pretraining data. Outputs should be critically evaluated for accuracy and fairness.
- Risks: As a base model, it may generate responses that are overly general or contextually inappropriate for specific tasks.
- Limitations: The model is not fine-tuned for specific domains, including legal or legislative text, and does not include the most recent developments in any field.
How to Get Started with the Model
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load model and tokenizer
model = AutoModelForCausalLM.from_pretrained("EryriLabs/llama-3.2-uk-legislation-instruct-3b", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("EryriLabs/llama-3.2-uk-legislation-instruct-3b")
# Sample question
input_text = "What are the main principles of UK legislation?"
# Tokenize and generate response
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")
outputs = model.generate(inputs["input_ids"], max_length=50)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Technical Specifications
- Model Architecture: Llama 3.2 3B, a transformer-based model designed for natural language processing tasks.
- Training Data: Pretrained on a diverse dataset of general text.
- Compute Infrastructure: Training conducted on high-performance GPUs (e.g., NVIDIA A100).
Citation
If you use this model, please cite:
@misc{llama3.2-uk-legislation-instruct-3b,
author = {GPT-LABS.AI},
title = {Llama 3.2 UK Legislation Instruct 3B},
year = {2024},
publisher = {Hugging Face},
url = {https://huggingface.co/EryriLabs/llama-3.2-uk-legislation-instruct-3b}
}
Model Card Authors
- GPT-LABS.AI
Contact
For questions or feedback, please visit gpt-labs.ai
- Downloads last month
- 10
Model tree for EryriLabs/llama-3.2-uk-legislation-instruct-3b
Base model
meta-llama/Llama-3.2-3B
Finetuned
unsloth/Llama-3.2-3B
Finetuned
EryriLabs/llama-3.2-uk-legislation-3b