Model Summary
OpenCelestial_1 is a compact and efficient language model fine-tuned on a greeting dataset. It demonstrates that small LLMs can achieve remarkable conversational capabilities, even when trained on consumer-grade hardware.
Based on the GPT-2 architecture, OpenCelestial_1 is optimized for clear, polite, and structured responses, making it ideal for use cases such as:
Chatbots
Instruction-following assistants
Lightweight deployments on limited hardware
Model Training
Base Model: openai-community/gpt2
Dataset: Custom greeting dataset with structured "User" and "AI" dialogue pairs.
Hardware: Fine-tuned on a single NVIDIA RTX 3060.
Optimization: Fine-tuning utilized LoRA (Low-Rank Adaptation) to improve memory efficiency.
Usage Example
To interact with OpenCelestial_1, use the following Python script:
pip install transformers torch
Copy and paste the following script:
from transformers import GPT2LMHeadModel, GPT2Tokenizer
import torch
# Load the model and tokenizer
model_path = "theaithinker/OpenCelestial_1"
model = GPT2LMHeadModel.from_pretrained(model_path)
tokenizer = GPT2Tokenizer.from_pretrained(model_path)
# Set the pad token to the EOS token if not already set
tokenizer.pad_token = tokenizer.eos_token
print("Chatbot is ready! Type 'exit' to quit.")
while True:
user_input = input("You: ")
if user_input.lower() == "exit":
print("Chatbot: Goodbye!")
break
# Define the system prompt and the full prompt
system_prompt = "You are an intelligent AI assistant that will answer every question to the best of your ability. Be clear and polite with your answers."
prompt = f"{system_prompt}\n### Instruction:\n{user_input}\n### Response:"
# Tokenize the input
inputs = tokenizer(
prompt,
return_tensors="pt",
padding=True,
truncation=True,
max_length=1024,
)
input_ids = inputs.input_ids.to(model.device)
attention_mask = inputs.attention_mask.to(model.device)
# Generate the response
with torch.no_grad():
outputs = model.generate(
input_ids=input_ids,
attention_mask=attention_mask,
max_new_tokens=150,
pad_token_id=tokenizer.eos_token_id,
do_sample=True,
temperature=0.7,
top_k=50,
top_p=0.95,
)
# Decode the response and clean it up
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
clean_response = response.split("### Response:")[-1].strip()
print(f"Chatbot: {clean_response}")
Example Outputs
Prompt: Hello there! Response: Hello there! I am just an AI assistant, but I’m here to help you with anything you need.
Prompt: Can you tell me a joke? Response: Sure! Why don’t skeletons fight each other? Because they don’t have the guts!
Prompt: What is the capital of France? Response: The capital of France is Paris. Training Details
LoRA Configuration:
Rank (r): 4
Alpha: 16
Dropout: 0.1
Target Modules: GPT-2’s attention layers (attn.c_attn)
Training Arguments:
Mixed precision: Enabled (fp16)
Epochs: 3
Batch size: 2 (to fit GPU memory)
Learning rate: 5e-5
Performance
OpenCelestial_1 demonstrates:
Clear conversational ability with polite, structured responses.
Low resource requirements, suitable for GPUs like the RTX 3060.
Consistency in instruction-following tasks.
Intended Use
This model is designed for:
Conversational AI applications.
Instruction-based assistants that respond politely and clearly.
Lightweight deployments for hobbyists, small-scale developers, or educational purposes.
Limitations
Responses may still contain hallucinations or factual inaccuracies.
Performance is limited to the dataset scope and GPT-2’s inherent capabilities.
Citation
If you use OpenCelestial_1 in your work, please consider citing:
@misc{OpenCelestial_1, author = {Your Name or Organization}, title = {OpenCelestial_1: A Compact GPT-2 Fine-Tuned Model}, year = {2024}, howpublished = {\url{https://huggingface.co/your_username/OpenCelestial_1}}, }
Acknowledgments
Base Model: openai-community/gpt2
Fine-tuned using the LoRA technique for efficient memory usage.
Developed on a single NVIDIA RTX 3060 GPU.
- Downloads last month
- 8