|
--- |
|
license: mit |
|
datasets: |
|
- neural-bridge/rag-dataset-12000 |
|
- neural-bridge/rag-dataset-1200 |
|
language: |
|
- en |
|
--- |
|
# VERY IMPORTANT |
|
- This model is in alpha phase and is NOT yet recommended for use. |
|
- This model is obsolete today, recommended model [here.](https://huggingface.co/BueormLLC/RAGPT-2_Turbo) |
|
# RAGPT-2 (unfunctional): Fine-tuned GPT-2 for Context-Based Question Answering |
|
|
|
## Model Description |
|
|
|
RAGPT-2 is a fine-tuned version of [GPT-2 small](https://huggingface.co/BueormLLC/CleanGPT), specifically adapted for context-based question answering tasks. This model has been trained to generate relevant answers based on a given context and question, similar to a Retrieval-Augmented Generation (RAG) system. |
|
|
|
### Key Features |
|
|
|
- Based on the GPT-2 small architecture (124M parameters) |
|
- Fine-tuned on the "neural-bridge/rag-dataset-12000" and others dataset from Hugging Face |
|
- Capable of generating answers based on provided context and questions |
|
- Suitable for various question-answering applications |
|
|
|
## Training Data |
|
|
|
The model was fine-tuned using the "neural-bridge/rag-dataset-12000" and "neural-bridge/rag-dataset-1200" dataset, which contains: |
|
- Context passages |
|
- Questions related to the context |
|
- Corresponding answers |
|
|
|
## Fine-tuning Process |
|
|
|
The fine-tuning process involved: |
|
1. Loading the pre-trained GPT-2 small model |
|
2. Preprocessing the dataset to combine context, question, and answer into a single text |
|
3. Training the model to predict the next token given the context and question |
|
|
|
### Hyperparameters |
|
|
|
- Base model: GPT-2 small |
|
- Number of training epochs: 8 |
|
- Batch size: 4 |
|
- Learning rate: Default AdamW optimizer settings |
|
- Max sequence length: 512 tokens |
|
|
|
## Usage |
|
|
|
To use the model: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("BueormLLC/RAGPT-2_unfunctional") |
|
model = AutoModelForCausalLM.from_pretrained("BueormLLC/RAGPT-2_unfunctional") |
|
|
|
context = "Mount Everest is the highest mountain in the world, with a height of 8,848 meters." |
|
question = "What is the height of Mount Everest?" |
|
input_text = f"Context: {context}\nquestion: {question}\nanswer:" |
|
|
|
input_ids = tokenizer.encode(input_text, return_tensors="pt") |
|
output = model.generate(input_ids, max_length=150, num_return_sequences=1) |
|
answer = tokenizer.decode(output[0], skip_special_tokens=True) |
|
|
|
print(f"Respuesta generada: {answer}") |
|
``` |
|
|
|
## Limitations |
|
|
|
- The model's knowledge is limited to its training data and the base GPT-2 model. |
|
- It may sometimes generate irrelevant or incorrect answers, especially for topics outside its training domain. |
|
- The model does not have access to external information or real-time data. |
|
|
|
## Ethical Considerations |
|
|
|
Users should be aware that this model, like all language models, may reflect biases present in its training data. It should not be used as a sole source of information for critical decisions. |
|
|
|
## Future Improvements |
|
|
|
- Fine-tuning on a larger and more diverse dataset |
|
- Experimenting with larger base models (e.g., GPT-2 medium or large) |
|
- Implementing techniques to improve factual accuracy and reduce hallucinations |
|
|
|
## Support us |
|
|
|
- [Paypal](https://paypal.me/bueorm) |
|
- [Patreon](https://patreon.com/bueorm) |
|
### We appreciate your support, without you we could not do what we do. |
|
|
|
## Citation |
|
|
|
If you use this model in your research, please cite: |
|
|
|
``` |
|
@misc{RAGPT, |
|
author = {Bueorm}, |
|
title = {RAGPT-2: Fine-tuned GPT-2 for Context-Based Question Answering}, |
|
year = {2024}, |
|
publisher = {GitHub}, |
|
journal = {None}, |
|
howpublished = {\url{https://huggingface.co/BueormLLC/RAGPT-2_unfunctional}} |
|
} |
|
``` |