|
--- |
|
library_name: transformers |
|
tags: |
|
- transformers |
|
- text-generation |
|
- conversational |
|
license: apache-2.0 |
|
datasets: |
|
- marmikpandya/mental-health |
|
--- |
|
|
|
# Model Name: Gemma-2-2B-IT |
|
|
|
This is a fine-tuned model based on `Gemma-2-2B-IT`, optimized for text generation tasks. It is compatible with Hugging Face's `transformers` library. |
|
|
|
## Fine-Tuning Details |
|
- Dataset: Mental Health / Therapist Dataset |
|
- Method: LoRA for low-resource adaptation |
|
|
|
## Usage Example |
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model = AutoModelForCausalLM.from_pretrained("YourModelRepoName") |
|
tokenizer = AutoTokenizer.from_pretrained("YourModelRepoName") |
|
|
|
input_text = "What is the meaning of life?" |
|
inputs = tokenizer(input_text, return_tensors="pt") |
|
outputs = model.generate(**inputs) |
|
print(tokenizer.decode(outputs[0])) |