rohitnagareddy's picture
Update README.md
be5f3c7 verified
---
license: apache-2.0
base_model: google/gemma-2b-it
tags:
- text-to-lora
- sakana-ai
- peft
- lora
- python
- code-generation
- programming
library_name: peft
---
# gemma-2b-python-expert-lora(Text to Model)
This LoRA adapter specializes the base model for expert-level Python programming. Created using Sakana AI's Text-to-LoRA technology.
## Model Details
- **Base Model**: `google/gemma-2b-it`
- **LoRA Rank**: 16
- **Target Modules**: q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
- **Task**: Python Code Generation
## Usage
```python
from peft import PeftModel
from transformers import AutoModelForCausalLM, AutoTokenizer
# Load base model and tokenizer
model = AutoModelForCausalLM.from_pretrained("google/gemma-2b-it")
tokenizer = AutoTokenizer.from_pretrained("google/gemma-2b-it")
# Load LoRA adapter
model = PeftModel.from_pretrained(model, "rohitnagareddy/gemma-2b-python-expert-lora")
# Generate Python code
prompt = "Write a Python function to implement binary search:"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=256)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Capabilities
- Clean, documented Python code
- Type hints and error handling
- PEP 8 compliance
- Algorithm implementation
- Web development
- Data processing
- Testing and debugging
## Citation
```bibtex
@misc{sakana2024texttolora,
title={Text-to-LoRA},
author={Sakana AI},
year={2024},
url={https://github.com/SakanaAI/text-to-lora}
}
```