ICEPVP8977's picture
Update README.md
346beef verified
|
raw
history blame
1.45 kB
---
language:
- en
pipeline_tag: text-generation
---
## This model was fine-tuned using a combination of 'uncensored' datasets available on Hugging-Face, as well as the 'Uncensored_mini'.
## In my opinion, this was a waste of time.
### **"too toxic."** I prefer the LLM to maintain a level of respect when addressing the user without being overly limited or censored.
```python
from huggingface_hub import snapshot_download
# Replace with your Hugging Face token (optional but recommended)
# token = ""
# Replace with the repository ID you want to download
repo_id = "ICEPVP8977/Uncensored_llama_3.2_3b_safetensors"
try:
snapshot_download(repo_id=repo_id,
token=token,
local_dir="./model")
print(f"Successfully downloaded {repo_id} to ./model")
except Exception as e:
print(f"Error downloading repository: {e}")
```
```python
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("./model", torch_dtype=torch.float16, device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("./model")
```
```python
prompt = "Your_question_here"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
max_new_tokens = 2000 # Set the maximum number of tokens in the response
outputs = model.generate(**inputs, max_new_tokens=max_new_tokens)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
```