|
--- |
|
library_name: transformers |
|
license: gpl-3.0 |
|
base_model: philippelaban/keep_it_simple |
|
datasets: |
|
- Yelp/yelp_review_full |
|
language: |
|
- en |
|
tags: |
|
- ppo |
|
--- |
|
|
|
# TAROT-PPO |
|
|
|
Task-Oriented Authorship Obfuscation Using Policy Optimization Methods |
|
|
|
Fine-tuned text rewriting model with **proximal policy optimization** for authorship obfuscation. |
|
|
|
ArXiv paper: https://arxiv.org/abs/2407.21630v1 |
|
|
|
## Model description |
|
- **Model type:** Authorship obfuscation model using GPT2-based text rewriting |
|
- **Reward models:** [rrivera1849/LUAR-MUD](https://huggingface.co/rrivera1849/LUAR-MUD) & [Alibaba-NLP/gte-large-en-v1.5](https://huggingface.co/Alibaba-NLP/gte-large-en-v1.5) |
|
- **Finetuned from model:** [philippelaban/keep_it_simple](https://huggingface.co/philippelaban/keep_it_simple) |
|
- **Dataset:** [Yelp/yelp_review_full](https://huggingface.co/datasets/Yelp/yelp_review_full) |
|
- **Repository:** https://github.com/hornetsecurity/tarot |
|
|
|
## Example use |
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("gabrielloiseau/TAROT-PPO") |
|
model = AutoModelForCausalLM.from_pretrained("gabrielloiseau/TAROT-PPO") |
|
|
|
paragraph = """I had dinner at Bella's Bistro last night, and it was a delightful experience. |
|
As soon as I walked in, I was greeted warmly by the hostess, and the cozy, rustic decor made me feel right at home. |
|
I started with the bruschetta, which was so fresh and flavorful—I could have eaten a whole meal of just that!""" |
|
|
|
inputs = tokenizer([paragraph + "<|endoftext|>"], return_tensors="pt", padding=True) |
|
outputs = model.generate(**inputs, do_sample=True, max_new_tokens=128) |
|
|
|
outputs = outputs[:, inputs["input_ids"].shape[1]:] |
|
tokenizer.batch_decode(outputs,skip_special_tokens=True) |
|
``` |