Edit model card

TAROT-PPO

Task-Oriented Authorship Obfuscation Using Policy Optimization Methods

Fine-tuned text rewriting model with proximal policy optimization for authorship obfuscation.

ArXiv paper: https://arxiv.org/abs/2407.21630v1

Model description

Example use

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("gabrielloiseau/TAROT-PPO")
model = AutoModelForCausalLM.from_pretrained("gabrielloiseau/TAROT-PPO")

paragraph = """I had dinner at Bella's Bistro last night, and it was a delightful experience. 
As soon as I walked in, I was greeted warmly by the hostess, and the cozy, rustic decor made me feel right at home. 
I started with the bruschetta, which was so fresh and flavorful—I could have eaten a whole meal of just that!"""

inputs = tokenizer([paragraph + "<|endoftext|>"], return_tensors="pt", padding=True)
outputs = model.generate(**inputs, do_sample=True, max_new_tokens=128)

outputs = outputs[:, inputs["input_ids"].shape[1]:]
tokenizer.batch_decode(outputs,skip_special_tokens=True)
Downloads last month
29
Safetensors
Model size
355M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for gabrielloiseau/TAROT-PPO

Finetuned
(2)
this model

Dataset used to train gabrielloiseau/TAROT-PPO

Collection including gabrielloiseau/TAROT-PPO