|
--- |
|
language: en |
|
license: mit |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
tags: |
|
- text-generation |
|
- ai-detection |
|
- paraphrasing |
|
- originality |
|
- privacy |
|
datasets: |
|
- checkgpt |
|
base_model: Qwen/Qwen2.5-3B-Instruct |
|
model_type: causal-lm |
|
--- |
|
|
|
# AuthorMist Originality |
|
|
|
[](https://huggingface.co/authormist/originality) |
|
[](https://opensource.org/licenses/MIT) |
|
|
|
## Overview |
|
|
|
AuthorMist Originality is a specialized language model designed to transform AI-generated text into more human-like writing while preserving the original meaning. This model was developed using reinforcement learning techniques to specifically evade AI text detection systems, with a focus on Originality.ai's detection algorithms. |
|
|
|
The model is based on Qwen2.5-3B Instruct and has been fine-tuned using Group Relative Policy Optimization (GRPO) with detector feedback as a reward signal. AuthorMist Originality demonstrates strong performance in reducing detectability across multiple AI text detection systems while maintaining high semantic similarity with the original text. |
|
|
|
## Key Features |
|
|
|
- **Detector Evasion**: Trained specifically to evade Originality.ai's detection algorithms, with strong cross-detector generalization |
|
- **Meaning Preservation**: Maintains high semantic similarity (>0.94) with the original text |
|
- **Natural Output**: Produces fluent, coherent text that reads naturally |
|
- **Broad Applicability**: Effective across various domains including academic, technical, and creative writing |
|
|
|
## Model Details |
|
|
|
- **Base Model**: Qwen2.5-3B Instruct |
|
- **Training Method**: Reinforcement Learning with Group Relative Policy Optimization (GRPO) |
|
- **Training Data**: 10,000 human-written abstracts from the CheckGPT dataset with corresponding AI-generated versions |
|
- **Domains Covered**: Computer Science, Humanities, Social Sciences, Physics, and more |
|
- **Text Length Support**: Optimized for texts ranging from 100 to 500 words |
|
|
|
## Performance |
|
|
|
AuthorMist Originality demonstrates exceptional performance in evading AI text detection: |
|
|
|
- **Mean AUROC**: 0.49 across six major detection systems |
|
- **Mean F1-score**: 0.09 across all tested detectors |
|
- **Semantic Similarity**: >0.94 with original text |
|
|
|
The model shows particularly strong performance against: |
|
- Hello SimpleAI (AUROC: 0.07) |
|
- Sapling (AUROC: 0.13) |
|
- Winston.ai (AUROC: 0.35) |
|
|
|
## Usage |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
# Load model and tokenizer |
|
model_name = "authormist/authormist-originality" |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
model = AutoModelForCausalLM.from_pretrained(model_name) |
|
|
|
# Prepare input text |
|
ai_text = "Your AI-generated text here..." |
|
prompt = f"""Please paraphrase the following text to make it more human-like while preserving the original meaning: |
|
|
|
{ai_text} |
|
|
|
Paraphrased text:""" |
|
|
|
# Generate paraphrased text |
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
outputs = model.generate( |
|
inputs.input_ids, |
|
max_new_tokens=512, |
|
temperature=0.7, |
|
top_p=0.9, |
|
do_sample=True |
|
) |
|
paraphrased_text = tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
print(paraphrased_text.split("Paraphrased text:")[1].strip()) |
|
``` |
|
|
|
## Ethical Considerations |
|
|
|
AuthorMist Originality is released for research purposes to advance understanding of AI text detection limitations and privacy-preserving technologies. We acknowledge the dual-use nature of this technology and emphasize the following ethical considerations: |
|
|
|
1. **Academic Integrity**: This model should not be used to misrepresent AI-generated content as human-written in academic settings where such distinctions are ethically relevant. |
|
|
|
2. **Transparency**: We encourage users to maintain transparency about the use of AI assistance in content creation, even when using privacy-enhancing tools like AuthorMist. |
|
|
|
3. **Privacy Protection**: The primary legitimate use case for this technology is protecting author privacy and preventing unfair discrimination against AI-assisted writing in contexts where such assistance is permissible. |
|
|
|
4. **Research Value**: This model provides valuable insights into the limitations of current AI detection systems and contributes to the ongoing research dialogue about AI text detection and privacy. |
|
|
|
## Citation |
|
|
|
If you use AuthorMist Originality in your research, please cite our paper: |
|
|
|
```bibtex |
|
@article{authormist2025, |
|
title={AuthorMist: Evading AI Text Detectors with Reinforcement Learning}, |
|
author={David, Isaac and Gervais, Arthur}, |
|
journal={arXiv preprint}, |
|
year={2025} |
|
} |
|
``` |
|
|
|
## License |
|
|
|
This model is released under the [MIT License](https://opensource.org/licenses/MIT). |
|
|
|
## Acknowledgments |
|
|
|
We thank the developers of Qwen2.5 for the base model and the creators of the CheckGPT dataset for providing valuable training data. |