AuthorMist Originality

Hugging Face License

Overview

AuthorMist Originality is a specialized language model designed to transform AI-generated text into more human-like writing while preserving the original meaning. This model was developed using reinforcement learning techniques to specifically evade AI text detection systems, with a focus on Originality.ai's detection algorithms.

The model is based on Qwen2.5-3B Instruct and has been fine-tuned using Group Relative Policy Optimization (GRPO) with detector feedback as a reward signal. AuthorMist Originality demonstrates strong performance in reducing detectability across multiple AI text detection systems while maintaining high semantic similarity with the original text.

Key Features

  • Detector Evasion: Trained specifically to evade Originality.ai's detection algorithms, with strong cross-detector generalization
  • Meaning Preservation: Maintains high semantic similarity (>0.94) with the original text
  • Natural Output: Produces fluent, coherent text that reads naturally
  • Broad Applicability: Effective across various domains including academic, technical, and creative writing

Model Details

  • Base Model: Qwen2.5-3B Instruct
  • Training Method: Reinforcement Learning with Group Relative Policy Optimization (GRPO)
  • Training Data: 10,000 human-written abstracts from the CheckGPT dataset with corresponding AI-generated versions
  • Domains Covered: Computer Science, Humanities, Social Sciences, Physics, and more
  • Text Length Support: Optimized for texts ranging from 100 to 500 words

Performance

AuthorMist Originality demonstrates exceptional performance in evading AI text detection:

  • Mean AUROC: 0.49 across six major detection systems
  • Mean F1-score: 0.09 across all tested detectors
  • Semantic Similarity: >0.94 with original text

The model shows particularly strong performance against:

  • Hello SimpleAI (AUROC: 0.07)
  • Sapling (AUROC: 0.13)
  • Winston.ai (AUROC: 0.35)

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

# Load model and tokenizer
model_name = "authormist/authormist-originality"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Prepare input text
ai_text = "Your AI-generated text here..."
prompt = f"""Please paraphrase the following text to make it more human-like while preserving the original meaning:

{ai_text}

Paraphrased text:"""

# Generate paraphrased text
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
    inputs.input_ids,
    max_new_tokens=512,
    temperature=0.7,
    top_p=0.9,
    do_sample=True
)
paraphrased_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(paraphrased_text.split("Paraphrased text:")[1].strip())

Ethical Considerations

AuthorMist Originality is released for research purposes to advance understanding of AI text detection limitations and privacy-preserving technologies. We acknowledge the dual-use nature of this technology and emphasize the following ethical considerations:

  1. Academic Integrity: This model should not be used to misrepresent AI-generated content as human-written in academic settings where such distinctions are ethically relevant.

  2. Transparency: We encourage users to maintain transparency about the use of AI assistance in content creation, even when using privacy-enhancing tools like AuthorMist.

  3. Privacy Protection: The primary legitimate use case for this technology is protecting author privacy and preventing unfair discrimination against AI-assisted writing in contexts where such assistance is permissible.

  4. Research Value: This model provides valuable insights into the limitations of current AI detection systems and contributes to the ongoing research dialogue about AI text detection and privacy.

Citation

If you use AuthorMist Originality in your research, please cite our paper:

@article{authormist2025,
  title={AuthorMist: Evading AI Text Detectors with Reinforcement Learning},
  author={David, Isaac and Gervais, Arthur},
  journal={arXiv preprint},
  year={2025}
}

License

This model is released under the MIT License.

Acknowledgments

We thank the developers of Qwen2.5 for the base model and the creators of the CheckGPT dataset for providing valuable training data.

Downloads last month
71
Safetensors
Model size
3.09B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for authormist/authormist-originality

Base model

Qwen/Qwen2.5-3B
Finetuned
(322)
this model