| --- |
| license: mit |
| datasets: |
| - saiteja33/DAMASHA |
| language: |
| - en |
| base_model: |
| - FacebookAI/roberta-base |
| - answerdotai/ModernBERT-base |
| pipeline_tag: token-classification |
| --- |
| |
| # DAMASHA-MAS: Mixed-Authorship Adversarial Segmentation (Token Classification) |
|
|
| This repository contains a **token-classification model** trained on the **DAMASHA-MAS** benchmark, introduced in: |
|
|
| > **DAMASHA: Detecting AI in Mixed Adversarial Texts via Segmentation with Human-interpretable Attribution** |
|
|
| The model aims to **segment mixed human–AI text** at *token level* – i.e., decide for each token whether it was written by a *human* or an *LLM*, even under **syntactic adversarial attacks**. |
|
|
| - **Base encoders:** |
| - [`FacebookAI/roberta-base`](https://huggingface.co/FacebookAI/roberta-base) |
| - [`answerdotai/ModernBERT-base`](https://huggingface.co/answerdotai/ModernBERT-base) |
| - **Architecture (high level):** RoBERTa + ModernBERT feature fusion → BiGRU + CRF with the **Info-Mask** gating mechanism from the paper. |
| - **Task:** Token classification (binary authorship: human vs AI). |
| - **Language:** English |
| - **License (this model):** MIT |
| - **Training data license:** CC-BY-4.0 via the DAMASHA dataset. |
|
|
| If you use this model, **please also cite the DAMASHA paper and dataset** (see Citation section). |
|
|
| --- |
|
|
| ## 1. Model Highlights |
|
|
| - **Fine-grained mixed-authorship detection** |
| Predicts authorship **per token**, allowing reconstruction of human vs AI **spans** in long documents. |
|
|
| - **Adversarially robust** |
| Trained and evaluated on **syntactically attacked texts** (misspelling, Unicode substitutions, invisible characters, punctuation swaps, case perturbations, and “all-mixed” attacks). |
|
|
| - **Human-interpretable Info-Mask** |
| The architecture incorporates **stylometric features** (perplexity, POS density, punctuation density, lexical diversity, readability) via an **Info-Mask** module that gates token representations in an interpretable way. |
|
|
| - **Strong reported performance (from the paper)** |
| On DAMASHA-MAS, the **RMC\*** model (RoBERTa + ModernBERT + CRF + Info-Mask) achieves: |
| - **Token-level**: Accuracy / Precision / Recall / F1 ≈ **0.98** |
| - **Span-level (strict)**: SBDA ≈ **0.45**, SegPre ≈ **0.41** |
| - **Span-level (relaxed IoU ≥ 0.5)**: ≈ **0.82** |
|
|
| > ⚠️ The exact numbers for *this* specific checkpoint may differ depending on training run and configuration. The values above are from the paper’s best configuration (RMC\*). |
| |
| --- |
| |
| ## 2. Intended Use |
| |
| ### What this model is for |
| |
| - **Research on human–AI co-authorship** |
| - Studying where LLMs “take over” in mixed texts. |
| - Analysing robustness of detectors under adversarial perturbations. |
| |
| - **Tooling / applications (with human oversight)** |
| - Assisting editors, educators, or moderators to **highlight suspicious spans** rather than making final decisions. |
| - Exploring **interpretability overlays** (e.g., heatmaps over tokens) when combined with Info-Mask outputs. |
| |
| ### What this model is *not* for |
|
|
| - Automated “cheating detector” / plagiarism court. |
| - High-stakes decisions affecting people’s livelihood, grades, or reputation **without human review**. |
| - Non-English or heavily code-mixed text (training data is English-centric). |
|
|
| Use this model as a **signal**, not a judge. |
|
|
| --- |
|
|
| ## 3. Data: DAMASHA-MAS |
|
|
| The model is trained on the **MAS** benchmark released with the DAMASHA paper and hosted as the Hugging Face dataset: |
|
|
| - **Dataset:** [`saiteja33/DAMASHA`](https://huggingface.co/datasets/saiteja33/DAMASHA) |
|
|
| ### 3.1 What’s in MAS? |
|
|
| MAS consists of **mixed human–AI texts with explicit span tags**: |
|
|
| - Human text comes from several corpora for **domain diversity**, including: |
| - Reddit (M4-Reddit) |
| - Yelp & /r/ChangeMyView (MAGE-YELP, MAGE-CMV) |
| - News summaries (XSUM) |
| - Wikipedia (M4-Wiki, MAGE-SQuAD) |
| - ArXiv abstracts (MAGE-SciGen) |
| - QA texts (MAGE-ELI5) |
|
|
| - AI text is generated by multiple modern LLMs: |
| - **DeepSeek-V3-671B** (open-source) |
| - **GPT-4o, GPT-4.1, GPT-4.1-mini** (closed-source) |
|
|
| ### 3.2 Span tagging |
|
|
| Authorship is marked using **explicit tags** around AI spans: |
|
|
| - `<AI_Start>` … `</AI_End>` denote AI-generated segments within otherwise human text. |
| - The dataset stores text in a `hybrid_text` column, plus metadata such as `has_pair`, and adversarial variants include `attack_name`, `tag_count`, and `attacked_text`. |
| - Tags are sentence-level in annotation, but the model is trained to output **token-level** predictions for finer segmentation. |
|
|
| > During training, these tags are converted into **token labels** (2 labels total; see `config.id2label` in the model files). |
|
|
| ### 3.3 Adversarial attacks |
|
|
| MAS includes multiple **syntactic attacks** applied to the mixed text: |
|
|
| - Misspelling |
| - Unicode character substitution |
| - Invisible characters |
| - Punctuation substitution |
| - Upper/lower case swapping |
| - All-mixed combinations of the above |
|
|
| These perturbations make tokenization brittle and test robustness of detectors in realistic settings. |
|
|
| --- |
|
|
| ## 4. Model Architecture & Training |
|
|
| ### 4.1 Architecture (conceptual) |
|
|
| The model follows the **Info-Mask RMC\*** architecture described in the DAMASHA paper: |
|
|
| 1. **Dual encoders** |
| - RoBERTa-base and ModernBERT-base encode the same input sequence. |
| 2. **Feature fusion** |
| - Hidden states from both encoders are fused into a shared representation. |
| 3. **Stylometric Info-Mask** |
| - Hand-crafted style features (perplexity, POS density, punctuation density, lexical diversity, readability) are projected, passed through multi-head attention, and turned into a **scalar mask per token**. |
| - This mask gates the fused encoder states, down-weighting style-irrelevant tokens and emphasizing style-diagnostic ones. :contentReference[oaicite:16]{index=16} |
| 4. **Sequence model + CRF** |
| - A BiGRU layer captures sequential dependencies, followed by a **CRF** layer for structured token labeling with a sequence-level loss. :contentReference[oaicite:17]{index=17} |
|
|
| ### 4.2 Training setup (from the paper) |
|
|
| Key hyperparameters used for the Info-Mask models on MAS: |
|
|
| - **Number of labels:** 2 |
| - **Max sequence length:** 512 |
| - **Batch size:** 64 |
| - **Epochs:** 5 |
| - **Optimizer:** AdamW (with cosine annealing LR schedule) |
| - **Weight decay:** 0.01 |
| - **Gradient clipping:** 1.0 |
| - **Dropout:** Dynamic 0.1–0.3 (initial 0.1) |
| - **Warmup ratio:** 0.1 |
| - **Early stopping patience:** 2 |
|
|
| **Hardware & compute** (as reported): |
|
|
| - AWS EC2 g6e.xlarge, NVIDIA L40S (48GB) GPU, Ubuntu 24.04 |
| - ≈ 400 GPU hours for experiments. |
|
|
| > The exact training script used for this checkpoint is available in the project GitHub: |
| > <https://github.com/saitejalekkala33/DAMASHA> |
|
|
| --- |
|
|
| --- |
| license: mit |
| --- |