--- library_name: transformers tags: - mdlm - diffusion license: apache-2.0 datasets: - turkish-nlp-suite/InstrucTurca language: - tr base_model: - diffutron/DiffutronLM-0.3B-Alpaca pipeline_tag: text-generation --- # DiffutronLM-0.3B-Instruct **Diffutron** is a parameter-efficient, Masked Diffusion Language Model (MDLM) specifically designed for the Turkish language. Unlike standard autoregressive models that generate text one token at a time, Diffutron generates text by iteratively refining sequences in parallel, allowing for simultaneous consideration of the entire sentence context. Despite its compact size of 307 million parameters, `DiffutronLM-0.3B-Instruct` achieves highly competitive performance against much larger, multi-billion-parameter autoregressive baselines on Turkish NLP benchmarks. ## 📌 Model Details * **Model Type:** Masked Diffusion Language Model (MDLM) * **Base Architecture:** `jhu-clsp/mmBERT-base` (Multilingual Encoder) * **Language:** Turkish * **Parameter Count:** 307M (0.3B) * **Context Length:** 256 tokens (Instruct), 512 tokens (Base) * **Training Libraries:** `dllm`, PyTorch ## 🚀 Architecture & Approach Diffutron departs from traditional next-token prediction. It treats text generation as a discrete diffusion process: 1. **Forward Process:** Clean text is gradually corrupted into a sequence of `` tokens. 2. **Reverse Process:** The model learns to denoise the sequence globally, attending to visible context bi-directionally to predict the original tokens. This non-autoregressive paradigm compresses linguistic knowledge efficiently, allowing this 0.3B model to punch significantly above its weight class. ## 📚 Training Pipeline The model was developed through a resource-efficient, multi-stage training pipeline: ### 1. Continual Pre-training (CPT) To adapt the multilingual backbone to Turkish without catastrophic forgetting, we employed a high-rank LoRA strategy (r=256, α=256) targeting all linear modules (Attention and MLP). * **Data:** ~2 million sequences sourced from Havadis (news), Temiz-OSCAR (web), and Turkish Wikipedia. * **Result:** Perplexity on the Bilkent Turkish Writings Dataset dropped significantly from 3.42 (base) to 2.75. ### 2. Progressive Instruction-Tuning (SFT) To unlock generative instruction-following capabilities, we utilized a two-stage supervised fine-tuning approach: * **Stage 1 (General Adaptation):** Trained on `metunlp/LlamaTurk-Instruction-Set` for 20 epochs to establish fundamental instruction-following behaviors. * **Stage 2 (Complex Specialization):** Trained on the nuanced `turkish-nlp-suite/InstrucTurca` dataset for 8 epochs with an increased batch size, enhancing the model's ability to handle intricate, domain-specific Turkish commands. ## 📊 Evaluation Results The model was evaluated on a representative subset of the **CETVEL Benchmark Suite**. DiffutronLM-0.3B (2nd Stage) demonstrates remarkable parameter efficiency, outperforming models up to 7x its size (e.g., Kumru-2B and TURNA-1.1B) on average scores. | Benchmark | Diffutron-1st-Stage (0.3B) | Diffutron-2nd-Stage (0.3B) | TURNA (1.1B) | Kumru (2B) | Kanarya (2B) | Llama-3.2 (3B) | Trendyol (7B) | Aya-101 (13B) | | :--- | :---: | :---: | :---: | :---: | :---: | :---: | :---: | :---: | | **Belebele_TR** | 22.22 | 27.00 | 22.56 | 29.00 | 28.11 | **55.78** | 36.22 | 22.89 | | **EXAMS_TR** | 25.95 | 27.74 | 23.66 | **30.03** | **30.03** | 26.21 | 28.50 | 22.90 | | **IronyTR** | 50.67 | **52.00** | 48.33 | 51.00 | 50.00 | 50.17 | 50.00 | **52.17** | | **News_Cat** | 23.20 | 32.40 | 32.80 | 26.40 | 66.80 | 64.00 | **81.20** | 20.00 | | **MNLI_TR** | 33.29 | 32.81 | 34.94 | **36.42** | 33.40 | 34.76 | 35.19 | 27.90 | | **STS_TR** | 17.77 | **18.78** | 14.21 | 11.75 | 12.91 | 12.91 | 15.52 | 16.97 | | **XCOPA_TR** | 53.80 | 52.00 | 55.80 | 54.00 | **64.20** | 54.60 | 61.00 | 59.60 | | **Average** | 32.41 | **34.68** | 33.19 | 34.09 | 40.78 | 42.63 | **43.95** | 31.78 | ## 💻 Usage Because Diffutron is a Masked Diffusion Language Model, it requires inference strategies distinct from standard causal generation. We recommend using the `dllm` library or custom generation loops tailored for discrete diffusion. ### Generation Parameters Used in Paper: * **Longer Context:** Steps: 128, Temp: 0.1, Block Length: 32, Repetition Penalty: 1.2 * **Shorter Context:** Steps: 64, Remask: `low_conf`, Stochastic: `False`, CFG: 0.0 ## ⚠️ Limitations * **Multilingual Backbone:** Built upon a multilingual encoder rather than a native Turkish foundation model. * **Context Window:** Restricted to a 256-token context window for generation, limiting its use in long-form summarization or document-level generation. * **Data Nuances:** Instruction datasets rely heavily on translations or synthetic data, which may occasionally miss subtle cultural contexts. ## 📝 Citation If you use Diffutron in your research, please cite our preprint: ```bibtex @misc{diffutron2026, author = {Kocabay, Şuayp Talha and Akkuş, Talha Rüzgar}, title = {Diffutron: A Masked Diffusion Language Model for Turkish Language}, year = {2026}, publisher = {Hugging Face}, howpublished = {\url{https://huggingface.co/collections/diffutron/diffutronlm}} } ```