compar:IA: The French Government's LLM arena to collect French-language human prompts and preference data Paper • 2602.06669 • Published 4 days ago • 6
Group-Evolving Agents: Open-Ended Self-Improvement via Experience Sharing Paper • 2602.04837 • Published 6 days ago • 7
QuantLRM: Quantization of Large Reasoning Models via Fine-Tuning Signals Paper • 2602.02581 • Published 10 days ago • 6
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. • 57 items • Updated 1 day ago • 4
ReMiT: RL-Guided Mid-Training for Iterative LLM Evolution Paper • 2602.03075 • Published 7 days ago • 4
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. • 57 items • Updated 1 day ago • 4
RaBiT: Residual-Aware Binarization Training for Accurate and Efficient LLMs Paper • 2602.05367 • Published 5 days ago • 7
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. • 57 items • Updated 1 day ago • 4
Canzona: A Unified, Asynchronous, and Load-Balanced Framework for Distributed Matrix-based Optimizers Paper • 2602.06079 • Published 6 days ago • 16
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. • 57 items • Updated 1 day ago • 4
Beyond Fixed Frames: Dynamic Character-Aligned Speech Tokenization Paper • 2601.23174 • Published 11 days ago • 2
Learning Rate Matters: Vanilla LoRA May Suffice for LLM Fine-tuning Paper • 2602.04998 • Published 6 days ago • 5
Late-to-Early Training: LET LLMs Learn Earlier, So Faster and Better Paper • 2602.05393 • Published 5 days ago • 6 • 3
WTF GENIUS PAPERS Collection Papers that made me appreciate my major and my life a little more. obs=Observation, innov=Innovation. Most papers are abt improving tiny models. • 57 items • Updated 1 day ago • 4
Late-to-Early Training: LET LLMs Learn Earlier, So Faster and Better Paper • 2602.05393 • Published 5 days ago • 6
DASH: Faster Shampoo via Batched Block Preconditioning and Efficient Inverse-Root Solvers Paper • 2602.02016 • Published 8 days ago • 11
HUMAN-WRITTEN & LEGALLY-SOURCED* Collection Datasets written by humans and/or reverse-engineered from text with deterministic algorithms. No illegal scraping or unethical synthesis *...mostly. • 155 items • Updated about 24 hours ago • 2
Privileged Information Distillation for Language Models Paper • 2602.04942 • Published 6 days ago • 23
HUMAN-WRITTEN & LEGALLY-SOURCED* Collection Datasets written by humans and/or reverse-engineered from text with deterministic algorithms. No illegal scraping or unethical synthesis *...mostly. • 155 items • Updated about 24 hours ago • 2