INT4 LLMs for vLLM Collection Accurate INT4 quantized models by Neural Magic, ready for use with vLLM! β’ 16 items β’ Updated Mar 2 β’ 12
Running Featured 1.04k Can You Run It? LLM version π 1.04k Calculate GPU needs for running LLMs on your hardware
meta-llama/Meta-Llama-3-8B-Instruct Text Generation β’ 8B β’ Updated Jun 18, 2025 β’ 1.33M β’ β’ 4.46k
swtb/XLM-RoBERTa-Base-Conll2003-English-NER-Finetune-FP16-BinaryClass-WeightedLoss Token Classification β’ 0.3B β’ Updated Jun 1, 2024 β’ 1
swtb/XLM-RoBERTa-Base-Conll2003-English-NER-Finetune-BinaryClass-WeightedLoss Token Classification β’ 0.3B β’ Updated Jun 1, 2024