-
-
-
-
-
-
Inference Providers
Active filters:
modelopt
lukealonso/MiniMax-M2.5-NVFP4
130B
•
Updated
•
15.4k
•
31
nvidia/Qwen3.5-397B-A17B-NVFP4
Text Generation
•
Updated
•
1.33k
•
12
lukealonso/MiniMax-M2.5-REAP-139B-A10B-NVFP4
80B
•
Updated
•
2.54k
•
12
vincentzed-hf/Qwen3.5-397B-A17B-NVFP4
Image-Text-to-Text
•
Updated
•
17.8k
•
9
425B
•
Updated
•
6.55k
•
6
nvidia/Qwen3-Next-80B-A3B-Thinking-NVFP4
Text Generation
•
Updated
•
76.5k
•
49
Text Generation
•
8B
•
Updated
•
116
•
4
Text Generation
•
Updated
•
51.5k
•
49
nvidia/Qwen3-Next-80B-A3B-Instruct-NVFP4
Text Generation
•
Updated
•
29.9k
•
27
vincentzed-hf/Qwen3-Coder-Next-NVFP4
Text Generation
•
Updated
•
5.52k
•
6
nvidia/Llama-4-Scout-17B-16E-Instruct-NVFP4
56B
•
Updated
•
12.2k
•
21
nvidia/Llama-3.1-8B-Instruct-NVFP4
5B
•
Updated
•
106k
•
7
Text Generation
•
15B
•
Updated
•
3.32k
•
4
shanjiaz/gpt-oss-120b-nvfp4-modelopt
59B
•
Updated
•
9.12k
•
3
nvidia/Kimi-K2-Thinking-NVFP4
Text Generation
•
Updated
•
125k
•
27
Text Generation
•
177B
•
Updated
•
3.65k
•
16
nvidia/Qwen3-235B-A22B-Thinking-2507-NVFP4
Text Generation
•
Updated
•
635
•
4
nvidia/Qwen3-235B-A22B-Instruct-2507-NVFP4
Text Generation
•
120B
•
Updated
•
1.91k
•
3
nvidia/Qwen3-Coder-480B-A35B-Instruct-NVFP4
Text Generation
•
241B
•
Updated
•
414
•
2
BenChaliah/Gemma3-27B-it-NVFP4
15B
•
Updated
•
10
•
3
baseten-admin/Kimi-2.5-text-nvfp4-v6-mlp-only
Updated
•
1.69k
•
1
nvidia/Llama-4-Maverick-17B-128E-Instruct-FP8
402B
•
Updated
•
606
•
12
nvidia/Llama-4-Scout-17B-16E-Instruct-FP8
109B
•
Updated
•
38.8k
•
11
ishan24/test_modelopt_quant
nvidia/Llama-4-Maverick-17B-128E-Eagle3
Updated
•
10
•
9
nvidia/Qwen3-30B-A3B-NVFP4
Text Generation
•
16B
•
Updated
•
54.9k
•
23
jiangchengchengNLP/L3.3-MS-Nevoria-70b-FP8
Text Generation
•
71B
•
Updated
•
6
NVFP4/Qwen3-30B-A3B-Instruct-2507-FP4
Text Generation
•
16B
•
Updated
•
1.14k
•
11
NVFP4/Qwen3-Coder-30B-A3B-Instruct-FP4
Text Generation
•
16B
•
Updated
•
21.2k
•
7
gesong2077/Qwen3-32B-NVFP4
19B
•
Updated
•
1