Inference Providers
Active filters: vLLM
mistralai/Mistral-Small-4-119B-2603
119B • Updated • 63.1k
• 344
unsloth/Mistral-Small-4-119B-2603-GGUF
119B • Updated • 37.3k
• 53
QuantTrio/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-v2-AWQ
Image-Text-to-Text
• 28B • Updated • 3.26k
• 9
mistralai/Mistral-Small-4-119B-2603-NVFP4
Updated • 3.29k
• 79
QuantTrio/Qwen3.5-27B-AWQ
Image-Text-to-Text
• 28B • Updated • 400k
• 31
Image-Text-to-Text
• 5B • Updated • 11.9k
• 7
QuantTrio/Qwen3.5-35B-A3B-AWQ
Image-Text-to-Text
• 36B • Updated • 161k
• 15
Image-Text-to-Text
• 10B • Updated • 283k
• 10
mistralai/Mistral-Small-4-119B-2603-eagle
Updated • 329
• 38
QuantTrio/Qwen3.5-122B-A10B-AWQ
Image-Text-to-Text
• 125B • Updated • 27.6k
• 23
mlx-community/Mistral-Small-4-119B-2603-4bit
19B • Updated • 2.3k
• 3
cyankiwi/Mistral-Small-4-119B-2603-AWQ-4bit
21B • Updated • 2.09k
• 6
brandonbeiler/Skywork-R1V3-38B-FP8-Dynamic
Image-Text-to-Text
• 38B • Updated • 10
• 2
QuantTrio/Qwen3-Coder-Next-E336
Text Generation
• 53B • Updated • 9
• 2
QuantTrio/MiniMax-M2.5-AWQ
Text Generation
• 229B • Updated • 61.7k
• 12
QuantTrio/Qwen3.5-397B-A17B-AWQ
Image-Text-to-Text
• Updated • 11.4k
• 8
Text Generation
• 586B • Updated • 3.71k
• 4
Simsema/Simsema_Small-4-119B-32226
119B • Updated • 26
• 1
mradermacher/Simsema_Small-4-119B-32226-i1-GGUF
119B • Updated • 14.9k
• 1
RedHatAI/Mistral-Small-4-119B-2603-NVFP4
Updated • 40
• 1
Xingyu-Zheng/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-INT4-FOEM
Image-Text-to-Text
• 27B • Updated • 1
model-scope/glm-4-9b-chat-GPTQ-Int4
Text Generation
• 9B • Updated • 93
• 6
model-scope/glm-4-9b-chat-GPTQ-Int8
Text Generation
• 9B • Updated • 28
• 2
tclf90/qwen2.5-72b-instruct-gptq-int4
Text Generation
• 73B • Updated • 108
• 2
tclf90/qwen2.5-72b-instruct-gptq-int3
Text Generation
• 69B • Updated • 108
prithivMLmods/Nu2-Lupi-Qwen-14B
Text Generation
• 15B • Updated • 5
• 2
mradermacher/Nu2-Lupi-Qwen-14B-GGUF
15B • Updated • 120
• 1
mradermacher/Nu2-Lupi-Qwen-14B-i1-GGUF
15B • Updated • 55
• 1
JunHowie/Qwen3-0.6B-GPTQ-Int4
Text Generation
• 0.6B • Updated • 124
• 1
JunHowie/Qwen3-0.6B-GPTQ-Int8
Text Generation
• 0.6B • Updated • 8