-
-
-
-
-
-
Inference Providers
Active filters: fp8
Qwen/Qwen3-VL-4B-Instruct-FP8
Image-Text-to-Text
• 5B • Updated
• 29.2k
• 50
Qwen/Qwen3-VL-4B-Thinking-FP8
Image-Text-to-Text
• 5B • Updated
• 2.71k
• 30
Qwen/Qwen3-VL-32B-Instruct-FP8
Image-Text-to-Text
• Updated
• 403k
• 43
Qwen/Qwen3-VL-2B-Instruct-FP8
Image-Text-to-Text
• 2B • Updated
• 182k
• 38
Qwen/Qwen3-VL-2B-Thinking-FP8
Image-Text-to-Text
• 2B • Updated
• 1.27k
• 26
RedHatAI/granite-4.0-h-small-FP8-dynamic
Text Generation
• 32B • Updated
• 718
• 2
ai-sage/GigaChat3-10B-A1.8B
Text Generation
• 11B • Updated
• 3.52k
• 64
Text-to-Image
• Updated
• 13.9k
• 40
XiaomiMiMo/MiMo-V2-Flash-Base
Text Generation
• 310B • Updated
• 114
• 44
RedHatAI/Ministral-3-14B-Instruct-2512
14B • Updated
• 332
• 2
sh0ck0r/Llama-3.3-70B-Vulpecula-r1-FP8-Dynamic
Text Generation
• 71B • Updated
• 5
• 1
Ilus-AI/Qwen-Image-2512-FP8
Text-to-Image
• Updated
• 58
• 1
YifeiDevs/Huihui-Qwen3-8B-abliterated-v2-FP8-Comfy
Ex0bit/MiniMax-M2.5-PRISM-PRO-Tensors
229B • Updated
• 16
• 2
alexliap/Qwen3-VL-Embedding-2B-FP8-DYNAMIC
Feature Extraction
• 2B • Updated
• 12.5k
• 2
mratsim/MiniMax-M2.5-FP8-INT4-AWQ
Text Generation
• 39B • Updated
• 5.56k
• 9
EliasOenal/MiniMax-M2.5-Hybrid-AWQ-W4A16G128-Attn-fp8_e4m3-KV-fp8_e4m3
Text Generation
• 34B • Updated
• 517
• 11
RedHatAI/Qwen3.5-397B-A17B-FP8-dynamic
397B • Updated
• 2.87k
• 3
prithivMLmods/Qwen-Image-Edit-AIO-FP8
Image-to-Image
• Updated
• 232
• 4
unsloth/Qwen3.5-397B-A17B-FP8
Image-Text-to-Text
• 403B • Updated
• 1.02k
• 2
prithivMLmods/Spatial-SSRL-Qwen3VL-4B-FP8
Image-Text-to-Text
• 4B • Updated
• 13
• 1
Text Generation
• 229B • Updated
• 33
• 1
lovedheart/Qwen3.5-26B-A3B-FP8
Image-Text-to-Text
• 28B • Updated
• 37
• 1
FriendliAI/Meta-Llama-3-8B-Instruct-fp8
Text Generation
• 8B • Updated
• 25
• 2
RedHatAI/Meta-Llama-3-8B-Instruct-FP8
Text Generation
• Updated
• 6.22k
• • 24
RedHatAI/Mixtral-8x7B-Instruct-v0.1-AutoFP8
Text Generation
• 47B • Updated
• 7
• 3
anyisalin/Meta-Llama-3-8B-Instruct-FP8
Text Generation
• 8B • Updated
• 6
anyisalin/Meta-Llama-3-8B-Instruct-FP8-D
Text Generation
• 8B • Updated
• 4
anyisalin/lzlv_70b_fp16_hf-FP8-D
Text Generation
• 69B • Updated
• 1
anyisalin/Meta-Llama-3-70B-Instruct-FP8-D
Text Generation
• 71B • Updated
• 4