-
-
-
-
-
-
Inference Providers
Active filters: FP8
duydq12/GLM-4-32B-0414-FP8-dynamic
Text Generation
• 33B • Updated
• 3
duydq12/GLM-Z1-32B-0414-FP8-dynamic
Text Generation
• 33B • Updated
• 3
duydq12/nomic-embed-code-FP8-dynamic
Text Generation
• 8B • Updated
• 112
• 1
duydq12/Qwen2.5-Coder-1.5B-Instruct-FP8-dynamic
Text Generation
• 2B • Updated
• 23
duydq12/Qwen2.5-Coder-3B-Instruct-FP8-dynamic
Text Generation
• 3B • Updated
• 1
nvidia/Qwen3-235B-A22B-FP8
Text Generation
• 235B • Updated
• 1.56k
• 3
Image-Text-to-Text
• 109B • Updated
• 1
EliovpAI/Qwen3-14B-FP8-KV
Text Generation
• 15B • Updated
• 3
• 2
clarifai/Qwen3-Coder-30B-A3B-Instruct-FP8-Dynamic
Text Generation
• 31B • Updated
• 15
• 4
EliovpAI/Qwen3-0.6B-FP8-KV
Text Generation
• 0.6B • Updated
• 1
RedHatAI/Devstral-Small-2507-FP8-Dynamic
Text Generation
• 24B • Updated
• 21
• 4
nvidia/Phi-4-reasoning-plus-FP8
15B • Updated
• 515
• 3
EliovpAI/Deepseek-R1-0528-Qwen3-8B-FP8-KV
Text Generation
• 8B • Updated
• 1
Text Generation
• 15B • Updated
• 4.16k
• 4
nvidia/Qwen2.5-VL-7B-Instruct-FP8
Text Generation
• 8B • Updated
• 347
• 7
QuantTrio/Qwen3-VL-235B-A22B-Instruct-FP8
Text Generation
• Updated
• 104
QuantTrio/Qwen3-VL-235B-A22B-Thinking-FP8
Text Generation
• 236B • Updated
• 27
tokenlabsdotrun/Llama-3.1-8B-ModelOpt-FP8-QAT
RedHatAI/Phi-4-reasoning-FP8-dynamic
Text Generation
• 15B • Updated
• 181
• 1