-
-
-
-
-
-
Inference Providers
Active filters:
8-bit
GadflyII/GLM-4.7-Flash-NVFP4
Text Generation
•
18B
•
Updated
•
153k
•
38
Text Generation
•
120B
•
Updated
•
3.01M
•
•
4.38k
Text Generation
•
22B
•
Updated
•
6.66M
•
•
4.24k
mlx-community/GLM-4.7-Flash-8bit
Text Generation
•
30B
•
Updated
•
4.36k
•
15
mlx-community/Qwen3-TTS-12Hz-0.6B-CustomVoice-8bit
Text-to-Speech
•
0.3B
•
Updated
•
770
•
7
microsoft/bitnet-b1.58-2B-4T
Text Generation
•
0.8B
•
Updated
•
5.95k
•
1.26k
MultiverseComputingCAI/HyperNova-60B
Text Generation
•
60B
•
Updated
•
1.42k
•
48
mlx-community/GLM-4.7-Flash-8bit-gs32
Text Generation
•
30B
•
Updated
•
374
•
5
AlicanKiraz0/Mihenk-LLM-14B-Turkish-Financial-Model-mlx-8Bit
15B
•
Updated
•
26
•
6
NVFP4/Qwen3-Coder-30B-A3B-Instruct-FP4
Text Generation
•
16B
•
Updated
•
3.31k
•
6
openai/gpt-oss-safeguard-20b
Text Generation
•
22B
•
Updated
•
11k
•
•
181
Text Generation
•
177B
•
Updated
•
4.92k
•
10
nvidia/DeepSeek-V3.2-NVFP4
Text Generation
•
394B
•
Updated
•
947
•
3
LiquidAI/LFM2.5-1.2B-Thinking-MLX-8bit
Text Generation
•
0.3B
•
Updated
•
168
•
3
lmstudio-community/GLM-4.7-Flash-MLX-8bit
Text Generation
•
30B
•
Updated
•
244k
•
3
MaziyarPanahi/Mistral-7B-Instruct-Aya-101-GGUF
Text Generation
•
7B
•
Updated
•
219
•
12
ragraph-ai/stable-cypher-instruct-3b
Text Generation
•
3B
•
Updated
•
355
•
31
MaziyarPanahi/Qwen2.5-1.5B-Instruct-GGUF
Text Generation
•
2B
•
Updated
•
145k
•
9
tiiuae/Falcon-E-3B-Instruct
Text Generation
•
0.9B
•
Updated
•
293
•
36
drwlf/medgemma-4b-it-abliterated
Text Generation
•
Updated
•
15
•
6
nvidia/Qwen3-30B-A3B-NVFP4
Text Generation
•
16B
•
Updated
•
32.8k
•
21
Text Generation
•
5B
•
Updated
•
4.97k
•
12
nvidia/Qwen2.5-VL-7B-Instruct-NVFP4
Text Generation
•
5B
•
Updated
•
3.13k
•
12
FabioSarracino/VibeVoice-Large-Q8
Text-to-Audio
•
9B
•
Updated
•
2.69k
•
78
Firworks/NVIDIA-Nemotron-3-Nano-30B-A3B-nvfp4
18B
•
Updated
•
2.01k
•
7
mlx-community/GLM-4.7-8bit
Text Generation
•
353B
•
Updated
•
1.21k
•
4
Tengyunw/MiniMax-M2.1-NVFP4
Text Generation
•
115B
•
Updated
•
183
•
6
mlx-community/translategemma-27b-it-8bit
Text Generation
•
27B
•
Updated
•
964
•
3
nightmedia/Qwen3-32B-Element5-Heretic-qx86-hi-mlx
Text Generation
•
33B
•
Updated
•
212
•
2
arcee-ai/Trinity-Nano-Preview-MLX-8bit
Text Generation
•
6B
•
Updated
•
64
•
2