MLX-GUI v1.2.4 Private AI API for Apple Silicon (Chat,Audio,Embeddings,Vision)

3 RamboRogers 2 7/23/2025, 6:39:40 PM mlxgui.com ↗

Comments (2)

RamboRogers · 7h ago
Hey HN! I've been building MLX-GUI as an open-source inference server that turns any Mac into a multi-user AI server. v1.2.4 just shipped with some major additions:

Complete Whisper ecosystem (99+ languages, word timestamps, any audio format) 23 embedding models across 13 families (E5, ModernBERT, Arctic, etc.) Mistral Small 24B with vision capabilities OpenAI-compatible API that's actually faster than Ollama on Apple Silicon

The goal was simple: I wanted to use my Mac Mini/Studio as proper inference servers without the complexity of managing Python environments or paying for cloud APIs while keeping data local. It's packaged as a native macOS app (no Python install needed) with a beautiful web GUI for model management. The API is drop-in compatible with OpenAI, so existing apps like Jan.ai work immediately. 900+ lines of tests ensure production reliability. G

NU GPL v3 licensed and actively maintained. GitHub: https://github.com/RamboRogers/mlx-gui

Would love feedback from the community - especially on the embedding pipeline and audio processing!

jki275 · 7h ago
It looks pretty cool. How is it different or better than LM Studio?