📦
Vllm Mlx
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.
0 installs
458 stars
59 forks
Trust: 69 — Good
Ai
Installation
npx vllm-mlxAsk AI about Vllm Mlx
Powered by Claude · Grounded in docs
I know everything about Vllm Mlx. Ask me about installation, configuration, usage, or troubleshooting.
0/500
Loading tools...
Reviews
Documentation
No README available
