High-performance LLM inference on Apple Silicon using MLX and vLLM
vLLM Metal is a plugin that enables vLLM to run on Apple Silicon Macs using MLX as the primary compute backend. It unifies MLX and PyTorch under a single lowering path.
- MLX-accelerated inference: faster than PyTorch MPS on Apple Silicon
- Unified memory: True zero-copy operations leveraging Apple Silicon's unified memory architecture
- vLLM compatibility: Full integration with vLLM's engine, scheduler, and OpenAI-compatible API
- Paged attention (experimental): Efficient KV cache management for long sequences — opt-in via
VLLM_METAL_USE_PAGED_ATTENTION=1(requirespip install 'vllm-metal[paged]'); default path uses MLX-managed KV cache - GQA support: Grouped-Query Attention for efficient inference
- macOS on Apple Silicon
curl -fsSL https://raw.githubusercontent.com/vllm-project/vllm-metal/main/install.sh | bashIf any issues occur, please use the following command to switch to the latest release version and check if the problem is resolved. If the issue continues to occur in the latest release, please report the details of the issue.
rm -rf ~/.venv-vllm-metal && curl -fsSL https://raw.githubusercontent.com/vllm-project/vllm-metal/main/install.sh | bash┌─────────────────────────────────────────────────────────────┐
│ vLLM Core │
│ Engine, Scheduler, API Server, Tokenizers │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ vllm_metal Plugin Layer │
│ ┌─────────────────┐ ┌────────────────┐ ┌──────────────────┐ │
│ │ MetalPlatform │ │ MetalWorker │ │ MetalModelRunner │ │
│ │ (Platform) │ │ (Worker) │ │ (ModelRunner) │ │
│ └─────────────────┘ └────────────────┘ └──────────────────┘ │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ Unified Compute Backend │
│ ┌───────────────────────────┐ ┌───────────────────────────┐ │
│ │ MLX Backend │ │ PyTorch Backend │ │
│ │ (Primary) │ │ (Model Loading/Interop) │ │
│ │ │ │ │ │
│ │ • SDPA Attention │ │ • HuggingFace Loading │ │
│ │ • RMSNorm │ │ • Weight Conversion │ │
│ │ • RoPE │ │ • Tensor Bridge │ │
│ │ • Cache Ops │ │ │ │
│ └───────────────────────────┘ └───────────────────────────┘ │
└─────────────────────────────────────────────────────────────┘
│
▼
┌─────────────────────────────────────────────────────────────┐
│ Metal GPU Layer │
│ Apple Silicon Unified Memory Architecture │
└─────────────────────────────────────────────────────────────┘
Environment variables for customization:
| Variable | Default | Description |
|---|---|---|
VLLM_METAL_MEMORY_FRACTION |
auto |
auto allocates just enough memory plus a minimal KV cache, or 0.? for fraction of memory |
VLLM_METAL_USE_MLX |
1 |
Use MLX for compute (1=yes, 0=no) |
VLLM_MLX_DEVICE |
gpu |
MLX device (gpu or cpu) |
VLLM_METAL_BLOCK_SIZE |
16 |
KV cache block size |
VLLM_METAL_USE_PAGED_ATTENTION |
0 |
Enable experimental paged KV cache (requires pip install 'vllm-metal[paged]') |
VLLM_METAL_DEBUG |
0 |
Enable debug logging |
VLLM_USE_MODELSCOPE |
False |
Set True to change model registry to https://www.modelscope.cn/ |
VLLM_METAL_MODELSCOPE_CACHE |
None | Specify the absolute path of the local model |
VLLM_METAL_PREFIX_CACHE |
(unset) | Set to enable prefix caching for shared prompt reuse |
VLLM_METAL_PREFIX_CACHE_FRACTION |
0.05 |
Fraction of MLX working set for prefix cache (0, 1] |