We welcome contributions to vllm-mlx!
# Clone the repository
git clone https://github.com/waybarrios/vllm-mlx.git
cd vllm-mlx
# Install with dev dependencies
pip install -e ".[dev]"# Run all tests
pytest tests/
# Run specific test file
pytest tests/test_paged_cache.py -v
# Run with coverage
pytest --cov=vllm_mlx tests/# Format code
black vllm_mlx/
isort vllm_mlx/
# Type checking
mypy vllm_mlx/# LLM benchmark
vllm-mlx-bench --model mlx-community/Qwen3-0.6B-8bit
# Image benchmark
vllm-mlx-bench --model mlx-community/Qwen3-VL-8B-Instruct-4bit
# Video benchmark
vllm-mlx-bench --model mlx-community/Qwen3-VL-8B-Instruct-4bit --video- Bug fixes - Fix issues and improve stability
- Performance optimizations - Improve inference speed
- New features - Add functionality
- Documentation - Improve docs and examples
- Benchmarks - Test on different Apple Silicon chips
- Model support - Test and add new models
- Fork the repository
- Create a feature branch
- Make your changes
- Run tests to ensure they pass
- Submit a pull request
See Architecture for details on the codebase structure.
If you have access to different Apple Silicon chips (M1, M2, M3, M4), benchmark results are valuable:
vllm-mlx-bench --model mlx-community/Qwen3-0.6B-8bit --output results_m4.jsonOpen an issue at GitHub Issues.