- Docker and Docker Compose
- NVIDIA GPU with CUDA support (recommended for faster inference)
- At least 40GB disk space for LLM model storage
git clone https://github.com/dbcls/bsllmner-mk2.git
cd bsllmner-mk2
# Build and start containers
docker compose up -d --buildThe compose.yml reserves all available NVIDIA GPUs by default. To restrict which GPUs are used, edit the CUDA_VISIBLE_DEVICES environment variable:
# compose.yml (ollama service)
environment:
- CUDA_VISIBLE_DEVICES=0,1 # Use GPU 0 and 1For Ollama performance tuning options, see Configuration - Ollama Performance Tuning.
- Python 3.10+
- uv
- Ollama server running locally or remotely
uv sync
# Install with test/development dependencies
uv sync --all-extrasIf the Ollama server is running on a different host, set the OLLAMA_HOST environment variable:
export OLLAMA_HOST=http://<ollama-host>:11434For Docker Compose:
# Check containers are running
docker compose psFor uv (local):
# Check the CLI is available
uv run bsllmner2_extract --help