LLM Foundry is the workshop around an LLM.
It is not the model itself. It is the memory, compression, tokenizer, tool use, traces, harnesses, training plumbing, and API orchestration that make a model useful.
- a backend abstraction for many model sources
- tokenizer support
qwenfor the built-in Hugging Face Qwen pathopenaifor OpenAI-compatible BYOK providersanthropicfor Anthropic BYOK providershffor any Hugging Face local modelmultioropenai-multifor failover bundles
Example:
python -m llm_foundry demo --provider qwen --model Qwen/Qwen2.5-0.5B-Instruct --prompt "Hello"Use proof for a terminal-first demo that shows the question, tool call trace, final answer, and JSON artefact. It is designed for screenshots and file explorer use on Windows, macOS, and Linux.
Example:
python -m llm_foundry proof --provider qwen --model Qwen/Qwen2.5-0.5B-InstructA live GitHub Pages demo can host the docs site from docs/. Use it to try the default Qwen model path or BYOK providers.
- Default model:
qwen - BYOK providers:
openai,anthropic - Multi-endpoint bundles:
multi/openai-multi
For the live page, use the commands on docs/index.html.