To create a seamless, "local-feeling" experience where open-source contributors have instant access to enterprise-grade compute (e.g., H100 clusters). It should feel as simple as running Ollama, but instead of running on your laptop, it transparently offloads to powerful infrastructure subsidized by providers.
- Merit-Based Access: Compute is not bought; it is earned. "Unlimited" enterprise-grade resources are available to those whose open-source contributions qualify them.
- "Ollama-like" Simplicity: No complex cloud consoles or provisioning tickets. You just run a command, and the compute is there.
- Invisible Infrastructure: The developer doesn't "request" a server; they just run their workload. The system handles the federation of resources from providers (NVIDIA, etc.) in the background.
- Merit Engine: Verifies contributor status and authorizes access based on public impact.
- Transparent Broker: Routes workloads to the best available provider without user intervention.
- Federated Resource Pool: Providers (NVIDIA, Academic Clusters, Clouds) plug in their idle or subsidized capacity.
- Standardized Runtime: Ensures workloads run consistently regardless of the backing hardware.
- Unified CLI: A single tool (e.g.,
cc run) that mimics local execution. - Seamless Tunneling: Ports and logs are forwarded instantly, making remote H100s feel like
localhost.
- Implement the logic to validate open-source contributions (e.g., GitHub/GitLab activity).
- Build the CLI that abstracts the remote execution.
cc run llama3should feel identical toollama run llama3, but running on a DGX Cloud instance.
- Connect the first major provider (e.g., NVIDIA) to back the experience.