Skip to content

End-to-end MLOps platform demonstrating a complete CI/CD lifecycle for LLMs. Tech stack: FastAPI, Docker, Kubernetes (EKS), ArgoCD, and GitHub Actions. Focuses on automation, testing, and scalable model serving.

Notifications You must be signed in to change notification settings

Sanket-Raj/opsgpt-gitops

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

14 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

OpsGPT: GitOps-Driven MLOps Inference Engine

OpsGPT is a production-grade inference engine designed to serve Hugging Face LLMs (DistilGPT-2/SmolLM) with high availability and automated delivery.

I built this project to demonstrate a Split-Repository GitOps Strategy, separating application source code from configuration manifests to ensure security, auditability, and clean release cycles.

OpsGPT: GitOps MLOps Pipeline

[ Badges here: CI Status | Python 3.10 | Docker ]

๐Ÿš€ Overview

OpsGPT is a production-ready LLM inference engine deployed via Kubernetes.

๐Ÿ› ๏ธ Local Development

  1. Install dependencies: pip install -r requirements.txt
  2. Run server: uvicorn app.main:app --reload

๐Ÿ”ฎ Future Roadmap

  • Add Redis Caching
  • Implement Model Quantization

๐Ÿ— Architecture

graph LR
    A[Developer] -->|Push Code| B(opsgpt-app Repo)
    B -->|GitHub Action| C{CI Pipeline}
    C -->|Build & Test| D[Docker Registry]
    C -->|Update Tag| E(opsgpt-gitops Repo)
    E -->|Sync| F[ArgoCD]
    F -->|Deploy| G[Kubernetes Cluster]
    G -->|Scale| H[HPA]
Loading

About

End-to-end MLOps platform demonstrating a complete CI/CD lifecycle for LLMs. Tech stack: FastAPI, Docker, Kubernetes (EKS), ArgoCD, and GitHub Actions. Focuses on automation, testing, and scalable model serving.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published