This repository is a Windows-first desktop toolkit for a small YOLO object-detection workflow. It provides three companion scripts for capture and mandatory labeling, temporary YOLO dataset preparation and training launch, and live inference with telemetry overlays.
The tools are intended for local desktop use, controlled imaging setups, and experiment-scale model work. Camera support, GPU detection, and runtime stability depend on the installed Windows drivers and Python packages, and the repo is not positioned as a production inspection system.
Use Python 3.12 and keep dependencies inside a dedicated .venv312 environment so the GUI tools and native packages stay aligned.
py -3.12 -m venv .venv312
.\.venv312\Scripts\activate
pip install --upgrade pip
pip install --only-binary=:all: "numpy<2.3.0" "opencv-python<4.13" pyqt5 ultralytics torch psutil pynvmlThe repo stores shared state on disk in captures/, classes.txt, class_colors.json, and config.json. Keep those files alongside the three top-level scripts when running locally.
Run the primary capture and labeling app:
.\.venv312\Scripts\python.exe .\main.py --camera 0Launch the training companion to build a temporary .yolo_training_cache/ dataset and start the Ultralytics CLI:
.\.venv312\Scripts\python.exe .\train_model.pyLaunch the inference viewer for live model checks and screenshot capture:
.\.venv312\Scripts\python.exe .\run_inference.pySee CONTRIBUTING.md for local workflow, validation expectations, and review readiness.