Skip to content

Crusherbolt/AI-DOCTOR-

Repository files navigation

AI-DOCTOR

An experimental, multimodal conversational health assistant using Large Language Models (LLMs) and computer vision to enhance preliminary health support accessibility.


Project Overview

AI-DOCTOR aims to make basic healthcare interaction more accessible using the latest advances in AI. The assistant engages users in natural, conversational health queries and supplements responses with image-based analysis—such as reading and understanding prescriptions. This project is in a prototype phase, with active development continuing.


Key Features

  • Conversational AI: Handles user symptom descriptions and health-related questions with contextual, AI-powered responses.
  • Prescription & Image Interpretation: Uses computer vision (OpenCV) to analyze and extract text from uploaded prescription images.
  • Multi-Modal Interaction: Integrates both text and image modalities for versatile user engagement.
  • Easy-to-Use Web Interface: Designed for accessibility, allowing hands-on interaction without technical barriers.
  • Customizable AI Backend: Supports selection of LLM endpoints (e.g., GPT-4, GPT-4 Vision, or Azure OpenAI) to power conversational flows.
  • Extensible Flask Backend: Built for rapid prototyping, enabling further expansion such as medical dataset integration, external API calls, or improved health reasoning.

Demo Video (Draft)

AI-DOCTOR Demo (Draft)

This is an early preview of the project. Note: The app is not fully completed; ongoing development is planned.


Technology Stack

  • Frontend: HTML/CSS (with plans for React-based upgrade)
  • Backend: Python (Flask)
  • AI Backend: Large Language Models (e.g., GPT-4, GPT-4 Vision), cloud-hosted or local
  • Vision: OpenCV for image processing and OCR (Optical Character Recognition)
  • APIs: Integration-ready for future health APIs and real-time data sources

Getting Started

Prerequisites

  • Node.js & npm
  • Python (>=3.8 recommended)
  • Flask, OpenCV, and required Python packages (see requirements.txt)
  • Access to your preferred LLM (API key or local deployment)

Setup

  1. Frontend

    npm install
    npm run dev
    
  2. Vision Service

    cd sinus verification
    python app.py
    
  3. Backend/Client

    cd Client
    python app.py dev
    
  4. Model Configuration

    • In main.py, set the LLM as "gpt-4-vision", "gpt-4", or your cloud provider setup (e.g., Azure, Gemini, etc.)

Usage

  • Ask health questions (symptoms, conditions, advice)
  • Upload photos of handwritten/printed prescriptions for interpretation
  • Get AI-driven text responses based on user input and image analysis

Roadmap & To-Do

  • Enhance medical reasoning and expand dataset coverage
  • Improve OCR/vision accuracy and prescription handling
  • Upgrade frontend to React for richer user experience
  • Add user authentication and data privacy enhancements
  • Integrate external health and symptom-checker APIs

License

[Specify your preferred license here.]


Acknowledgments

  • OpenAI (and API providers) for language models
  • OpenCV community for computer vision tools
  • Any contributors and testers

DEMO LINKC= https://youtu.be/Vf7T_Yo6Cjc

  • For feedback, feature requests, or collaboration—please open an issue or pull request! *

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors