An AI-powered, sensor-integrated wearable device for obstacle detection, text recognition, navigation, and real-time assistanceβbuilt to empower visually impaired individuals with safe, independent mobility.
GuardianEye is a multi-sensory smart eyewear system designed to assist visually impaired individuals in navigating their surroundings confidently and independently. By combining advanced AI, computer vision, and sensor fusion, the device offers:
- Real-time obstacle detection
- Currency note recognition
- Printed text reading via OCR
- Multi-language voice feedback
- Location tracking for caregivers
- Hands-free voice command control
- Emergency alert system with haptic and audio feedback
| Feature | Description |
|---|---|
| π£οΈ Real-Time Obstacle Detection | Detects and classifies static/dynamic obstacles using LiDAR + YOLO |
| π¬ Multi-Language Voice Output | NLP-based output in multiple languages for localized interaction |
| π· Optical Character Recognition (OCR) | Reads text from newspapers, signboards, etc., and converts to speech |
| ποΈ Voice Command System | Microphone-powered hands-free interaction via natural language |
| πΈ Currency Note Detection | Identifies denomination using computer vision to prevent fraud |
| π§ GPS Location Tracking | Enables family members to track userβs location remotely |
| π€ Person & Object Recognition | Recognizes known individuals and frequently seen objects |
| π³ Haptic Feedback | Vibrates for obstacle alerts, turns, or important environmental cues |
| π¨ Emergency Beep Alert | Loud beeping sound triggers when immediate danger is detected |
- π§ AI Models: YOLOv8 (Object Detection), Tesseract (OCR), NLP (Speech-to-Text / Text-to-Speech)
- π Communication: Bluetooth Low Energy (BLE)
- π Audio Feedback: Bone conduction speakers
- π· Sensors: LiDAR, Pi camera, GPS
- π¦ Hardware Platform: Microcontroller (Raspberry Pi), Rechargeable battery, Microphone

- Navigate urban streets, stairs, and corridors
- Recognize people at work or home
- Read public signs or packaging labels
- Detect currency during monetary exchange
- Alert surroundings in case of emergencies
- Stay connected with caregivers for safety
Coming Soon: Demo Video
![]()
- β Real-time object detection
- β Text-to-speech conversion of printed material
- β GPS tracking dashboard for caregivers
- β Voice-command feature demo
- Research and literature review
- Hardware component selection
- AI module prototyping (YOLO + OCR)
- NLP-based multi-language system
- Integration & Testing
- Real-world trials with visually impaired users
- Final production and optimization
- Anirudh Garg β Computer vision(Team Lead)
- Aaradhya Sharma β Computer vision
- Abhiroop Singh β IoT
- Rajveer Singh β Speech Processing, Documentation
- Bhavneet Kaur β NLP, Speech Processing
This project is licensed under the MIT License.
Have feedback, ideas, or want to collaborate? Reach out at:
"Letβs build a world where everyone can see possibilitiesβeven without sight." π
