Speech Assistance for Blind People is a Python application that leverages the YOLO (You Only Look Once) object detection model and OpenCV to assist visually impaired individuals in navigating their surroundings. The application detects and recognizes objects in real-time and provides spoken descriptions to the user through a text-to-speech interface.
- Real-time Object Detection: The application uses YOLO to detect and recognize objects in the user's environment.
- Text-to-Speech (TTS) Integration: It provides audible descriptions of detected objects using TTS.
- User-Friendly Interface: Designed with simplicity in mind to ensure ease of use for visually impaired individuals.
- Customizable Voice Output: Users can configure the TTS voice and speech rate to suit their preferences.
- Object Recognition: In addition to detection, the application recognizes and identifies objects when possible.
- Detect Humans: Can spell the name of the person who is standing in front of you.
- Voice Assistant: Result is assisted to person through audio output.
- Run the 01_face_dataset.py - Add the unique id in the terminal(for ex: 1,2,3,...).
- Run the 02_face_training.py - The faces generated in the dataset folder will be trained.
- Run 03_face_recognition.py - add your name in the list (names = [none,"Shreyas"]) depending on the number of faces trained in the model.
- run the file, following output will be displayed on the screen.