Virtual Blind Assistance is being developed to guide blind persons via voice instructions in indoor and outdoor environment. The assistance will detect objects, recognize them and guide the blind person to plan his path. The objects to be detected and recognized include house furniture, humans
Virtual Assistant for blind people
Virtual Blind Assistance is being developed to guide blind persons via voice instructions in indoor and outdoor environment. The assistance will detect objects, recognize them and guide the blind person to plan his path. The objects to be detected and recognized include house furniture, humans and walls , vehicles , and other things
The blind person wear glasses in which camera is attached also that person have hand free and a mobile phone so that when a person see anything from camera send the input to your mobile phone then mobile phone process it and give the audio to the hand free by this method blind person know the things that is in front of him
To assist a blind person in his daily life without interaction of human being.
To include feature of object detection and object recognition.
After recognizing object, the system will play an audio to guide the blind person.
it includes all aspect of life , like moving ,news paper reading and other things reading , meeting people and their behavior , shopping , face detection , food checking , and all aspect of life
During the Implementation phase, the project plan is put into motion and the work of the project is performed. It is important to maintain control and communicate as needed during implementation. Progress is continuously monitored and appropriate adjustments are made and recorded as variances from the original plan.
Hence as part of execution we broke our Virtual Assistance to modules for a broad view of the project and to strategize easily for further implementation and tracking through out.
Modules
Developing your own Voice Assistant App
For now we will be using custom voice assistance provided to us by the gTTS, STT or TTS free libraries provided open source, for later on broad production and our own identity we might be developing our personal assistant voice api.
Voice/speech to text (STT)
This is the process of converting speech signal into digital data (e.g., text data). The voice may come as a file or a stream. You can use CMU Sphinx for its processing.
Text to speech (TTS)
This is the process of converting digital data (e.g., text data) into speech signal. The voice may come as a file or a stream. We might use gTTS for its processing.
This process translates text / images in a human speech. It is very useful when, for instance, a user wants to hear the correct pronunciation of a foreign word.
Intelligent tagging and decision making
Intelligent tagging and decision making serve for interpreting the user's request. For example, in our current perspective the user may ask: 'I want to reach certain destination'. The technology will tag few appropriate path and suggest you a few according to your interests. We are overseeing multiple APIs for our certain needs.
Image recognition
Image recognition is the most important part of our virtual assistant project and the mighty module we are focusing on, we’ve considered and have been working on OpenCV because of the bigger community size that could help us in any point of error. We will be considering deep learning model be practiced and matching accuracy of our need in OpenCV through image recognition/ object detection/ classification, clustering/ or mapping.
Virtual assistants can make it feel like secretaries have come back in style. Some examples includes for major perspective for Blinds are:
In the final deliverable we will integrated system in which Bluetooth glasses , Bluetooth hand free and mobile are attach to each other through hand free , and mobile is connected to a cloud server
, so we will show following functionality during final testing of the project
A blind person wear a glasses along with the hand free and a mobile phone in his pocket, so when person start walking the things that can see by camera will detect using Open CV and then that result will send to mobile and then mobile process it through Cloud server using deep learning algorithms and then cloud server tell the result of the things to the mobile and mobile send that result to the hand free using text to speech and blind understand the
It tell about the following things:
here is the diagram for our final Deliverable :

| Item Name | Type | No. of Units | Per Unit Cost (in Rs) | Total (in Rs) |
|---|---|---|---|---|
| Jetson nano | Equipment | 1 | 21000 | 21000 |
| SD card 32 GB | Equipment | 1 | 800 | 800 |
| Charger | Equipment | 1 | 250 | 250 |
| Camera (8 MP) for jetson nano | Equipment | 1 | 8000 | 8000 |
| Audio speaker for jetson | Equipment | 1 | 800 | 800 |
| Bluetooth camera glasses | Equipment | 1 | 21000 | 21000 |
| Cloud servers | Equipment | 1 | 5000 | 5000 |
| Bluetooth Hand free | Equipment | 1 | 600 | 600 |
| Document printing | Miscellaneous | 1 | 2000 | 2000 |
| Total in (Rs) | 59450 |
The ceramic tiles industrial sector is a comparatively young industry that has taken signi...
The aim of our work is to present an inexpensive, light-weight and easily controlled robot...
Edge Detection is a key low-level task in computer vision which can greatly help in severa...
Recently Data science is advancing at a fast pace and its application makes things easier...
Harvester is an autonomous robot having an ability to recognize and pluck the fruits. In i...