
- Projects 2
- Followers 1
Soumo Gorai
Kolkata, West Bengal
According to the World Health Organization, around 40 million people in the world are blind, while another 250 million have some form of visual impairment. The main objective of this project is to solve the daily life problems faced by them using AI and IoT technologies. Here, the users will give voice commands to get the assistance from the device. The device which will assist the users in various ways as listed:- 1. Environment Activity. 2. Lane Detection. 3. Basic Information ( Date, time, location, calendar). 4. Drop a message. 5. Emotion Detection. ...learn more
Project status: Under Development
Internet of Things, Artificial Intelligence, Graphics and Media
Intel Technologies
Intel Opt ML/DL Framework
The main objective of this project is to solve the daily life problems faced by them using AI and IoT technologies.
Here, the users will give voice commands to get the assistance from the device. The device which will assist the users in various ways as listed:-
I have created a prototype version to detect the various environmental activities, the working video is attached under in the video links.
Dataset for Activity Detection : https://forms.illinois.edu/sec/1713398
In this prototype version instead of using the camera, the user is manually selecting images for the activity prediction.
The final device will be totally different, It will be an IoT device where it will take the input from the camera and microphone and compute them using complex deep learning models to assist the users, the architecture of the deep learning models as well as the device is attached in the image section.
The entire Approach:
The user asks for the assistance --> Microphone --> IoT device -->Camera--> IoT device -->Natural Language Processing and Image Processing --> Deep Learning Models --> Prediction --> text to voice --> speaker.
The Prototype version of activity detection is trained in Google Colab.
While training the model(activity detection) I have gone through the following steps:
After the model is ready, it is made ready for inference via conversation with the help of gTTs and speech_recognition.
Python
Keras(TensorFlow)
gTTs
speech_recognition