This repository is a submission for MLDA Deep Learning Week Hackathon 2022 held on 30 September 2022 - 3 October 2022.
This project is created by:
- Dhairya Rungta (@dhairyarungta)
- Joshua Adrian Cahyono (@JvThunder)
- Karan Andrade (@kj-andrade)
- Michael Andrew Chan (@michac789)
FIX6DSENSE.AI
is an AI-Powered real-time virtual assistant that enables the visually impaired to see the world by hearing.
- Detecting 80 different categorical object types with their confidence level
- Priority-based detection
- Distance prediction & scaling algorithm
- Mobile phone activated camera for various items
- Speech-activated mode selection
- Multi-threading usage for asynchronous voice warnings and feedbacks
- SSD Mobilenet COCO API Model
- Aware mode: continuously speak out loud 3 items with the topmost priority
- Warn mode: only shout out warnings if any item is too close based on priority
- Search mode: focus on searching a particular item until it is found
Please click the link below to view our other resources regarding to this project submission.
-
Clone this repository and make sure you have all the files downloaded
-
Set your environment and install the required python packages
It is recommended for you to use Anaconda, and specifically
python version 3.9.0
(newer version of python has some compatibility issues related to thecollections
package, other older python version is untested and might have some compatibility issues as well). Then, install all the required python packages by typing:pip install -r requirements.txt
-
Execute
main.py
ormain.ipynb
fileIf you want to use an external camera, you can connect via IP Webcam Pro and edit the
url
variable when running the model. By default, it is connected to your PC / Laptop default camera. You can also toggle theUSE_SPEECH
global variable toTrue
orFalse
. If it isTrue
, it will use speech recognition to choose and change the mode, or else it will manually prompt you for the mode input.