In this project, I compensated for modern virtual assistants by adding website and app interaction features. I used Python due to its wide range of modules that help simplify many tasks. I used Google's speech recognition model in order to transcribe commands; afterward, I picked out keywords and used services such as PyAutoGUI, Selenium, and the Sys module to interact with website and app elements. On top of that, the assistant also contains a feature that uses eye recognition through cv2 in order to move the cursor and detect clicks through winks.
This project was very fun as I experimented with creating an assistant and interacting with websites through speech recognition.
Screenshots of prototype:
Here are some screenshots of my prototype and it clicking on a tab within the Youtube page.