Project 1: Out of Stock Detection

Retailers suffer billions of dollars in lost sales due to persistent out of stock products on their shelves. In this work, we devise a system based on explainability, computer vision and deep learning to leverage the power of AI into detecting out of stock labels by processing images of the aisle. The system is a combination of 5 deep learning models performing detection of products and labels, classification of labels into various types, and segmentation of shelves to get fixture information, along with other heuristic modules based on logic and inference. The system was deployed in 550 Walmart stores across the US where ~100 aisles from each store were processed to get the OOS data and delivered to the store in 15 minutes. This work is part of my Phd thesis and I will post a link to the thesis once I am done if anybody is interested in the details.

Project 2: Multistage OCR

Project 4: Robohelper project

During my masters at UIC, I worked with Prof.Zefran at the Computer Vision and Robotics Library. This was a pre-deep learning era where the huge language models hadn't taken off quite yet. We worked on developing a multi-modal communication interface with dialogue act classification, and haptic force feedback from the hand to assist the elderly in their daily activities. In the video, I integrated a Dialogue act classifier into the ROS framework and can be heard toying with it through a speech to text generator from CMU called pocketsphinx. I used the rviz tool to simulate arbitrary robot responses to each predicted dialogue act tag for the spoken sentence.

Want to get in touch?
Drop me a line!

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.