@root: Tinkerer. Developer. Innovator
☑ Leading interactive labs for ~60 students, totaling 120 minutes,
teaching fundamentals of coding, C, and data structures.
☑ Orchestrate office hours for personalized student support, focusing on largely
theoretical discussions about C, alongside debugging sessions.
☑ Working with professors to develop new material and problem sets to effectively
challenge students in a collaborative and engaging format that enhances learning.
Teach over 60 students the fundamentals of coding, C, and data structures, alongside working with professors to develop new problem sets to challenge students in a format that works alongside Generative AI.
☑ Developed Computer Vision implementation based off ResNet to segment
internals of 36 different iPhone models
☑ Implemented SORT algorithm for segmenting parts on a conveyor belt moving at 5 m/s at
real-time (> 60 FPS) with 96% IOU accuracy
☑ Used YOLO V8 Image Classification to train a top-down model to detect screws from 3
different angles on 720p cameras on 36 models of iPhones, 12 models of iPads, and 4 models of
Apple Watches. Led the creation of a demonstration video that was presented to Apple and leads
of 100+ member lab.
Developed CV implementation to segment internals of 36 different iPhone models into 10 classes, and created a screw detection software for iPhones, iPads, and Apple Watches.
☑ Led a team of 10 in building out autonomous vehicle driving and passing, including
position communication, implementing a Model Predictive Controller, and racing at 30 mph.
☑ Innovating a LiDAR + Stereo Camera setup for perception of surrounding vehicles for
autonomous zero aid passing.
☑ Assisted in development of a new website to improve traffic, potential member interest,
and promotion for sponsors.
☑ Control security of 3 remote workstations, a variety of 3D printers that can be accessed
by the 100+ members of the club.
Led a team of 10 in building out autonomous vehicle driving and passing, including onboard LiDAR based object detection.
☑ Developed performance modeling software to assist in chip development
with 97% accurate benchmarks on a variety of ML workloads (BERT, ResNet50, etc.).
☑ Improved prior software hardware resource utilization by over 46% with a novel weighted
round-robin load management solution.
☑ Created a memory modeling software that modeled memory packets throughout the chip, and
the variety of paths to identify bottlenecks in current designs.
Developed performance modeling software to assist in chip development and created a memory modeling software for 5+ ML Workloads
☑ Published and presented this paper to over 30 audience members, passing
a rigorous approval period with 2 separate peer-revision trials.
☑ Built universal modeling software for variety of different silicons to test against 10+
ML Workloads, compatible with CPUs, GPUs, and even TPUs.
☑ Modeled memory and performance accuracy of Google TPU and other hardware within 5% of
lab-tested measurements.
Published peer-reviewed paper at AAAI discussing a universal ML modeling software for a variety of different silicon
I Want That is a mobile application created to help users find the best artists to create the custom comissions that they envision. It does this through providing a reference image for the comissioned piece, which is described by Google's Gemini AI, which afterwards we use to semantically search our existing database of artists in order to find the top 3 artist.
Audit-AI is a chrome extension built around websites like Pinterest and Reddit, that determines if an image is AI-Generated or not. This is done by training a dataset over a YoloV8 image classification model, that had an overall accuracy of 98% on the provided dataset, and an estimated accuracy of 75% based upon real-world testing. The backend was written in primarily Python and deployed on a local web server, while the front end consisted of a ReactJS based chrome extension.
Braille Score is an application that translates the sheet music for visually impaired musicians. It does this using a custom Computer Vision model called OEMER, which creates a braille format for the music. Afterwards, musicians have a variety of ways of understanding the music, from braille sheets to audio for whichever format they best learn through.
GlobaLex is an application that does real time translation of a call into a different language. To create this application, we built a full web-calling application using Azure cloud services and ReactJS. The authentication is based upon Mongo DB. After receiving an audio stream on the opposite end, using OpenAI's Whisper model, we translate it in 30 second audio chunks.
Eco Bin was a project that focused on improving the environment through auto-sorting waste into recycling, compost, and trash. It comprised of 2 different sensors, a camera for computer vision, alongside of a depth camera to detect the shape of the object. Using this information along with a custom-built deep learning model, the robotic trash can was able to accurately classify 80% of the tested objects, with a 95% accuracy on the training dataset. This software was coupled with a hardware solution that would automatically ingest and sort the respective waste.