This is where I currently work. I originally interned at LLNL over the summer of 2017, and joined part-time during university to support the software side of the project longer-term. I am continuing to develop vision algorithms for scientific measurements, but I am also working on streamlining our team’s workflow by using the latest workflow automation tools like Dagster.
I designd flight hardware and software for CubeSats, developed hardware and software interfaces with testability in mind, and led the lab through a sprint where the lab was producing multiple flagship missions at once.
I developed image processing pipelines for autonomous vehicles (AVs) and architected an extensible simulator for AVs using Microsoft AirSim and Unreal Engine.
I interfaced the JPL operations team with the operations team at Cal Poly, in addition to my day-to-day duties which include drafting and verifying uplink products, analyzing telemetry, and developing data management tools for streamlining and automating spacecraft operations.
I interned at LLNL working on the Film Scanning and Re-Analysis Project. My research project was to design a semi-automated image processing pipeline to measure the size of the shockwave from a nuclear blast. The pipeline I developed provides was developed in Python using the OpenCV library and can detect and accurately measure the shockwave and fireball with minimal user input.