A cloud-native ETL pipeline for ingesting, transforming, and analyzing financial market data using AWS, GCP, Docker, and Apache Airflow. This project is currently under active development and serves as a hands-on demonstration of modern data engineering and DevOps practices.
A data‑science capstone from my final semester at Purdue University. We built several classification models (Logistic Regression, SVM, KNN, Random Forest) to predict how many games Purdue would win in the 2024 NCAA Tournament, and evaluated them against both business and data‑science success criteria.