Project Overview
Waldinger's data platform ran Snowflake across 3,000+ objects with no automated validation, and Azure ML compute clusters sat idle at peak provisioning between jobs. Two fixable problems. This internship was about fixing them.
The core deliverable was a production LLM inference pipeline on Azure ML using Mistral-7B-Instruct, supporting both real-time and batch modes from a single registered model artifact. Infrastructure was written entirely in Terraform so deployments are reproducible and configuration drift is structurally impossible.
The schema drift detection system runs on every pipeline execution, snapshots Snowflake object definitions, and surfaces deviations before they cascade into downstream failures. Between that and compute right-sizing, idle spend dropped 70%.