Data platforms that
scale with the business.

I design and build the data infrastructure, tooling, and engineering practices that let organizations move faster, spend less, and make better decisions.

10+

Years Delivering

10+

Organizations Served

3

Cloud Platforms

80%

Cost Reduction (Best Case)

Services

What I deliver

End-to-end data engineering, from greenfield architecture to production operations. I build the systems and lead the teams that make data infrastructure a competitive advantage.

Platform Architecture

Streaming and batch architectures, data lake design, warehouse modeling, infrastructure-as-code. Full lifecycle from whiteboard to production on AWS, Azure, or GCP.

Custom Tooling

When off-the-shelf tools don't fit, I build what's needed. Orchestration engines, SQL frameworks, API automation layers, CI/CD pipelines. Production software in Python, Rust, Go.

Analytics & ML

BI platforms, predictive models, customer-facing analytics products. From inventory forecasting to customer segmentation to self-service reporting that stakeholders actually use.

Governance & Compliance

Data governance frameworks, retention policies, access controls, ISO certification enablement. The operational foundation that lets engineering move fast without creating risk.

Team Leadership

Building and leading data engineering teams. Defining standards, mentoring engineers, establishing practices that scale. Currently managing data engineering at a high-growth DTC brand.

AI Readiness

Positioning organizations for AI adoption through modern data infrastructure, clean modeling practices, and the engineering foundations that make ML actually work in production.

Case Studies

Selected work

Production systems designed and shipped across adtech, financial services, food services, consumer health, and automotive.

Data Modeling & Lineage Framework

Platform Tooling - Python, AWS Athena, QuickSight

Challenge

No cloud-compatible data modeling tool existed for the stack. Development lacked standardization for code tracking, environment management, documentation, and orchestration.

Solution

Built a Python-native framework with automated column-level lineage, dependency graphs, self-documenting BI datasets, catalog sync, and full CI/CD integration.

Result

Validated the modeling pattern at scale. Successfully transitioned the organization to sqlMesh when a mature open-source option emerged - the framework made migration seamless.

Unified Orchestration Engine

Infrastructure - Rust, ECS

Challenge

Scheduling logic fragmented across Airflow, Step Functions, and Dagster. Each had its own monitoring, error handling, and operational semantics.

Solution

Engineered a high-performance job scheduler in Rust deployed on ECS. Unified all scheduling into a single system with consistent alerting and error handling.

Result

Single pane of glass for all pipeline operations. Reduced on-call burden and improved reliability across the board.

Inventory Prediction Engine

Analytics & ML - Python, SciKit-Learn, Step Functions

Challenge

Advertisers needed precise inventory availability across many dimensions. Predictions had to be accurate enough to commit campaigns against and communicate with confidence.

Solution

Built an ML pipeline with 500-segment clustering and per-cluster regression. Model statistics revealed a mean-based approach produced identical results - replaced the system with SQL queries.

Result

Same accuracy, fewer anomalies, dramatically simpler operations. A deliberate simplification driven by data, not compromise.

Pipeline Automation at Scale

Automation - Python, REST APIs, 4,000+ pipelines

Challenge

4,000+ advertising pipelines across throttled vendor APIs (20K requests/day cap). Just loading configuration consumed 12,000 calls. Manual management was unsustainable.

Solution

API wrapper with incremental state caching reduced startup calls from 12,000 to under 100. Dynamic queue management using historical job durations prevented resource overloading.

Result

Fully self-healing system. Automated detection, retries, and escalation freed the team entirely from routine pipeline management.

Enterprise Workload Platform

Full Stack - Azure, Data Factory, SQL Server

Challenge

80+ associates spent up to 90 minutes daily downloading the same report from a mainframe. Bi-weekly consolidation took two analysts nearly a week.

Solution

Automated mainframe extraction via event-driven pipelines. Built a custom platform with real-time database sync, standardized workflows, and persistent tracking.

Result

Eliminated 90 min/day for 80+ users. Bi-weekly reporting went from a week to real-time. Became the foundation for unified operational finance.

Legacy CI/CD Automation

DevOps - Azure Logic Apps, GitHub Actions, Informatica API

Challenge

Legacy orchestration tool supported git but had no way to continuously deploy merged changes. Cross-environment object IDs embedded in XML didn't translate between environments.

Solution

Built a deployment pipeline with custom ID resolution and XML transformation logic that the vendor's own API couldn't handle.

Result

Git merges automatically propagated to test and production. Eliminated manual deployments and direct environment changes entirely.

Open Source

Platforms & Tools

Software I build and maintain publicly. Click any card to view the source on GitHub.

Background

Credentials

Certifications

AWS Developer Associate

Snowflake SnowPro Core

Databricks Data Engineer Associate

dbt Analytics Engineering

Apache Airflow DAG Authoring

Fivetran Technical Certification

Education

MS in IT, Data Science

Walsh College

Big Data & Data Science Post Graduate

Duke University

BAS, Accounting

Ferris State University

Core Stack

Python SQL Rust Go AWS Azure GCP Snowflake Databricks Airflow dbt Terraform

Let's build something.

Whether you need a data platform architect, a hands-on engineering leader, or someone to come in and solve a hard problem - I'd like to hear about it.