Senior Data Engineer with 15+ years in infrastructure, systems, databases and cloud. The last 3 years focused on data engineering and data platform work.
I build ingestion pipelines from PostgreSQL, SQL Server and APIs into curated Iceberg tables — PySpark for large volumes, Python (DuckDB, Pandas, Arrow) for high-performance batch jobs.
Kubernetes · ArgoCD · Helm · Kustomize · Spark Operator · Hive Metastore · MinIO
Iceberg · Trino · Dremio · Spark/PySpark · Airflow
AWS (S3, MWAA, Glue, Athena) · OCI · Python · SQL
- On-prem lakehouse: Designed and built a Kubernetes-based lakehouse with GitOps (ArgoCD/Helm/Kustomize). Added Dremio so analysts can query data on their own, without routing requests through the data team.
- Optimization: Replaced 8h+ financial reports running on PostgreSQL with PySpark on AWS Glue (MWAA). Results land in Parquet tables queried by Athena (8h → 40min).
- Legacy migration: Moved ETL workflows from Pentaho and SSIS to the Kubernetes lakehouse platform.
