About me
I'm a Data Engineer with 5 years of experience architecting enterprise-scale data platforms that drive strategic business insights. I specialize in building robust ETL/ELT pipelines, designing Azure-based lakehouse architectures, and enabling advanced analytics through highly performant and governed data systems.
My toolkit includes Python, PySpark, SQL, Azure Data Factory, Databricks, and Apache Airflow, with strong expertise in Delta Lake, Unity Catalog, and MLOps integration. I thrive on optimizing data workflows for scale, reliability, and cost efficiency—ensuring 99.9% data accuracy and system availability across cloud environments. From automating CI/CD pipelines to enabling clean, production-ready data for machine learning models, I bridge the gap between data engineering, governance, and business value with a product-driven mindset.
What i'm doing
-
Data Engineering & Pipelines
Designing and orchestrating scalable ETL/ELT pipelines using Databricks, ADF, and PySpark—optimized for performance, reliability, and cost efficiency across Azure.
-
Cloud Data Architecture
Implementing Azure-based lakehouse solutions with Delta Lake, Unity Catalog, and medallion architecture for unified analytics and governed data access.
-
Data Enablement for AI & ML
Preparing clean, production-grade datasets and feature stores that accelerate machine learning workflows, experimentation, and model deployment in collaboration with data science teams.
-
Analytics & BI Enablement
Supporting advanced analytics and decision-making through dimensional modeling, curated datasets, and interactive dashboards using Power BI.
Tools & Technologies
Azure
Databricks
ADF
Fabric
Python
SQL / T-SQL
Apache Airflow
Azure DevOps
Unity Catalog
SQL Server
Snowflake
Power BI
