Building enterprise-scale data pipelines at Walmart — transforming raw data into reliable, queryable assets that power business decisions.
I am a Data Engineer at Deloitte, currently building enterprise-scale data pipelines for Walmart's data platform on Google Cloud. I engineer ingestion pipelines for 100+ tables across SAP systems into BigQuery, build Scala-based transformation JARs processing billions of rows, and orchestrate 20+ Airflow DAGs.
I hold an M.Tech in Computer Science (Data Science) from the Indian Statistical Institute, Kolkata — where I secured All India Rank 11 in the entrance exam. Before ISI, I spent 2 years at Tata Power building backend APIs, data pipelines, ML models, and automated reporting systems.
My professional journey in data engineering and software development.
Strong foundation in computer science and data science.
Technologies and tools I work with daily.
Highlights from my professional experience.
End-to-end ingestion pipeline for 100+ SAP tables flowing through GCS Raw Zone, Catalog Zone, and into BigQuery External Tables for downstream analytics.
Python tool that auto-detects column sensitivity, maps target schemas, renames SAP columns with business descriptions, and provisions GCS buckets — replacing hours of manual setup.
Optimized SQL logic across the consumption layer for large-scale datasets, reducing query execution time by 40% and cutting compute costs.
ML forecasting model for day-ahead electricity trading on Indian Energy Exchange, achieving 88% accuracy with real-time PI Server integration.
End-to-end automated Power BI dashboards for HR, customer operations, and engineering teams. REST APIs with FastAPI and Node.js serving 10+ dashboards.
Optimized XGBoost model improving payment default prediction from 77% to 85% accuracy using geographic feature engineering and hyperparameter tuning.