ANSHUL DHAWAN

[email protected] +91-9873919352 Gurugram, India
LinkedIn: https://linkedin.com/in/anshul-dhawan

PROFESSIONAL SUMMARY

I am a Data Engineer with strong expertise in developing scalable data pipelines and frameworks. With extensive experience in utilizing Databricks, Spark, Scala, and Python, I have successfully optimized data processing for large-scale financial data. My background includes working with advanced technologies such as Azure, AWS, Teradata, and Airflow. I am passion-driven, constantly seeking to enhance operational efficiency and automate processes in data engineering

WORK EXPERIENCE

Data Engineer 2
01/2025 - Present
Wio , Gurugram, India
Designed and developed streaming data pipelines using Azure Databricks Delta Live Tables (DLT) to process customer interactions and operational data in real time
Implemented Change Data Capture (CDC) and Slowly Changing Dimension (SCD) Type 1 logic with Auto-CDC APIs for efficient handling of incremental data updates
Built staging and gold layers in a Delta Lakehouse architecture with PySpark, enabling reliable and scalable data transformations
Automated data quality checks, schema evolution handling, and incremental ingestion to ensure high data reliability for downstream analytics and reporting
Collaborated with cross-functional teams to enable real-time dashboards and customer service analytics by integrating pipeline outputs with BI tools
Improved pipeline performance and reduced latency by 20–30% through partitioning, caching strategies, and watermarking in streaming jobs
Data Engineer 2
08/2023 - 09/2024
Expedia , Gurugram, India
Development of a generic framework to process and load travel rewards data across multiple layers using Spark, Scala and Python
Working with Teradata and S3 as storage layer
Orchestrating the jobs using Airflow DAGS – Spark Operator, Python Operator, Livy Operator, etc
Data ingestion and processing of data using Spark, Scala and Python
Data Engineer
03/2022 - 07/2023
Paytm , Noida, India
Development of data pipelines using Spark, Scala and Python
Processing huge volumes of financial data (in billions) every day across various products
Automation of Data Science Models and providing data to downstream models on time
Created an automated recon process to deliver data gaps to stakeholders
Wrote a common utility for data writing to an internal Datalake tool
Working with AWS S3 as storage layer
Orchestrating the jobs using Airflow DAGS – Python Operator, Livy Operator, etc
Data ingestion and processing of data using Spark, Scala and Python
Data Engineer 2
08/2023 - 01/1970
Expedia (via TEKsystems) , Gurugram, India
Development of a generic framework to process and load travel rewards data across multiple layers using Spark, Scala and Python
Working with Teradata and S3 as storage layer
Orchestrating the jobs using Airflow DAGS – Spark Operator, Python Operator, Livy Operator, etc
Data ingestion and processing of data using Spark, Scala and Python

EDUCATION

B.Tech
08/2015 - 05/2019
Bharati Vidyapeeth’s College of Engineering , New Delhi
12th Non-Medical (PCM)
01/2015 - 05/2015
Hansraj Model School, Punjabi Bagh , New Delhi

SKILLS

PROJECTS

Analysis and Prediction of Suicide Attempt
Analyzed the suicide dataset to identify attributes contributing to suicide attempts
Predicted future attempts with significant precision using machine learning models