MY SKILLS

Data Engineerring Skills

Python 85%
Snowlflake 80%
Datamodeling 85%
GCP 80%
Apache hive 72%
AWS 75%
S3,Glue 70%
SQL 68%
Azure Services 73%

Soft Skills

Cross function collaboration 80%
stakeholder Communication 80%
Critical Thinking 75%
ProblemSolving 90%
Attention to Detail90

Databases

  • MySql
  • Postgre Sql
  • MongoDB
  • Oracle

Project Tools

  • Agile
  • Scrum
  • Git
  • SVN
  • Jira

Bigdata tools

  • Apache Hive
  • Apache HBase
  • Apache Spark
  • Kafka
  • Parquet

Cloud Platforms

  • AWS
  • GCP
  • Azure

Work Experience

JP Morgan Chase | Data Engineer

  • Apache Airflow
  • Git
  • GCP
  • Pipelines
  • DataLake

At JPMorgan Chase, I worked on modernizing legacy financial data pipelines into a cloud-based data lake architecture. Using PySpark, Apache Airflow, and AWS (S3, Redshift, Lambda), I built efficient, scalable ETL pipelines that processed transactional data for compliance and risk reporting. My work ensured timely delivery of regulatory reports (CCAR, Basel III) and improved data quality across finance and risk domains. I also partnered with data governance teams to implement data lineage tracking, enabling auditability and transparency in financial operations.

Key Contributions

Elevance Health | Senior Data Engineer

  • PySpark
  • AWS
  • Snowflake
  • Data Modeling
  • HIPPA

At Elevance Health, I led the design and development of large-scale data engineering solutions focused on healthcare claims and electronic health records (EHR). Using PySpark and AWS services (Glue, S3, Kinesis, Lambda), I built end-to-end ETL pipelines capable of handling millions of records daily, ensuring HIPAA compliance and data integrity. I also contributed to the migration of on-prem systems to Snowflake, optimizing cost and performance for real-time analytics. The project enabled timely insights for care management teams, significantly reducing data latency and supporting better member outcomes.

Key Contributions

Swiggy | Associate Data Engineer

  • Python
  • DataLakes
  • S3
  • Glue
  • Lambda

At Swiggy, I designed and maintained real-time data pipelines using Apache Kafka, Spark Streaming, and Hive to handle high-velocity order, delivery, and user behavior data. My work enabled live tracking, dynamic pricing, and restaurant ETA calculations across multiple cities. I also contributed to building a data quality monitoring framework, ensuring consistency in downstream analytics used by marketing and operations teams. This project enhanced Swiggy’s ability to deliver personalized experiences and optimize delivery logistics.

Key Contributions

Home Goods | Data Engineer

  • PySpark
  • AWS
  • snowflake
  • Data Modeling
  • Apache Airflow

At HomeGoods, I built robust ETL pipelines to support retail sales analytics, integrating data from POS systems, supply chain sources, and vendor feeds. Using SQL, Informatica, and Teradata, I automated the data ingestion and transformation processes, significantly reducing manual effort and increasing reporting accuracy. I also supported inventory forecasting models by delivering clean, aggregated datasets to data scientists, leading to more accurate replenishment strategies and reduced stockouts across stores.

Key Contributions