Experienced Data Engineer with 10 years of expertise in designing, building, and optimizing large-scale data pipelines, databases, and architectures. Proficient in modern data engineering tools and platforms, including Azure, Databricks, Snowflake, and Hadoop. Adept at ensuring data quality, scalability, and performance while collaborating with cross-functional teams to deliver business insights.
Client: ABI InBev | Domain: CPG | Role: Data Engineer
Technologies Used: Azure, Azure Data Factory (ADF), Azure Databricks (ADB), Snowflake
Project: Migrated data from Snowflake to Azure Data Lake (Brewdat) and converted dataflows from ADF to Databricks using PySpark.
Architecture: Implemented Unity Catalog and Medallion Architecture with data transformation at the Silver layer.
Key Contributions:
Client: Mars | Domain: CPG | Role: Data Engineer
Technologies Used: Azure, Azure Data Factory (ADF), Azure Databricks (ADB), PostgreSQL
Project: Developed and enhanced Databricks notebooks to create Delta tables for the Price Pack Architecture (PPA), used by the Power BI team for sales dashboards.
Data Management: Stored ingested data in PostgreSQL alongside Delta tables for consumption by a web application.
Scaling: Expanded the project from a single-country to a multi-country deployment.
Key Contributions:
Client: Credit Suisse | Domain: Banking | Role: Data Engineer
Project Overview:
Developed an ingestion framework to migrate data from Cloudera (CDH) to Palantir Foundry (SFI), facilitating the transition from an on-premises platform to a cloud-based solution.
Key Responsibilities:
Client: ABN-AMRO | Domain: Banking | Role: Big Data Engineer
Project Overview:
Developed an ingestion framework to migrate on-premises data from PostgreSQL and Oracle databases to Hive. This included both full and incremental loads through Spark jobs. Additionally, created tables using Spark SQL transformations and ingested final data into Elasticsearch for querying via Kibana.
Key Responsibilities:
Client: Allstate | Domain: Insurance |Role: Developer
Project Overview:
Worked on the Interlink project for rules, rating, and underwriting, supporting the second-largest personal lines property and casualty insurer in the U.S.
Key Responsibilities:
Pyspark, Databricks, Azure Data Factory (ADF)
Sparksql, Hive, Sqoop, HDFS
Secondary : GitHub, Agile Practitioner, Scala, Azure,PostgreSQL, Oozie, Putty