Resume
Swagat Dash
Kolkata, West Bengal, India
s************1@g****.com
*******893
Male
Personal Information
About me
10+ years of experience in Big Data Engineering, Cloud Data Platforms, and Data Pipeline Development. 6+ years of experience on GCP services including BigQuery, Dataflow, Pub/Sub, Cloud SQL, Firestore, and GKE. Expertise in Google Cloud Platform (GCP), Azure, Databricks, and Hadoop Ecosystem. Strong hands-on experience with Spark, Kafka, Airflow, Hive, Snowflake, and PySpark. Skilled in designing scalable ETL/ELT pipelines across batch and streaming architectures using ADF, Apache Spark, and Databricks. Involved in real-time data processing, machine learning operations, prompt engineering, and LLM integrations. Experience across multiple cloud platforms (Azure, AWS, GCP) for scalable and fault-tolerant system architecture. Proficient in DevOps, CI/CD, containerization (Docker, Kubernetes), and monitoring with GCP tools. Excellent communication skills, collaborative with cross-functional teams, and experienced in Agile methodologies.
Job Title
Data Engineer
Work Experience

IBM India Pvt Ltd
Full time• Dec 01 2014 - Present
Project: Enterprise Big Data Platform Development
Created enterprise data platform using Hadoop ecosystem tools to support ingestion, processing, and analytics.
Technologies: Hadoop, Hive, Sqoop, Cloudera, Kafka, Shell Scripting, Linux, Jenkins, Eclipse, Git, Oozie, Talend, Agile Methodology

Infosys
Full time• Oct 01 2018 - Present
Project: GCP-Based Data Lake Platform
Built Spark-based data pipelines for both batch and streaming using PySpark, integrated with GCP and containerized environments.
Technologies: GCP, PySpark, Airflow, BigQuery, GCS, Kubernetes, Cloud Functions, Docker, Jenkins, PostgreSQL, Hive, SQL, Terraform, Golang

EY GDS
Full time• Mar 01 2021 - Present
Project: Cloud Data Transformation and Pipeline Management
Built and managed GCP-based infrastructure and pipelines to support batch and streaming analytics use cases. Focused on performance tuning, cost optimization, and scalable streaming applications.
Technologies: Azure Databricks, ADF, GCP (BigQuery, GKE, Composer), PySpark, Git, Bitbucket, SQL Server)

Humancloud Technologies
Full time• Aug 01 2022 - Present
Project: End-to-End Data Lakehouse Platform using Azure & GCP
Led the development of a modern data architecture using Azure and GCP components for ingestion, transformation, and analytics using Databricks, ADF, and Snowflake. Focused on delivering scalable ETL pipelines, optimizing performance, and ensuring high data quality.
Technologies: Azure Data Factory, Azure Databricks, PySpark, Azure Functions, Snowflake, Kafka, NiFi, Power Bi, Docker, Kubernetes, SQL, Server, DB2, JSON, Git, Hive, ADLS Gen2, HDInsight, Python, Shell Scripting
Education

BPUT
Bachelor’s in Electrical and Electronics Engineering • 2012
Skills
Certifications