Job Description
Job Description
Migration & Architecture
Lead Big Data platform migrations from on‑prem Hadoop / legacy data platforms to GCP
Design target-state architectures using BigQuery, Dataproc, Dataflow, GCS
Define migration strategies (re-host, re-platform, re-engineer)
Modernize ETL pipelines to cloud-native or Spark-based solutions
Development & Engineering
Architect and develop large-scale PySpark applications
Refactor existing Spark, Hive, or MapReduce jobs for GCP
Optimize data pipelines for performance, scalability, and cost
Implement batch and streaming workloads on GCP
Responsibilities
BigQuery
Cloud Storage (GCS)
Dataproc
Dataflow
Pub/Sub