
Search by job, company or skills
Hiring Now: GCP Data Engineer (68 Years Experience)
Location: Chennai, Hyderabad, Kolkata
Employment Type: Full-time (On-site)
We are seeking a proactive GCP Data Engineer proficient in PySpark to develop and maintain robust, large-scale data solutions.
Key Responsibilities
Data Pipeline Development: Design and implement batch and real-time data ingestion pipelines using Cloud Dataflow, Pub/Sub, and PySpark on Dataproc.
Transformation & Processing: Develop, optimize, and maintain distributed data processing jobs using PySpark and Apache Beam to clean, enrich, and aggregate massive datasets.
Data engineer with strong coding and technical architectural experience in Google Data Solution
Should be able to guide a new team to build algorithms and data products, resolve blockers and bugs and be the code custodian.
Work across the full data pipeline on projects (ingestion, data modelling, data warehousing, Data Quality validity).
Knowledge on CI/CD Pipelines
Hands-on Scheduling the Data Pipelines with dependency flow using any Cloud Services Airflow/Composer, Schedulers etc.
Preferably worked on Data Analytics Projects before
Value addition if hands-on on any of the Reporting tool Tableau /Power BI/Looker
Job ID: 135870881