Job Title: GCP Data Engineer
Location: Pune, India
Experience: 4 to 7 Years
Job Type: Full-Time
Job Summary:
We are looking for a highly skilled
GCP Data Engineer with 4 to 7 years of experience to join our data engineering team in
Pune. The ideal candidate should have strong experience working with
Google Cloud Platform (GCP), including
Dataproc,
Cloud Composer (Apache Airflow), and must be proficient in
Python/Shell Scripting, SQL, and
Apache Spark. The role involves designing, building, and optimizing data pipelines and workflows to support enterprise-grade analytics and data science initiatives.
Key Responsibilities:
- Design and implement scalable and efficient data pipelines on GCP, leveraging Dataproc, BigQuery, Cloud Storage.
- Develop and manage ETL/ELT workflows using Apache Spark, SQL, and Python.
- Orchestrate and automate data workflows using Cloud Composer (Apache Airflow).
- Build batch and streaming data processing jobs that integrate data from various structured and unstructured sources.
- Optimize pipeline performance and ensure cost-effective data processing.
- Collaborate with data analysts, scientists, and business teams to understand data requirements and deliver high-quality solutions.
- Implement and monitor data quality checks, validation, and transformation logic.
Required Skills:
- Strong hands-on experience with Google Cloud Platform (GCP)
- Proficiency with Dataproc for big data processing and Apache Spark
- Expertise in Python and SQL for data manipulation and scripting
- Experience with Cloud Composer / Apache Airflow for workflow orchestration
- Knowledge of data modeling, warehousing, and pipeline best practices
- Solid understanding of ETL/ELT architecture and implementation
- Strong troubleshooting and problem-solving skills
Preferred Qualifications:
- GCP Data Engineer or Cloud Architect Certification.
- Familiarity with BigQuery, Dataflow. Exposure to Agile methodologies and team collaboration tools.