We are seeking an experienced Engineer with strong expertise in Spark, Scala, GCP, Python/PySpark, Shell Scripting and SQL to build and optimize large-scale data solutions in a cloud environment.
4+ years of experience in Big Data and Data Engineering.
Strong hands-on expertise in Apache Spark (batch & streaming).
Proficiency in Scala, Python/PySpark, and Shell scripting.
Solid experience with GCP services (BigQuery, Dataproc, Airflow, etc.).
Advanced knowledge of Spark-SQL and SQL.
Understanding of distributed systems and performance optimization.
Roles & Responsibilities
Design and develop scalable data pipelines using Apache Spark and PySpark.
Implement data processing applications in Scala and Python.
Leverage Google Cloud Platform (GCP) services for data engineering workflows.
Write and optimize Spark-SQL and SQL queries for analytics and transformations.
Automate processes using Shell scripting and ensure system reliability.
Collaborate with cross-functional teams to deliver high-quality data solutions.