Search by job, company or skills

Awign

Sr. GCP Data Engineer

new job description bg glownew job description bg glownew job description bg svg
  • Posted a day ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Job Description : Sr. GCP Data Engineer

Experience : 7 to 12 Years

Work location: Chennai, Bangalore, Hyderabad, Pune-Hybrid

Shift Timing : 2 to 11 PM

Interview process : L1 and L2 round

Job description:

  • 5+ years experience
  • Should have experience in GCP BigQuery, DataProc(PySpark)
  • Good to have experience on Informatica

Requirements

  • How do you upload files to GCS
  • Using Google Cloud Console (UI) drag and drop files into a bucket.
  • Using gsutil CLI:
  • How do you query data in BigQuery

Write SQL queries in:

  • BigQuery Console (UI)
  • bq CLI bq query SELECT ...

APIs or client libraries (Python, Java, etc.)Cluster monitoring with Cloud Monitoring & alerts.Upload the script to the DAGs folder in Composer's GCS bucket Composer automatically deploys it.

  • What is the purpose of a Dataproc job Dataproc job is a task you submit to a cluster to process data, such as:
  • Spark job (Scala/Python/Java)
  • Hive query
  • Pig script
  • Hadoop MapReduce job
  • How do you handle errors in Dataproc jobs
  • Job retry policies (set retries in workflow templates).
  • Error logs in Cloud Logging (examine stdout/stderr).
  • Graceful failure handling with workflow DAG (e.g., skip/stop).
  • How do you create an Airflow DAG in Cloud Composer
  • Write a Python script defining the DAG (dag_id, schedule_interval, default_args).
  • Define tasks using operators (e.g., PythonOperator, BashOperator, BigQueryOperator)

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 135072545