Job Description
Key Responsibilities:
- Design, develop, and maintain data pipelines and ETL/ELT workflows using GCP-native tools and services.
- Build and optimize data warehouses using Snowflake.
- Write complex and efficient SQL queries for data transformation, analysis, and reporting.
- Collaborate with analysts, data scientists, and business stakeholders to understand data needs and deliver reliable solutions.
- Implement data governance, security, and monitoring best practices across GCP projects.
- Tune queries and optimize performance of large-scale datasets.
- Automate workflows using Cloud Composer (Airflow) or similar orchestration tools.
Required Skills & Qualifications:
- 3+ years of experience in a data engineering or data platform role.
- Strong hands-on experience with Snowflake data warehousing
- Expert-level skills in SQL able to write optimized, scalable, and complex queries.
- Experience with data modeling (star/snowflake schema), partitioning, clustering, and performance tuning in a data warehouse.
- Familiarity with modern ELT tools such as dbt, Fivetran, or Cloud Data Fusion.
- Experience in Python or similar scripting language for data engineering tasks.
- Understanding of data governance, privacy, and Google Cloud Platform services, especially BigQuery, Cloud Storage, Dataflow, Pub/Sub, Composer.