Search by job, company or skills

Blue Yonder

Sr. Technical Consultant - Cloud - Snowflake

new job description bg glownew job description bg glownew job description bg svg
  • Posted a day ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Scope:

  • The Data Platform team partners with business and engineering groups to deliver scalable Snowflake solutions.

  • The L4 role focuses heavily on execution-translating architectural patterns into highly optimized, secure, and reliable data pipelines and models for specific business domains.

  • The position involves automating ingestion, optimizing complex queries, and ensuring platform stability while guiding junior data engineers.

Technical Environment :

  • Snowflake Platform: Multi-Cluster Warehouses, Snowpipe, Tasks, Streams, Zero-Copy Cloning, Time Travel, Apache Iceberg tables.

  • Data Engineering & Scripting: Advanced SQL, Python (Snowpark), Java/Scala (UDFs/UDTFs), dbt (Data Build Tool).

  • Integrations & Orchestration: Apache Airflow, Fivetran, Kafka, Spark, Trino, external catalogs (AWS Glue, Polaris).

  • Governance & Security: Hierarchical RBAC, Dynamic Data Masking, Row Access Policies, Object Tagging, Secure Data Sharing.

  • Platform Enhancements: Snowpark Container Services, Snowflake Cortex (AI/ML), Search Optimization Service, Materialized Views.

  • DataOps/Agile: CI/CD pipelines, Git, GitHub Actions/GitLab, Terraform (Infrastructure-as-Code), Agile delivery.

What you'll do:

  • Design & Architect: Design robust dimensional data models and domain-specific data pipelines. Implement standard FinOps and security patterns defined by senior architects.

  • Develop & Deliver: Configure continuous data ingestion (Snowpipe/Streams), write modular transformations using dbt and Python (Snowpark), and build task orchestrations.

  • Guide & Govern: Review code and data models for team members, ensuring adherence to CI/CD standards and SQL best practices.

  • Operate & Optimize: Identify and rewrite bottleneck queries, optimize micro-partition clustering, and handle L3 technical escalations for pipeline failures.

What we are looking for:

  • Bachelor's degree in Computer Science, Data Engineering, or a related technical field.

  • 6-8 years of IT/Data experience, with 3-5 years specifically in deep Snowflake development and pipeline architecture.

  • Strong expertise in Snowflake core architecture, caching layers, and warehouse sizing.

  • Deep proficiency in Advanced SQL, Python, and data modeling (Dimensional/Kimball).

  • Hands-on experience with dbt, Airflow, and CI/CD pipelines for database deployments.

Our Values


If you want to know the heart of a company, take a look at their values. Ours unite us. They are what drive our success - and the success of our customers. Does your heart beat like ours Find out here:

All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status.

More Info

Job Type:
Employment Type:

About Company

Job ID: 144994677