Search by job, company or skills

Hitya Global

Senior Data Engineer - ETL

6-8 Years
new job description bg glownew job description bg glownew job description bg svg
  • Posted 2 days ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Key Responsibilities

  • Design, build, and optimize scalable data pipelines and data models.
  • Manage and administer Snowflake environments, including virtual warehouses, secure shares, clustering, tagging, masking, dynamic tables, and performance tuning.
  • Ensure cost efficiency in Snowflake workloads through monitoring and optimization techniques.
  • Implement and maintain data governance, data quality checks, and metadata standards.
  • Collaborate with cross-functional teams to understand data needs and deliver robust solutions.
  • Develop automated scripts and transformations using Python for data processing.
  • Leverage AI-assisted tools (GitHub Copilot, Claude Code) to improve design, coding, documentation, and reviews.
  • Monitor system health, performance, and data reliability using industry tools and best practices.
  • Contribute to architectural discussions, documentation, and design specifications.
  • Use Git effectively for branching strategies, code reviews, and version control.
  • Support production environments and troubleshoot performance, quality, and pipeline issues.

Requirements

  • 6+ years of experience as a Data Engineer, Software Developer, or Administrator focused on data modeling, governance, and platform administration.
  • 3+ years of hands-on Snowflake experience including virtual warehouse management, clustering, secure shares, cost control, tagging, masking, dynamic tables, and performance optimization.
  • Expert-level SQL capability for analytics and database engineering.
  • Intermediate Python skills for data pipelines and automation.
  • Experience using AI-assisted code tools (GitHub Copilot, Claude Code).
  • Strong experience across MSSQL, PostgreSQL, OLAP/OLTP systems.
  • Familiarity with ETL/ELT tools, schema design for high-volume systems, and data health monitoring.
  • Experience implementing data governance principles and standards.
  • Solid understanding of cloud design principles (SaaS, PaaS, IaaS).
  • Experience with structured file formats : XML, JSON, Parquet, CSV, Fixed Length.
  • Hands-on experience with Git including branching, merging, and git-flow.

Good To Have

  • Experience with large-scale cloud data platforms.
  • Knowledge of infrastructure automation.
  • Exposure to DevOps-driven workflows.

(ref:hirist.tech)

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 138102489