Search by job, company or skills

Exl

Senior Data Engineer

Save
new job description bg glownew job description bg glownew job description bg svg
  • Posted 27 days ago
  • Be among the first 20 applicants
Early Applicant

Job Description

About the Role

We are looking for a curious and driven Data Engineer to join our team. In this role, you will architect and build scalable data lakehouse solutions across major cloud providers. You will move beyond simple script writing to engineer robust, production-grade pipelines using the Modern Data Stack.

You will act as both an individual contributor and a technical mentor, ensuring high code quality and leveraging AI-assisted development tools (like Cursor, Claude, or GitHub Copilot) to maximize efficiency and innovation.

Key Responsibilities

  • Big Data Engineering: Design, build, and maintain scalable ETL/ELT pipelines using PySpark and Advanced SQL to process massive datasets.
  • Platform Architecture: Architect and implement infrastructure on Databricks or Snowflake, leveraging cloud storage (S3/ADLS), serverless services, and modern data warehouses.
  • Transformation & Orchestration: Utilize DBT (Data Build Tool) for effective data transformation and manage job orchestration/scheduling.
  • Code Quality & Best Practices: Champion software engineering best practices, including version control (Git), writing comprehensive unit tests, and maintaining design/API documentation.
  • AI-Augmented Development: Actively utilize AI coding assistants (Cursor, Copilot, etc.) to accelerate development cycles and improve code efficiency.
  • Collaboration & Review: Conduct rigorous peer code reviews for PySpark logic, ETL analytics, and Machine Learning integration.
  • Mentorship: Lead and mentor junior developers, ensuring the team adheres to best coding practices and helping them grow into future leaders.

Qualifications

Must-Have (Core Competencies):

  • Expert PySpark Proficiency: Deep experience processing large-scale data using Spark/PySpark. You understand how distributed computing works under the hood.
  • Advanced SQL: You can write complex, performant queries and understand database optimization deeply.
  • Python Scripting: Strong ability to write clean, modular, and efficient Python code for data engineering pipelines.
  • Platform Experience: Proven track record working within Databricks or Snowflake environments.
  • Data Modeling: Strong understanding of database systems, data modeling (Star schema, Snowflake schema), and data architecture.
  • Engineering Mindset: Experience with CI/CD, unit testing, and integrating with existing codebases.
  • AI Adaptability:Proficiency/Comfort with AI-enabled software development. You should be comfortable using IDEs with GenAI tools (Cursor, VS Code with Copilot, etc.) to iterate faster.

Good-to-Have (Preferred Qualifications):

  • Experience with DLT Hub and orchestration platforms like Airflow/Prefect.
  • Experience with Modern Data Stack (Fivetran, Airbyte, DBT, etc.).
  • Familiarity with DuckDB for analytical processing.
  • Exposure to building applications using LLMs/GenAI (OpenAI SDK, Gemini, Anthropic)

What We Offer

  • Exposure to Cutting-Edge Tech: Hands-on work with multiple cloud providers, prominent data platforms and the latest AI tools.
  • Leadership Pathway: While you will contribute individually, we provide a clear pathway to grow into a leadership role where you guide the technical direction of the team.
  • Collaborative Environment: A supportive work culture that values curiosity, where you will work alongside data scientists and consultants solving real-world sports business problems.
  • Competitive Package: A competitive salary and benefits package designed to reward top talent.

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 145756183

Similar Jobs

Noida, India

Skills:

snowflake JavaSQL ServerTableauSqlJenkinsData QualityGitDockerData MigrationShell scriptingSparkVerticaMongoDBRest ApisOracleAzurePythonAWSEtlmicroservices architecture

Gurugram, Gurugram, India

Skills:

PysparkData ModelingAzureBuilding data pipelines

Gurugram, India

Skills:

snowflake Data QualityPysparkAWS GlueData WarehousingData GovernancePythonEtlELTData Catalog

Gurugram, Gurugram, India

Skills:

snowflake DevopsGoogle Cloud PlatformPysparkDatabricksAzurePythonSqlAWSAirflowdbt

Gurugram, Gurugram, India

Skills:

HadoopScalaPysparkAvroUnix Shell ScriptingJenkinsGitHiveSpark StreamingRDBMSBitbucketApache KafkaSparkorcParquetNoSQL databasesHDFS