Search by job, company or skills

CoinDCX

Data Engineer - 3

new job description bg glownew job description bg glownew job description bg svg
  • Posted 6 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

The CoinDCX Journey: Building the Future of Finance

At CoinDCX, our mission is clear - to make crypto and blockchain accessible to every Indian and enable them to participate in the future of finance.

As India's first crypto unicorn valued at $2.45B, we are reshaping the financial ecosystem by building safe, transparent, and scalable products that power adoption at scale.

We believe that change starts together. It begins with bold ideas, relentless execution and people who want to build what's next.

If you're driven by purpose and thrive in environments where your work defines the next chapter of an industry, you'll feel right at home here.

About The Role

  • This role combines core data engineering excellence with optional exposure to financial correctness & reconciliation pipelines.
  • You will design, build, and optimize data systems that handle high-volume, high-variety datasets , power analytics, reporting, financial systems, and operational products.
  • The work will be across ingestion, transformation, data modeling, data marts, and workflow orchestration, ensuring reliability, scalability, and data correctness across the entire platform.
  • Design and build scalable, high-performance data platforms leveraging Databricks, Lakehouse architectures, Spark ETL, and Kafka-based streaming pipelines.

What You'll Do

  • Build & Optimize Large-Scale Data Pipelines
  • Own end-to-end ETL/ELT pipelines using Databricks (PySpark, Spark SQL).
  • Build streaming and batch workflows processing billions of events per day.
  • Implement efficient ingestion frameworks using Kafka/MSK, Auto Loader, CDC.
  • Optimize jobs for performance, cost efficiency, cluster utilisation, and scalability.
  • Ensure pipelines meet defined SLAs for latency, throughput, and freshness.
  • Own Lakehouse Design & Data Modeling
  • Design bronze silver gold layers using Delta Lake best practices.
  • Build curated data marts for analytics, BI, finance, risk, and product teams.
  • Implement dimensional models, fact/event stores, and conformed dimensions.
  • Drive schema governance, catalog organization, versioning, and lineage.
  • Build High-Quality Data Integrations
  • Integrate with internal microservices, third-party APIs, streaming sources, and databases.
  • Develop connectors/pipelines for structured, semi-structured, and unstructured data.
  • Implement robust idempotency, incremental ingestion, and change data capture.
  • Build scalable workflows for multi-system integration (custody, payments, trading, CRM, etc.)
  • Workflow Orchestration & Reliability
  • Build highly reliable workflows using Airflow/Databricks Workflows/Lambda/Step Functions.
  • Implement retry logic, DLQs, backfill strategies, and pipeline auto-recovery.
  • Maintain >99.5% pipeline uptime for critical systems.
  • Data Quality, Observability & Governance
  • Implement DQ checks: schema, completeness, freshness, referential integrity.
  • Build monitoring and alerting for ingestion lag, pipeline failures, and cost anomalies.
  • Own documentation, runbooks, dashboards (Datadog/CloudWatch/Lakehouse monitoring).
  • Ensure compliance with internal governance (S3 structure, catalog rules, PII handling).
  • Cross-Functional Collaboration
  • Work closely with Analytics, Product, Finance, Ops, Risk, and Engineering teams.
  • Translate requirements into scalable data pipelines and optimized datasets.
  • Participate in design reviews, code reviews, architecture discussions.
  • Mentor SDE-1/SDE-2 engineers and uplift engineering excellence.

You'll Excel in This Role If You

Have strong hands-on experience in:

Data Engineering Core

  • Databricks (PySpark, Spark SQL, Delta Lake) mandatory
  • Kafka/MSK / Streaming pipelines
  • PySpark, Python expert level
  • Spark SQL optimisation & query tuning
  • ETL engineering for large-scale workflows
  • Designing data marts, fact tables, dimensions, business layers
  • Strong experience in data warehouse concepts

Cloud & Infrastructure

  • AWS (S3, Glue, Lambda, EMR/Databricks, DMS, IAM, CloudWatch)
  • Cluster management, autoscaling, job tuning
  • CI/CD for data pipelines

Data Integration

  • REST APIs, event ingestion, CDC, RDBMS ingestion (Postgres/MySQL)
  • Handling schema evolution & incremental loads

You'll Know You're Winning When

  • 4-6 key pipelines redesigned or optimized for performance/cost
  • High-reliability ingestion for multiple systems in production
  • Data marts operational with measurable improvement in query performance
  • Zero severe data-quality issues attributed to owned systems
  • Documented and automated workflows replacing manual processes

Hiring Process

Here's what your journey with us looks like:

  • Application Review We assess for skills, alignment, and intent
  • Recruiter Connect A short conversation to understand you better
  • Functional Round(s) Deep dive into your approach, craft, and problem-solving
  • Assignment / Simulation Round A take-home task or live problem-solving exercise to understand how you think and execute in real scenarios
  • Culture & Values Discussion A conversation to understand our ways of working and how you thrive best
  • Founder Conversation (Optional) For certain roles and senior levels, you may meet our founders to explore strategic alignment and long-term fit

Where We Work

We believe the best ideas emerge when people build together. Collaboration, speed and trust come alive when teams share the same space.

With this belief, we operate as a work-from-office organisation. This role is based out of our Bengaluru office, where energy, alignment and innovation move in real time.

Perks That Empower You

We believe great people deserve great experiences.

  • Design Your Own Benefits: Flexible perks to match your lifestyle
  • Unlimited Wellness Leaves: Rest and recharge as you need
  • Mental Wellness Support: Access to therapy and wellness resources
  • Learning Sessions: Bi-weekly learning and growth opportunities

Ready to Build What's Next

If you're looking for a role that gives you direct access to high-stakes decisions, deep impact and a chance to build the future of finance, this is it.

Join CoinDCX and help us make crypto accessible to every Indian, together.

More Info

Job Type:
Industry:
Function:
Employment Type:

About Company

Job ID: 137121041