Primary Title: Data Engineer
About The Opportunity
A specialist talent and HR solutions firm serving enterprise technology, fintech, and retail clients is hiring an on-site Data Engineer in India. This role sits within high-growth data platform programs focused on building scalable, reliable ETL pipelines and analytics-ready data sets for business stakeholders and downstream ML consumers.
Role & Responsibilities
- Design, build and operate robust ETL/ELT data pipelines to ingest, transform and deliver large-scale datasets for analytics and ML.
- Author PySpark jobs and SQL-based transformations to implement reusable, performant data models and tables.
- Orchestrate workflows using Apache Airflow; implement scheduling, retries, and monitoring for production jobs.
- Work with cloud storage and warehousing solutions (S3, Redshift) to optimise cost, partitioning and query performance.
- Collaborate with data consumers and engineers to define schemas, data contracts, and data quality checks; implement automated testing and observability.
- Troubleshoot production incidents, tune pipeline performance, and implement CI/CD for data workloads.
Skills & Qualifications
Must-Have (skills)
- Python
- PySpark
- SQL
- Apache Airflow
- AWS S3
- ETL pipelines
Preferred (skills)
- AWS Redshift
- Apache Kafka
- dbt
Additional Qualifications
- 3+ years of hands-on experience building production data pipelines and data models in cloud environments.
- Proven track record working on on-site, collaborative engineering teams in India; strong ownership of production systems.
- Familiarity with CI/CD for data engineering, automated testing, and data observability practices.
Benefits & Culture Highlights
- Work on enterprise-level data platform projects with clear ownership and growth pathways.
- Collaborative engineering culture with emphasis on automation, code reviews and operational excellence.
- Competitive compensation aligned to market and opportunities for technical skill development.
Location: On-site in India. Candidates must be available for full-time on-site engagement. To apply, please submit a resume highlighting pipeline projects, code samples or links to relevant repositories and a short note on your role in delivered data solutions.
Skills: aws,hadoop,big data,dbt,sql,spark,apache kafka,python,azure,etl