- Full-time
- Remote in India
- Required Working hours: 9 AM-5 PM EST
Duties & Responsibilities
- Collaborate with the Team Lead and crossfunctional teams to gather and refine data requirements for Denials AI solutions.
- Design, implement, and optimize ETL/ELT pipelines using Python, Dagster, DBT, and AWS data services (Athena, Glue, SQS).
- Develop and maintain data models in PostgreSQL; write efficient SQL for querying and performance tuning.
- Monitor pipeline health and performance; troubleshoot data incidents and implement preventive measures.
- Enforce data quality and governance standards, including HIPAA compliance for PHI handling.
- Conduct code reviews, share best practices, and mentor junior data engineers.
- Automate deployment and monitoring tasks using infrastructure-as-code and AWS CloudWatch metrics and alarms.
- Document data workflows, schemas, and operational runbooks to support team knowledge transfer.
Qualifications
- Bachelor's or Master's degree in Computer Science, Data Engineering, or related field.
- 5+ years of handson experience building and operating productiongrade data pipelines.
- Solid experience with workflow orchestration tools (Dagster) and transformation frameworks (DBT) or other similar tools such (Microsoft SSIS, AWS Glue, Air Flow).
- Strong SQL skills on PostgreSQL for data modeling and query optimization or any other similar technologies (Microsoft SQL Server, Oracle, AWS RDS).
- Working knowledge with AWS data services: Athena, Glue, SQS, SNS, IAM, and CloudWatch.
- Basic proficiency in Python and Python data frameworks (Pandas, PySpark).
- Experience with version control (GitHub) and CI/CD for data projects.
- Familiarity with healthcare data standards and HIPAA compliance.
- Excellent problemsolving skills, attention to detail, and ability to work independently.
- Strong communication skills, with experience mentoring or leading small technical efforts.