Responsibilities
- Build and optimize ETL/ELT pipelines using Python, SQL, and AWS services.
- Architect scalable data solutions across S3, Redshift, Lambda, EKS/ECS, Step Functions, API Gateway.
- Develop highquality, productionready Python code (APIs, packages, unit tests).
- Optimize SQL workloads and Redshift performance.
- Manage orchestration using Airflow/Prefect and IaC using Terraform/CloudFormation.
- Lead CI/CD development (GitHub Actions/CodePipeline) and container workflows (Docker, Kubernetes).
- Drive production reliability, debugging, and performance tuning.
- Mentor engineers and collaborate with analytics/product teams.
Required Skills
- Advanced Python for scalable data engineering.
- Advanced SQL with query tuning and complex data transformation experience.
- Strong AWS expertise across compute, storage, and data services.
- Hands-on with Docker, Kubernetes (EKS), Git, Linux shell.
- Experience building enterprise-grade data platforms and pipelines.
- Strong problem-solving and production troubleshooting skills.
Good to Have
- Data modeling experience (Kimball, Data Vault).
- Streaming tools (Kafka/Kinesis).
- Data quality frameworks (Great Expectations, Soda).
AWS,Redshift DB,Python,SQL,Glue,Lamda,Athena