About the Role:
We are looking for a highly skilled AWS Data Architect with deep hands-on expertise in data lake development and data engineering. The ideal candidate will combine strong technical capabilities with architectural oversight, ensuring efficient, scalable, and governed AWS data ecosystems.
Experience Required: 10 - 14yrs
Location: Bangalore/Chennai
Key Responsibilities:
- Design, build, and maintain AWS-based data lake ecosystems using S3, Glue, Athena, covering ingestion, transformation, and curation.
- Develop PySpark-based ETL pipelines with schema evolution, partitioning strategies, and query optimization for performance and cost-efficiency.
- Implement and enforce data governance, security, and compliance using Lake Formation, IAM, KMS, and CloudTrail.
- Integrate data pipelines with downstream services such as Redshift Spectrum, QuickSight, and SageMaker for analytics and ML enablement.
- Define data architecture frameworks covering data modeling, ingestion design, integration patterns, and analytical enablement.
- Implement data quality frameworks like Deequ or Great Expectations to ensure reliable and accurate data.
- Automate deployments and CI/CD pipelines using AWS CDK, Terraform, GitHub Actions, and CodePipeline.
- Ensure observability, monitoring, and recovery using CloudWatch and other AWS-native tools.
- Guide engineering teams in best practices, performance optimization, cost management, and governance alignment.