10+ years of experience designing and implementing scalable data solutions, including end-to-end ETL/ELT pipelines using Databricks (PySpark) and Snowflake.
Build and manage secure, scalable data platforms using PostgreSQL and DynamoDB tailored to application needs.
Develop real-time and batch ingestion pipelines from diverse sources such as APIs, logs, files, and databases.
Apply transformation logic to clean, enrich, and normalize data for analytics and reporting.
Optimize pipeline performance and storage costs using best practices in partitioning, indexing, and query tuning.
Implement robust data quality checks, access controls, backup strategies, and governance policies.
Lead data strategy and unify large-scale existing platforms.
Establish data architecture principles, standards, and patterns for consistent adoption.
Design and oversee data lakes, data warehouses, and data mesh architectures.
Guide engineering teams on ingestion, transformation, storage strategies, and CI/CD best practices in production environments.