We are seeking a Senior Data Architect with deep expertise in Microsoft Azure and Microsoft Fabric to design, build, and modernize enterprise-scale data platforms.
This role requires strong solutioning capability, experience building reusable accelerators for data migration and modernization, and hands-on development skills using Python, Spark, Azure Databricks, and Azure Synapse.
The architect will lead end-to-end architecture, guide delivery teams, and remain actively involved in implementation, optimization, and governance.
Location: Bangalore/ Hyderabad/ Chennai
Experience: 14+ years
Key Responsibilities
Architecture & Solution Design
- Design modern Azure data architectures using:
- Microsoft Fabric (Lakehouse, Warehouse, OneLake, Direct Lake, Pipelines)
- Azure Synapse Analytics
- Azure Databricks
- Azure Data Lake Gen2
- Define medallion architecture (Bronze / Silver / Gold) and domain-driven data products.
- Architect Lakehouse, Warehouse, and hybrid analytics solutions.
- Lead legacy platform Azure / Fabric modernization initiatives.
- Create reference architectures, design standards, and best practices.
Hands-on Implementation
- Develop and optimize PySpark / Spark SQL workloads at scale.
- Build and orchestrate data pipelines using:
- Fabric Pipelines
- Azure Data Factory
- Synapse Pipelines
- Implement batch and near-real-time ingestion patterns.
- Design and manage Delta Lake-based Lakehouse architectures.
- Build ingestion, transformation, and validation frameworks using Python.
Accelerators & Migration Frameworks
- Design and build reusable accelerators for:
- On-prem / legacy DW Fabric migration
- SQL / Stored Procedures Spark & Fabric
- Metadata-driven ingestion and transformation
- Automated reconciliation, audit logging, and data validation
- Create CI/CD accelerators for Fabric, Synapse, and Databricks using:
- Azure DevOps
- GitHub Actions
- Standardize deployment templates, pipelines, and operational patterns.
Analytics & Semantic Layer
- Architect semantic models and curated layers for analytics using:
- Power BI
- Fabric Warehouse
- Direct Lake / Direct Query / Import modes
- Enable self-service analytics with governed datasets.
- Optimize data models for performance, scale, and cost.
Governance, Security & Optimization
- Define data governance and security frameworks:
- RBAC, RLS, OLS
- Workspace and domain strategies in Fabric
- Implement data cataloging, lineage, and quality using Microsoft Purview.
- Optimize Fabric capacity usage and cost.
- Establish monitoring, alerting, and operational runbooks.
Leadership & Stakeholder Engagement
- Act as technical authority for Azure data solutions.
- Mentor data engineers and review architecture and code.
- Collaborate with business, platform, and security teams.
- Support pre-sales, PoCs, architecture reviews, and proposals.
Required Technical Skills
Azure Data Platform
- Microsoft Fabric
- Azure Synapse Analytics
- Azure Databricks
- Azure Data Factory
- Azure Data Lake Gen2
- Power BI
Programming & Processing
- Python (advanced)
- Apache Spark / PySpark
- Spark SQL
- Strong SQL optimization and performance tuning
Data Engineering & Architecture
- Data modeling (Star / Snowflake / Dimensional)
- Batch and near-real-time processing
- Metadata-driven pipelines
- Data validation, reconciliation, and audit frameworks
- CI/CD for Azure data platforms
Good to Have
- Experience with Dynamics 365 / ERP / SaaS data sources
- Exposure to Apache Iceberg and open table formats
- Knowledge of Real-Time Analytics / Event Streams
- Microsoft certifications (Fabric, Azure Data Engineer, Azure Architect)
Soft Skills
- Strong solutioning and architectural thinking
- Ability to stay hands-on while leading
- Excellent stakeholder communication
- Experience working in large enterprise programs