Job Summary
Synechron is seeking a highly skilled
Azure Data Engineer with extensive expertise in Databricks and Lakehouse architecture to lead the development of scalable, high-performance data pipelines supporting enterprise analytics solutions. In this role, you will design, implement, and optimize large-scale data workflows across cloud platforms, ensuring compliance, security, and operational efficiency. Your contributions will enable data-driven decision-making and support our organization's digital transformation objectives with innovative data management practices.
Software Requirements
Required:
- Proven experience with Azure Cloud Services including Azure Data Factory, Azure Data Lake, Azure Databricks (latest runtime versions preferred) (5+ years)
- Deep expertise in Databricks components: Delta Lake, Unity Catalog, Lakehouse architecture, Delta Live Pipelines, and Table Triggers
- Strong hands-on knowledge of Spark and PySpark for large-scale data processing and transformation
- Experience developing and managing Databricks Asset Bundles for deployment automation
- Experience with SQL (Azure SQL, Data Warehouse, or relational databases) for data modeling and validation
- Familiarity with GitLab or similar version control tools for collaboration and artifact management
Preferred:
- Experience with streaming frameworks like Spark Streaming or Structured Streaming
- Knowledge of advanced Databricks runtime optimizations and configuration tuning
- Exposure to end-to-end data governance and security compliance frameworks (e.g., GDPR, HIPAA)
Overall Responsibilities
- Lead design and implementation of data pipelines supporting enterprise analytics, reporting, and AI integrations across the Azure cloud ecosystem
- Build, test, and deploy Lakehouse solutions using Databricks, Delta Lake, and related components
- Optimize big data workflows for performance, cost-efficiency, and scalability in cloud environments
- Develop automated deployment strategies using Databricks Asset Bundles and CD pipelines supporting continuous integration and delivery
- Manage data security, access controls, and compliance using tools like Unity Catalog and Azure security features
- Collaborate with analytics, data science, and security teams to incorporate AI/ML models into data workflows
- Conduct performance tuning, troubleshoot issues, and continuously improve data quality and pipeline reliability
- Maintain detailed documentation of architecture, data flow, security policies, and system configurations
- Lead efforts to migrate legacy systems to cloud-native architectures supporting scalability and operational resilience
Technical Skills (By Category)
Programming and Data Processing (Essential):
- PySpark, Spark SQL, Delta Lake (latest runtime preferred)
- Python for data scripting and automation
- SQL: Azure SQL, Data Warehouse, or equivalent
Frameworks & Libraries:
- Databricks Delta Lake, Unity Catalog, Delta Live Pipelines
- Spark structured streaming (preferred)
- Data governance and metadata management tools
Cloud Technologies:
- Azure Data Factory, Azure Data Lake, Azure Databricks, Azure Synapse (preferred)
- Cloud security best practices and access management
Data Management & Governance:
- Data lineage, data quality tools, security policies compliant with regulations
DevOps & Automation:
- CI/CD pipelines supporting Azure DevOps, GitLab, or Jenkins
- Infrastructure as Code: Terraform, Azure Resource Manager (ARM) templates
Experience Requirements
- 10+ years of experience designing, developing, and supporting large-scale data pipelines within cloud environments, preferably Azure
- Proven expertise in Databricks and Lakehouse architecture in enterprise settings
- Demonstrated experience integrating AI/ML workflows within data pipelines (preferred)
- Sound knowledge of data governance, security, and compliance practices supporting enterprise standards
- Experience supporting migration from legacy data platforms to cloud-native architectures
Day-to-Day Activities
- Architect and develop scalable data workflows supporting enterprise analytics, reporting, and AI initiatives
- Build, tune, and optimize Delta Lake and Lakehouse data solutions for performance and reliability
- Automate data pipeline deployments and infrastructure using Asset Bundles and CI/CD tools
- Manage data security, privacy, and compliance leveraging Unity Catalog and cloud security features
- Troubleshoot and resolve production issues, analyze system bottlenecks, and optimize data workflows
- Collaborate with data scientists, BI teams, and enterprise architects to refine infrastructure and data models
- Document architecture, data lineage, and security controls for audit and compliance readiness
- Support data migration, platform upgrades, and cloud infrastructure provisioning
Qualifications
- Bachelor's or Master's degree in Data Engineering, Computer Science, or related fields
- Certifications such as Azure Data Engineer Associate or equivalent are preferred
- Hands-on experience supporting regulated and secure data environments in enterprise organizations
Professional Competencies
- Strong analytical and troubleshooting skills for complex data workflows and pipelines
- Excellent stakeholder communication and collaboration skills
- Leadership qualities for mentoring team members and guiding best practices
- Adaptability to evolving data technologies, security standards, and industry regulations
- Results-oriented focus on operational excellence, security, and scalability in data environments
S YNECHRON'S DIVERSITY & INCLUSION STATEMENT
Diversity & Inclusion are fundamental to our culture, and Synechron is proud to be an equal opportunity workplace and is an affirmative action employer. Our Diversity, Equity, and Inclusion (DEI) initiative Same Difference is committed to fostering an inclusive culture promoting equality, diversity and an environment that is respectful to all. We strongly believe that a diverse workforce helps build stronger, successful businesses as a global company. We encourage applicants from across diverse backgrounds, race, ethnicities, religion, age, marital status, gender, sexual orientations, or disabilities to apply. We empower our global workforce by offering flexible workplace arrangements, mentoring, internal mobility, learning and development programs, and more.
All employment decisions at Synechron are based on business needs, job requirements and individual qualifications, without regard to the applicant's gender, gender identity, sexual orientation, race, ethnicity, disabled or veteran status, or any other characteristic protected by law.
Candidate Application Notice