Search by job, company or skills

threatxintel

Databricks Architect (Lakehouse Platform)

Save
new job description bg glownew job description bg glownew job description bg svg
  • Posted 4 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

Company Description

ThreatXIntel is a growing Cybersecurity, IT Staffing, and Consulting company delivering end-to-end technology and security solutions.

We are hiring for our corporate client. ThreatXIntel is the official hiring partner for this requirement.

Job Overview

We are seeking a highly skilled Databricks Architect to lead the design and implementation of enterprise-scale Lakehouse data platforms.

The ideal candidate will have deep expertise in Databricks, Apache Spark, Delta Lake, and cloud-native architectures, with proven experience delivering end-to-end data engineering solutions across batch and real-time pipelines.

This role requires strong architectural leadership, hands-on development capability, and stakeholder collaboration.

Key Responsibilities

Architecture & Design

  • Architect and implement enterprise-grade Lakehouse solutions using Databricks
  • Design scalable cloud-based data platforms integrating multiple data sources
  • Define data architecture standards, governance, and best practices

Data Engineering & Pipelines

  • Build end-to-end ETL/ELT pipelines using PySpark, Scala, and SQL
  • Develop batch and real-time streaming pipelines using Spark
  • Design incremental loading frameworks and metadata-driven ingestion pipelines

Databricks Ecosystem

  • Implement and manage:
  • Delta Live Tables
  • Autoloader
  • Structured Streaming
  • Databricks Workflows
  • Integrate with orchestration tools like Apache Airflow

Data Modeling & Governance

  • Design 3NF, dimensional models, and enterprise data warehouse solutions
  • Implement data quality frameworks and governance standards
  • Manage Unity Catalog with fine-grained security and access control

Performance Optimization

  • Optimize Spark jobs, pipelines, and compute resources
  • Ensure scalability, reliability, and cost efficiency

DevOps & CI/CD

  • Implement CI/CD pipelines and deployment strategies
  • Drive DevOps best practices for data engineering

Leadership & Collaboration

  • Provide technical leadership and architectural guidance
  • Collaborate with stakeholders to translate business needs into scalable solutions

Required Skills (Mandatory)

  • Deep expertise in Databricks and Lakehouse architecture
  • Strong experience with Apache Spark (batch & streaming)
  • Hands-on experience with Delta Lake
  • Strong programming skills in Python, PySpark, Scala, and SQL
  • Experience with distributed data processing systems
  • Strong experience in data modeling and data warehousing concepts
  • Experience with real-time data processing architectures
  • Hands-on experience with Unity Catalog and data governance

More Info

About Company

Job ID: 147364697