Search by job, company or skills

Awign

Data Engineer

new job description bg glownew job description bg glownew job description bg svg
  • Posted 6 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

This is a remote position.

No of positions- 2
Experience- 5-7 + years
Work Location: Remote
Screening Checklist:
Proficiency in interpreting data transformation logic written in T-SQL and implementing equivalent processes within Databricks
Ability to design and implement data ingestion pipelines using Azure Data Factory (from source to RAW layer)
Basic knowledge of C# and Sql(atleast read the coding, no need to write)
Experience in collecting and analyzing performance metrics to optimize data ingestion pipelines
Competence in performing performance optimizations for Databricks read/write queries as needed
Job Overview
We are currently seeking experienced Data Engineers (57 years of experience) with strong expertise
in Databricks, PySpark, and Data Fabric concepts to contribute to an ongoing enterprise data
transformation initiative. The ideal candidates will have solid hands-on engineering skills, a good
understanding of modern data architectures, and the ability to work collaboratively within cross-functional
teams.
Key capabilities and expectations include:
Strong experience in understanding and translating data transformation logic written in TSQL and implementing equivalent, efficient transformations in Databricks using PySpark,
aligned with Data Fabric design principles.
Hands-on experience in designing and implementing data ingestion pipelines using Azure
Data Factory, enabling reliable data movement from source systems to the RAW and curated
data layers within a Data Fabric ecosystem.
Working knowledge of Data Fabric concepts, including metadata-driven pipelines, data
integration, orchestration, data lineage, and governance, with the ability to apply these principles
in day-to-day engineering tasks.
Experience in monitoring, collecting, and analyzing pipeline performance metrics to identify
inefficiencies and support optimization of data ingestion and processing workflows.
Practical experience in performance tuning and optimization of Databricks read and write
operations, including partitioning, file formats, and query optimization techniques.
Ability to collaborate closely with senior engineers and architects, contribute to design
discussions, follow best practices, and support the continuous improvement of the data platform.
Strong problem-solving skills, eagerness to learn, and the ability to work effectively with cross
functional teams, including data analysts, data scientists, and business stakeholders.
This role is ideal for professionals looking to deepen their expertise in Databricks and Data Fabric
architectures while contributing to scalable, well-governed, and high-performance enterprise data
solutions.

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 138856979

Similar Jobs