Search by job, company or skills

DATAECONOMY

AWS Pyspark Data Engineer

5-7 Years
new job description bg glownew job description bg glownew job description bg svg
  • Posted 23 hours ago
  • Be among the first 10 applicants
Early Applicant

Job Description

About Us

About DATAECONOMY: We are a fast-growing data & analytics company headquartered in Dublin with offices inDublin, OH, Providence, RI, and an advanced technology center in Hyderabad,India. We are clearly differentiated in the data & analytics space via our suite of solutions, accelerators, frameworks, and thought leadership.

Job Description

We are seeking a highly skilled and experienced Senior Data Engineer to lead the end-to-end development of complex models for compliance and supervision. The ideal candidate will have deep expertise in cloud-based infrastructure, ETL pipeline development, and financial domains, with a strong focus on creating robust, scalable, and efficient solutions.

Key Responsibilities



  • Model Development: Lead the development of advanced models using AWS services such as EMR, Glue, and Glue Notebooks.



  • Cloud Infrastructure: Design, build, and optimize scalable cloud infrastructure solutions with a minimum of 5 years of experience.



  • ETL Pipeline Development: Create, manage, and optimize ETL pipelines using PySpark for large-scale data processing.



  • CI/CD Implementation: Build and maintain CI/CD pipelines for deploying and maintaining cloud-based applications.



  • Data Analysis: Perform detailed data analysis and deliver actionable insights to stakeholders.



  • Collaboration: Work closely with cross-functional teams to understand requirements, present solutions, and ensure alignment with business goals.



  • Agile Methodology: Operate effectively in agile or hybrid agile environments, delivering high-quality results within tight deadlines.



  • Framework Development: Enhance and expand existing frameworks and capabilities to support evolving business needs.



  • Documentation and Communication: Create clear documentation and present technical solutions to both technical and non-technical audiences.

Requirements

Required Qualifications:



  • 05+ years of experience with Python programming.



  • 5+ years of experience in cloud infrastructure, particularly AWS.



  • 3+ years of experience with PySpark, including usage with EMR or Glue Notebooks.



  • 3+ years of experience with Apache Airflow for workflow orchestration.



  • Solid experience with data analysis in fast-paced environments.



  • Strong understanding of capital markets, financial systems, or prior experience in the financial domain is a must.



  • Proficiency with cloud-native technologies and frameworks.



  • Familiarity with CI/CD practices and tools like Jenkins, GitLab CI/CD, or AWS CodePipeline.



  • Experience with notebooks (e.g., Jupyter, Glue Notebooks) for interactive development.



  • Excellent problem-solving skills and ability to handle complex technical challenges.



  • Strong communication and interpersonal skills for collaboration across teams and presenting solutions to diverse audiences.



  • Ability to thrive in a fast-paced, dynamic environment.

Benefits

Standard Company Benefits

More Info

Job Type:
Industry:
Employment Type:

Job ID: 134609713