Search by job, company or skills

Aligned Automation

Data Engineer (PySpark and Apache Airflow)

new job description bg glownew job description bg glownew job description bg svg
  • Posted an hour ago
  • Be among the first 10 applicants
Early Applicant

Job Description

About Aligned Automation

At Aligned Automation, we live by our Better Together philosophy to build a better world. As a strategic service provider to Fortune 500 companies, we help digitize enterprise operations and drive impactful business strategies. Our purpose goes beyond projectswe strive to deliver meaningful, sustainable change that shapes a more optimistic and equitable future.

Our culture is deeply rooted in our 4CsCare, Courage, Curiosity, and Collaborationensuring that each employee is empowered to grow, innovate, and thrive in an inclusive workplace.

Job Summary

We are seeking a skilled Data Engineer with strong expertise in PySpark and Apache Airflow to design, build, and optimize scalable data pipelines. The ideal candidate should have experience in big data processing, workflow orchestration, and cloud-based data platforms.

Key Responsibilities

  • Design, develop, and maintain scalable ETL/ELT pipelines using PySpark
  • Build and manage workflow orchestration using Apache Airflow
  • Process large datasets using distributed computing frameworks (Spark)
  • Optimize data pipelines for performance, reliability, and scalability
  • Implement data quality checks and monitoring mechanisms
  • Work closely with Data Analysts, Data Scientists, and BI teams
  • Manage data ingestion from various sources (APIs, databases, flat files, streaming)
  • Troubleshoot and resolve pipeline failures
  • Implement CI/CD for data pipelines
  • Ensure data governance and security best practices

Required Skills

Technical Skills:

  • Strong hands-on experience in PySpark
  • Experience in Apache Airflow (DAGs, Operators, Scheduling)
  • Good understanding of Spark architecture
  • Strong SQL knowledge
  • Experience with data warehousing concepts
  • Experience with:
    • S3 / ADLS / GCS
    • Redshift / Snowflake / BigQuery
  • Knowledge of Git and version control
  • Understanding of REST APIs and data ingestion

Good to Have:

  • Experience with cloud platforms (AWS / Azure / GCP)
  • Experience with Kafka or streaming pipelines
  • Docker & Kubernetes knowledge
  • Delta Lake / Iceberg knowledge
  • Experience in CI/CD tools (Jenkins, GitHub Actions)
  • Experience in monitoring tools (Prometheus, Grafana)

Educational Qualification

  • Bachelor's or Master's degree in Computer Science, IT, Engineering, or related field

Soft Skills

  • Strong problem-solving skills
  • Good communication and collaboration skills
  • Ability to work in an agile environment
  • Ownership mindset and attention to detail

More Info

Job Type:
Industry:
Employment Type:

About Company

Job ID: 144598429