Overview:
Join a rapidly growing team building world-class, large-scale Big Data architectures. This is a hands-on coding role focused on developing and optimizing data solutions using modern cloud and distributed computing technologies.
Key Responsibilities:
Development & Engineering
- Write high-quality, scalable code using Python, Scala, or similar languages.
- Work with SQL, PySpark, Databricks, and Azure cloud environments.
- Optimize Spark performance and ensure efficient data processing pipelines.
- Apply sound programming principles including version control, unit testing, and deployment automation.
- Design and implement APIs, abstractions, and integration patterns for distributed systems.
- Define and implement ETL, data transformation, and automation workflows in parallel processing environments.
Client Interaction
- Collaborate directly with Fortune 500 clients to understand strategic data requirements.
- Align project deliverables with client goals and Sigmoid's strategic initiatives.
Innovation & Continuous Learning
- Stay current with emerging Big Data and cloud technologies to maximize ROI.
- Explore and productize new tools and techniques for large-scale data engineering.
Culture & Mindset
- Demonstrate strategic and out-of-the-box thinking.
- Be analytical, data-driven, and entrepreneurial in approach.
- Thrive in a fast-paced, agile, start-up environment.
- Balance leadership with hands-on contribution.
Qualifications
- Bachelor's or higher in Computer Science or related technical field.
- Experience: 8 to 14 Years
- Proven experience in Big Data, application development, and data management.
- Strong skills in Python, PySpark (functional and OOP).
- Effective written and verbal communication skills.