Job description
The Role
You'll maintain our data platform, working with pioneering AWS technologies and industry-standard tools like Databricks to build robust, scalable data pipelines and integrations. This isn't just about writing code – it's about creating systems that transform raw data into actionable insights and provisioning of data to core applications. You'll work on a rotation base with another engineer to provide 16 hours of coverage to the distributed team.
Responsibilities:
- Maintain and improve IT systems infrastructure, ensuring flawless operation and integration
- Develop, implement, and optimize data pipelines and integrations using industry-standard tools
- Automate data acquisition processes and optimize data delivery for real-time analytics
- Collaborate with cross-functional teams across various time zones to ensure seamless operation
- Provide timely updates and recommendations on operational issues and improvements
- Develop customized SQL queries for database solutions and ad hoc requests
- Ensure data integrity and fidelity across all systems
Requirements:
- Bachelor's degree in computer science or information technology
- 3+ years of experience in IT, including 2 years in data integration and pipeline development using AWS S3, Redshift, Athena ecosystem
- Extensive experience with SQL Queries, Python development, and common Python libraries
- Proficiency in PySpark/Databricks Delta Lake and AWS Cloud data integration
- Proven ability to manipulate, process, and extract value from large datasets
- Excellent communication skills and strong leadership capabilities
- High energy level and dedication to Thermo Fisher Scientific's 4i values
Preferred Qualifications:
- Experience in Life Sciences or Service organizations
- Familiarity with data mining, data science, and predictive analytics
- Experience with visualization tools such as Microsoft Power BI