Understand the business case and translate to a holistic solution involvingAWS CloudServices,PySpark, EMR, Python, DataIngestionandCloudDB Redshift / Postgres
PL/SQL developmentfor high volume data sets, logical and physical schema design
Proven experience with large, complex databaseor data lakeprojects in environments producing high-volume data
Demonstrated problem solving skills; familiarity with various root cause analysis methods; experience in documentingidentifiedproblems and driving resolutions.
Communicationwith stakeholders and business partners across all worldwide locations.
Understanding and aligningbusiness partnersacross locations doing requirements engineering andbusiness case mapping.
Leading projects:Working on complex, cross-functional projects, acting as a subject matter expert, but also experience in project manager role
Designing data products: Creating trusted, reusable data products building on data collection, transformation and curation.
Developing documentation: Creating functional and technical documentation that supports best practices
Design data pipeline:Establish design artifacts based on user requirements and engage with ETL Build teams(ETL framework design, data modeling, source-target-mapping, architecture diagrams)
Interaction with Data Governance teams- aligning on data models,lineageand relationships.
Data Analysisfor trouble shooting (e.g. data issues, performance issues) but also for BI & analytics interaction, potentially on large data sets
Interfacingwith architecture team, cloud engineering teams and vendors on designingstate of the art solutions.
Make recommendationsregardingenhancements and/or improvements, provide consulting for operational aspects
Other Responsibilities:
Customer/stakeholder focus.Ability to build strong relationships with Application teams, cross functionalITand global/local IT teams- as well as interfacing with vendors (e.g.AWS)
Advisingjunior engineers& Build Team: Providing guidance to junior engineers andETL Build team and BI/Analytics teams
Required Qualifications:
Bachelor or master sdegree.Preferablyin information technologyor electricalengineering- but sincejobfocuson business interaction also background in otherfields is welcome.
Verygood knowledge of data architecture approaches and trends, andhighinterestto applyand furtherdevelopthat knowledge, including understanding ofOLAP/OLTP, ML,genAI, modelling, statistics.
Problem solving experience& analyticsskills-e.g.with L6Scurriculum(green/black belt)
Good experienceinAWS Services, Big data, PySpark,EMR,Python, Cloud DB RedShift
Proven experience with large, complex data projects in environments producing high-volume data,
Proficiencyin SQL and PL/SQL
Excellent conceptual abilities pared withvery goodtechnical documentation skills,e.g.ability to understand and document complex data flows as part of business / production processes,
Familiarity with SDLC concepts and processes
Translatebusiness requirements into technical specifications. This involves critical thinking, problem-solving, and thecapacityto work with large datasets.
AdditionalSkill:
Experience using and developing on AWS services, AWScertification, AWS Solutions Architect
Proficiencyin programming languages such as Python,PySpark,SQL, Java, C++ is highly recommended.
Understanding of machine learning and AI technologies is becoming increasingly important for data architects
Experience in semiconductor industry
Knowledge ofSemistructureddatasets
Experience with analytics & reporting solutions and business intelligence tools
Experience in collecting, structuring and summarizing requirements in a data warehouse environment,
Knowledge of statistical data analysis and data mining,
Experience in test management, test case definition and test processes