- Design, develop, and maintain data pipelines using GCP services like Dataflow, Dataproc, and Pub/Sub.
- Develop and implement data ingestion and transformation processes using tools like Apache Beam and Apache Spark.
- Manage and optimize data storage solutions on GCP, including Big Query, Cloud Storage, and Cloud SQL.
- Implement data security and access controls using GCP's Identity and Access Management (IAM) and Cloud Security Command Center.
- Monitor and troubleshoot data pipelines and storage solutions using GCP's Stackdriver and Cloud Monitoring tools.
- Collaborate with data experts, analysts, and product teams to understand data needs and deliver effective solutions.
- Automate data processing tasks using scripting languages like Python.
- Participate in code reviews and contribute to establishing best practices for data engineering on GCP.
- Stay up to date on the latest advancements and innovations in GCP services and technologies.
Your skills and experience
- 5+ years of experience as a Data Engineer or similar role.
- Proven expertise in designing, developing, and deploying data pipelines.
- In-depth knowledge of Google Cloud Platform (GCP) and its core data services (GCS, BigQuery, Cloud Storage, Dataflow, etc.).
- Strong proficiency in Python & SQL for data manipulation and querying.
- Experience with distributed data processing frameworks like Apache Beam or Apache Spark (a plus).
- Familiarity with data security and access control principles.
- Excellent communication, collaboration, and problem-solving skills.
- Ability to work independently, manage multiple projects, and meet deadlines
- Knowledge of Sustainable Finance / ESG Risk / CSRD / Regulatory Reporting will be a plus
- Knowledge of cloud infrastructure and data governance best practices will be a plus.
- Knowledge of Terraform will be a plus