Эта вакансия в архиве

Посмотреть похожие вакансии ↓
Company hidden
обновлено 8 дней назад

GCP Data Platform Engineer (Cloud)

Формат работы
hybrid
Тип работы
project
Грейд
middle/senior
Английский
b2
Страна
Poland

Описание вакансии

Текст:
/

TL;DR

GCP Data Platform Engineer (Data Engineering): Developing reusable frameworks for data processing and testing on GCP (e.g., BigQuery, Dataflow/Dataproc, Composer) with an accent on building and maintaining batch and streaming data ingestion pipelines from various sources into GCP. Focus on monitoring, optimizing and securing data pipelines in line with data governance and compliance standards.

Location: Warszawa, ul. Marynarska 12. Hybrid (2-3 days a week in the office)

Company

hirify.global is working with technologies like GCP, Spark, Python, Kubernetes, BigQuery, Vertex AI, Terraform, Looker to move analytics from onpremise to GCP and build its data architecture and data model from the ground up.

What you will do

  • Develop reusable frameworks for data processing and testing on GCP (e.g., BigQuery, Dataflow/Dataproc, Composer).
  • Build and maintain batch and streaming data ingestion pipelines from various sources (databases, Kafka/MQ, APIs, files) into GCP.
  • Implement automated tests and data quality checks for data pipelines.
  • Collaborate with analysts and data scientists to deliver reliable, well‑documented datasets.
  • Monitor, optimize and secure data pipelines in line with data governance and compliance standards.

Requirements

  • 3+ years of experience as a Data Engineer in a data‑driven environment.
  • Experience in large‑scale data migration or cloud transformation projects.
  • Experience with modern data platform patterns, including data lakehouse architectures on GCP (Cloud Storage + BigQuery).
  • Hands‑on experience with GCP data services (BigQuery, Cloud Storage, Pub/Sub, Dataflow/Dataproc, Composer, Looker, Vertex AI).
  • Hands‑on experience with Infrastructure‑as‑Code (IaC) tools, including Terraform.
  • Strong SQL skills and experience with large‑scale data processing (Spark required; batch and streaming).
  • Proficiency in Python and/or Scala or Java.
  • Very good command of English (spoken and written).

Nice to have

  • Degree in Computer Science, Data Science or a related field.
  • Experience with data governance, metadata and data quality tools.
  • Experience collaborating with business stakeholders.

Culture & Benefits

  • Join a new, strategic data transformation project.
  • Strong focus on business value creation and CX of customers.
  • Integrate diverse, high volume data sources, design streaming and batch processing layers, implement data governance, lineage, data quality and data security.
  • Set up CI/CD and monitoring/SLOs to shorten the path from question to answer for our business and create a solid foundation for AI/LLM driven solutions.