TL;DR
Senior Data Engineer (GCP, Python): Designing, building, monitoring, and supporting large-scale data processing pipelines with an accent on secure, automated, and scalable solutions on GCP. Focus on maximizing cloud compute efficiency, ensuring data governance, quality, and efficient consumption for operational and scientific purposes.
Location: Mumbai, India (Hybrid - 3 days onsite)
Company
hirify.global is an audience-first platform focused on big data analytics and advertising, delivering scaled, relevant, and high-performing audiences without relying on personal data.
What you will do
- Design, build, monitor, and support large scale data processing pipelines.
- Support and mentor team members to advance capabilities and capacity.
- Help hirify.global explore and exploit new data streams for commercial and technical growth.
- Work closely with Product to deliver against fast-paced decisions.
Requirements
- 8+ years direct experience delivering robust performant data pipelines within the constraints of direct SLA’s and commercial financial footprints.
- Proven experience in architecting, developing, and maintaining Apache Druid and Imply platforms, with a focus on DevOps practices and large-scale system re-architecture.
- Mastery of building Pipelines in GCP maximizing the use of native and native supporting technologies e.g. Apache Airflow.
- Mastery of Python for data and computational tasks with fluency in data cleansing, validation and composition techniques.
- Hands-on implementation and architectural familiarity with all forms of data sourcing i.e streaming data, relational and non-relational databases, and distributed processing technologies (e.g. Spark).
- Excellent working understanding of server-side Linux.
Nice to have
- Experience optimizing both code and config in Spark, Hive, or similar tools.
- Practical experience working with relational databases, including advanced operations such as partitioning and indexing.
- Knowledge and experience with tools like AWS Athena or Google BigQuery.
- Understanding and ability to innovate, apply, and optimize complex algorithms and statistical techniques to large data structures.
- Experience with Python Notebooks, such as Jupyter, Zeppelin, or Google Datalab.
Culture & Benefits
- Work in a growing team with big responsibilities and exciting challenges, aiming for the next 10x level of scale and intelligence.
- Adherents of Lean Development, working with significant amounts of freedom and ambitious goals.
- Committed to creating a powerful culture of inclusivity and belonging that embraces diversity and encourages authentic selves at work.
- Values leadership, client-obsession, teamwork, and being action-oriented.
- Equal opportunity employer, opposing unlawful discrimination and harassment.
Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →