Назад
Company hidden
1 день назад

Senior Software Engineer - Data Pipeline

160 800 - 193 000$
Формат работы
remote (только USA)/onsite
Тип работы
fulltime
Грейд
senior
Английский
b2
Страна
US
Вакансия из списка Hirify.GlobalВакансия из Hirify Global, списка международных tech-компаний
Для мэтча и отклика нужен Plus

Мэтч & Сопровод

Для мэтча с этой вакансией нужен Plus

Описание вакансии

Текст:
/

TL;DR

Senior Software Engineer (Data Pipeline): Design and develop high‑performance data converters for multi‑sensor autonomous‑driving data with an accent on accurate time alignment and robust handling of raw sensor logs. Focus on building and optimizing large‑scale ingestion and transformation pipelines capable of processing petabyte‑scale autonomous‑driving sensor data.

Location: Remote (US), Ann Arbor, MI

Salary: US Pay Range $160,800 - $193,000 USD

Company

hirify.global develops software for automated trucks to transform how the world moves freight.

What you will do

  • Design and develop high‑performance data converters for multi‑sensor autonomous‑driving data, ensuring accurate time alignment and robust handling of raw sensor logs.
  • Design, build, and optimize large‑scale ingestion and transformation pipelines (ETL/ELT) capable of processing petabyte‑scale autonomous‑driving sensor data, and automate them for reliable, production‑grade deployment.
  • Work with data formats such as ROS bags, MCAP, and custom binary encodings; establish standards for schema evolution and metadata integrity.
  • Implement automated data validation, quality checks, and lineage tracking to ensure reliability of production datasets.
  • Collaborate closely with ML, annotation, simulation, and perception teams to ensure cross‑team ownership of data products and deliver datasets that are consistent, semantically correct, and ready for downstream consumption.
  • Proactively assess current capabilities to identify areas for improvement proposing solutions that align with core strategy and operation.

Requirements

  • Bachelor's or Master's degree in STEM related field with 5+ years of working experience with cloud technologies & data operations.
  • Experience building or maintaining converters, decoders, or transformation pipelines for sensor‑rich data (e.g., lidar point clouds, camera streams, radar detections).
  • Understanding of multimodal data synchronization, timestamp alignment, and multi‑sensor calibration workflows.
  • Experience with distributed compute frameworks (Ray, Spark, Beam) and cloud‑based platforms like Anyscale and Databricks for large‑scale data‑pipeline execution.
  • Experience with high‑performance computing techniques, including vectorized data processing (NumPy), multithreaded or parallel execution, and GPU‑accelerated compute for optimizing large‑scale sensor‑data workloads.
  • Proficiency in Python, SQL, Shell Scripting.
  • Experience with major cloud providers like AWS, Google Cloud Platform (GCP) or Azure.

Nice to have

  • Working experience with design patterns & frameworks development for ML & operational data pipelines in the cloud
  • Familiarity with 3D labeling and CV annotation workflows.
  • Experience optimizing I/O‑heavy workloads , including columnar formats (Parquet, Arrow).
  • Knowledge of orchestration tools (Airflow, Argo, Prefect).
  • Hands‑on experience designing CI/CD automation for data services, including GitHub Actions, Databricks pipelines, and cloud‑native deployment workflows

Culture & Benefits

  • A competitive compensation package that includes a bonus component and stock options
  • 100% paid medical, dental, and vision premiums for full-time employees
  • 401K plan with a 6% employer match
  • Flexibility in schedule and generous paid vacation (available immediately after start date)

Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →