Назад
19 часов назад

AWS Databricks Data Engineer

Формат работы
hybrid
Тип работы
fulltime
Грейд
middle/senior
Английский
b2
Страна
US
vacancy_detail.hirify_telegram_tooltip Загружаем источник...

Мэтч & Сопровод

Покажет вашу совместимость и напишет письмо

Описание вакансии

AWS Data Engineer

Posted Date Feb 25, 2026 Employment Type Contract Experience Level Mid-Senior level Location Los Angeles Metropolitan Area Category ** Data Science ** Company **Rivago Infotech Inc **

H1b WorkableNear by Relocation only

Job Title: AWS Databricks Data EngineerLocation : Los Angeles CA (Hybrid)Hire type : FTE / CTH **Implementation partner - ************ End Client - ConfidentialInterview mode: Video/Virtual


Job Description – We are seeking a highly skilled AWS Data Engineer with strong expertise in SQL, Python, PySpark, Data Warehousing, and Cloud-based ETL to join our data engineering team. The ideal candidate will design, implement, and optimize large-scale data pipelines, ensuring scalability, reliability, and high performance. This role requires close collaboration with cross-functional teams and business stakeholders to deliver modern, efficient data solutions.

Key Responsibilities1. Data Pipeline Development

  • Build and maintain scalable ETL/ELT pipelines using Databricks on AWS.
  • Leverage PySpark/Spark and SQL to transform and process large, complex datasets.
  • Integrate data from multiple sources including S3, relational/non-relational databases, and AWS-native services.

2. Collaboration & Analysis Searching for Data Science roles that provide visa sponsorship? Connect with international employers through Data Science Jobs with Visa Sponsorship opportunities actively seeking talented professionals.

  • Partner with downstream teams to prepare data for dashboards, analytics, and BI tools.
  • Work closely with business stakeholders to understand requirements and deliver tailored, high‑quality data solutions.

3. Performance & Optimization

  • Optimize Databricks workloads for cost, performance, and efficient compute utilization.
  • Monitor and troubleshoot pipelines to ensure reliability, accuracy, and SLA adherence.
  • Apply query optimization, Spark tuning, and shuffle minimization best practices when handling tens of millions of rows.

4. Governance & Security

  • Implement and manage data governance, access control, and security policies using Unity Catalog.
  • Ensure compliance with organizational and regulatory data‑handling standards.

5. Deployment & DevOps

  • Use Databricks Asset Bundles for deployment of jobs, notebooks, and configuration across environments.
  • Maintain effective version control of Databricks artifacts using GitLab or similar tools.
  • Use CI/CD pipelines to support automated deployments and environment setups.

Technical Skills (Required)

  • Strong expertise in Databricks (Delta Lake, Unity Catalog, Lakehouse Architecture, Table Triggers, Workflows, Delta Live Pipelines, Databricks Runtime, etc.).
  • Proven ability to implement robust PySpark solutions. Explore our comprehensive directory of visa sponsorship jobs from employers worldwide who are ready to sponsor talented international professionals.
  • Hands‑on experience with Databricks Workflows & orchestration.
  • Solid knowledge of Medallion Architecture (Bronze/Silver/Gold).
  • Significant experience designing or rebuilding batch‑heavy data pipelines.
  • Strong background in query optimization, performance tuning, and Spark shuffle optimization.
  • Ability to handle and process tens of millions of records efficiently.
  • Familiarity with Genie enablement concepts (understanding required; deep experience optional).
  • Experience with CI/CD, environment setup, and Git-based development workflows.
  • Solid understanding of AWS cloud, including:
  • IAM
  • Networking fundamentals
  • Storage integration (S3, Glue Catalog, etc.)

Preferred Experience

  • Experience with Databricks Runtime configurations and advanced features.
  • Knowledge of streaming frameworks such as Spark Structured Streaming.
  • Experience developing real-time or near real-time data solutions.
  • Exposure to GitLab pipelines or similar CI/CD systems.

Certifications (Optional)

  • Databricks Certified Data Engineer Associate / Professional
  • AWS Data Engineer or AWS Solutions Architect certification

Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →

Текст вакансии взят без изменений

Источник -