Описание вакансии подробное, но отсутствует информация о зарплате, что затрудняет оценку общей привлекательности. Роль требует смешанных навыков, что может указывать на высокую нагрузку.
Кликните для подробной информации
Нет информации о зарплатеПерегруженная рольГибридный формат работыСовременный стек технологий
Оценка от Hirify AI
Мэтч & Сопровод
Покажет вашу совместимость и напишет письмо
Создать профиль и узнать мэтч
Описание вакансии
AWS Data Engineer
Posted Date Feb 25, 2026 Employment Type Contract Experience Level Mid-Senior level Location Los Angeles Metropolitan Area Category ** Data Science ** Company **Rivago Infotech Inc **
H1b WorkableNear by Relocation only
Job Title: AWS Databricks Data EngineerLocation : Los Angeles CA (Hybrid)Hire type : FTE / CTH **Implementation partner - ************ End Client - ConfidentialInterview mode: Video/Virtual
Job Description – We are seeking a highly skilled AWS Data Engineer with strong expertise in SQL, Python, PySpark, Data Warehousing, and Cloud-based ETL to join our data engineering team. The ideal candidate will design, implement, and optimize large-scale data pipelines, ensuring scalability, reliability, and high performance. This role requires close collaboration with cross-functional teams and business stakeholders to deliver modern, efficient data solutions.
Key Responsibilities1. Data Pipeline Development
Build and maintain scalable ETL/ELT pipelines using Databricks on AWS.
Leverage PySpark/Spark and SQL to transform and process large, complex datasets.
Integrate data from multiple sources including S3, relational/non-relational databases, and AWS-native services.
2. Collaboration & Analysis Searching for Data Science roles that provide visa sponsorship? Connect with international employers through Data Science Jobs with Visa Sponsorship opportunities actively seeking talented professionals.
Partner with downstream teams to prepare data for dashboards, analytics, and BI tools.
Work closely with business stakeholders to understand requirements and deliver tailored, high‑quality data solutions.
3. Performance & Optimization
Optimize Databricks workloads for cost, performance, and efficient compute utilization.
Monitor and troubleshoot pipelines to ensure reliability, accuracy, and SLA adherence.
Apply query optimization, Spark tuning, and shuffle minimization best practices when handling tens of millions of rows.
4. Governance & Security
Implement and manage data governance, access control, and security policies using Unity Catalog.
Ensure compliance with organizational and regulatory data‑handling standards.
5. Deployment & DevOps
Use Databricks Asset Bundles for deployment of jobs, notebooks, and configuration across environments.
Maintain effective version control of Databricks artifacts using GitLab or similar tools.
Use CI/CD pipelines to support automated deployments and environment setups.
Technical Skills (Required)
Strong expertise in Databricks (Delta Lake, Unity Catalog, Lakehouse Architecture, Table Triggers, Workflows, Delta Live Pipelines, Databricks Runtime, etc.).
Proven ability to implement robust PySpark solutions. Explore our comprehensive directory of visa sponsorship jobs from employers worldwide who are ready to sponsor talented international professionals.
Hands‑on experience with Databricks Workflows & orchestration.
Solid knowledge of Medallion Architecture (Bronze/Silver/Gold).
Significant experience designing or rebuilding batch‑heavy data pipelines.
Strong background in query optimization, performance tuning, and Spark shuffle optimization.
Ability to handle and process tens of millions of records efficiently.
Familiarity with Genie enablement concepts (understanding required; deep experience optional).
Experience with CI/CD, environment setup, and Git-based development workflows.
Solid understanding of AWS cloud, including:
IAM
Networking fundamentals
Storage integration (S3, Glue Catalog, etc.)
Preferred Experience
Experience with Databricks Runtime configurations and advanced features.
Knowledge of streaming frameworks such as Spark Structured Streaming.
Experience developing real-time or near real-time data solutions.
Exposure to GitLab pipelines or similar CI/CD systems.
Certifications (Optional)
Databricks Certified Data Engineer Associate / Professional
AWS Data Engineer or AWS Solutions Architect certification
Показать контакты
Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →
Текст вакансии взят без изменений
Источник - Telegram канал. Название доступно после авторизации