Эта вакансия в архиве
Посмотреть похожие вакансии ↓обновлено 1 месяц назад
Lead Engineer, Inference Platform (AI)
137 000 - 270 000$
Описание вакансии
Текст:
TL;DR
Lead Engineer, Inference Platform (AI Engineering): Building the inference platform for embedding models powering semantic search, retrieval, and AI-native features in Atlas with an accent on real-time, high-scale, and low-latency inference. Focus on guiding technical direction, mentoring junior engineers, and ensuring delivery of impactful features in a multi-tenant, cloud-native environment.
Location: Must be based in Palo Alto or Seattle for our hybrid working model.
Salary: $137,000—$270,000 USD
Company
is redefining the database for the AI era with a unified database platform, Atlas, available across AWS, Google Cloud, and Microsoft Azure.
What you will do
- Partner with Search Platform and Voyage.ai AI engineers and researchers to productionize embedding models and rerankers, supporting batch and real-time inference.
- Lead projects around performance optimization, GPU utilization, autoscaling, and observability for the inference platform.
- Design and build components of a multi-tenant inference service that integrates with Atlas Vector Search.
- Contribute to platform features like model versioning, safe deployment pipelines, latency-aware routing, and model health monitoring.
- Collaborate with peers across ML, infra, and product teams to define architectural patterns and operational practices.
- Guide decisions on model serving architecture using tools like vLLM, ONNX Runtime, and container orchestration in Kubernetes.
Requirements
- 8+ years of engineering experience in backend systems, ML infrastructure, or scalable platform development, with technical leadership experience.
- Expertise in serving embedding models in production environments.
- Strong systems skills in languages like Go, Rust, C++, or Python.
- Comfortable working on cloud-native distributed systems, with a focus on latency, availability, and observability.
- Familiarity with inference runtimes and vector search systems.
- Proven ability to collaborate across disciplines and experience levels.
- Experience with high-scale SaaS infrastructure, particularly in multi-tenant environments.
Nice to have
- Prior experience working with model teams on inference-optimized architectures.
- Background in hybrid retrieval, prompt-based pipelines, or retrieval-augmented generation (RAG).
- Contributions to relevant open-source ML serving infrastructure.
Culture & Benefits
- Shape the future of AI-native developer experiences.
- Collaborate with ML experts from Voyage.ai.
- Solve hard problems in real-time inference, model serving, and semantic retrieval.
- Work in a culture that values mentorship, autonomy, and strong technical craft.
- Competitive compensation, equity, and career growth in a hands-on technical leadership role.
Похожие вакансии
1 день назад
Software Engineer (ML Infrastructure)
4 дня назад
Lead ML/AI Engineer (MLOps)
175 000 - 215 000$
2 дня назад
Staff Software Engineer, Intelligence Tooling and Enablement (AI)
165 000 - 220 000$
7 дней назад
Artificial Intelligence Lead (Engineer) (AI)
124 780 - 146 800$
1 день назад
Software Engineer, Data (AI)
240 000 - 280 000$
3 дня назад
Principal Software Engineer (AI)
139 900 - 304 200$