Token-as-a-Service Technical Program Manager (AI)
Мэтч & Сопровод
Для мэтча с этой вакансией нужен Plus
Описание вакансии
TL;DR
Token-as-a-Service Technical Program Manager (AI): Leading the delivery of external compute capacity to serve model workloads with an accent on infrastructure execution and systems readiness. Focus on transforming third-party infrastructure into usable tokens at scale, managing cross-functional programs, and optimizing production throughput.
Location: Based in San Francisco, CA (Hybrid: 3 days per week in office). Relocation assistance is available.
Salary: $342K – $555K
Company
is an AI research and deployment company dedicated to ensuring that general-purpose artificial intelligence benefits all of humanity.
What you will do
- Lead end-to-end delivery programs that convert external infrastructure capacity into production-ready token supply.
- Own readiness across compute, storage, networking, security, and operational dependencies for third-party environments.
- Build integrated plans across internal engineering teams and external partners with clear milestones, owners, and risks.
- Drive launch execution for new partner regions, clusters, and capacity expansions.
- Create operating mechanisms to measure deployed capacity versus usable token output.
- Identify and resolve bottlenecks preventing token generation, such as network constraints or hardware readiness.
Requirements
- 8+ years of Technical Program Management, Engineering Program Management, or Infrastructure Delivery experience.
- Experience leading large-scale technical programs involving cloud, data centers, networking, hardware, or distributed systems.
- Strong understanding of compute infrastructure, clusters, networking, storage, and production systems.
- Proven ability to drive cross-functional execution across engineering, operations, finance, and external vendors.
- Experience managing executive stakeholders and communicating complex tradeoffs clearly.
- Must be based in or be able to relocate to San Francisco, CA.
Nice to have
- Experience with GPU clusters, AI infrastructure, or large-scale model serving environments.
- Familiarity with token economics, inference capacity planning, or workload scheduling.
- Experience scaling global infrastructure through third-party providers.
- Background in systems engineering, networking, or hardware deployment programs.
Culture & Benefits
- Relocation assistance is provided for eligible candidates.
- High-visibility role with direct impact on the ability to scale model training and inference globally.
- Work at the forefront of AI research and deployment within a fast-scaling environment.
- Commitment to diversity, equity, and inclusion as an equal opportunity employer.
Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →