TL;DR
Technical Program Manager, Infrastructure (AI): Driving complex programs to scale hirify.global's AI infrastructure, covering developer productivity, tooling, and reliability for massive training clusters and production systems. Focus on coordinating strategic initiatives across research, engineering, and product teams, solving novel scaling challenges, and ensuring security and reliability.
Location: Hybrid in San Francisco, CA, New York City, NY, or Seattle, WA. Requires in-office presence at least 25% of the time. Visa sponsorship is available.
Salary: $290,000–$365,000 USD
Company
hirify.global is a public benefit corporation focused on creating reliable, interpretable, and steerable AI systems.
What you will do
- Drive programs to improve developer environments, CI/CD infrastructure, and release processes while maintaining high security standards.
- Coordinate large-scale migrations and platform modernization efforts across engineering teams.
- Partner with teams to measure and improve developer productivity, identifying bottlenecks and driving systematic improvements.
- Lead initiatives to integrate AI tools into development workflows for AI-assisted research and engineering.
- Drive programs to establish and achieve reliability targets across training infrastructure and production services.
- Serve as the critical bridge between infrastructure teams, research, and product, translating technical complexities for various audiences.
Requirements
- 5+ years of technical program management experience, with a track record of successfully delivering complex infrastructure programs in ML/AI systems or large-scale distributed systems.
- Deep technical understanding of infrastructure systems, sufficient to engage substantively with engineers and identify technical risks.
- Excel at creating structure and processes in ambiguous environments, bringing clarity to complex cross-team initiatives.
- Strong stakeholder management skills and ability to build trust with technical and non-technical partners.
- Comfortable navigating competing priorities and using data to drive technical decisions.
- Passion for AI infrastructure and understanding the unique challenges of building and operating systems at frontier scale.
Nice to have
- Experience with Kubernetes, cloud platforms (AWS, GCP, Azure), and ML infrastructure (GPU/TPU/Trainium clusters).
- Background working with research teams and translating their needs into concrete technical requirements.
- Familiarity with observability tooling and practices.
Culture & Benefits
- Work as a single cohesive team on a few large-scale research efforts.
- Value impact, advancing long-term goals of steerable, trustworthy AI.
- Extremely collaborative group with frequent research discussions.
- Competitive compensation and benefits, including optional equity donation matching.
- Generous vacation and parental leave.
- Flexible working hours and a lovely office space for collaboration.
Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →