TL;DR
Member Of Technical Staff - Pre-Training (AI): Researching and building open foundational models with an accent on scaling laws, data processing, and distributed training. Focus on designing experiments to improve data efficiency, optimizing training infrastructure, and implementing state-of-the-art deep learning methods.
Location: On-site in San Francisco, New York, or London
Company
An AI startup mission-driven to build and open-source superintelligence with a high-caliber team from top research labs.
What you will do
- Build solutions across algorithms, scaling laws, and model architecture.
- Design and run scientific experiments to advance understanding of scaling LLMs.
- Implement state-of-the-art methods from deep learning literature.
- Lead independent research projects while collaborating on broader initiatives.
- Optimize training infrastructure for efficient large-scale scaling.
- Contribute across the stack from low-level optimizations to high-level model design.
Requirements
- Graduate degree (MS or PhD) in Computer Science, Machine Learning, or related field.
- Solid software engineering capabilities with experience building large-scale systems.
- Experience with large-scale ETL workflows and training data preparation.
- Deep understanding of large-scale ML, distributed training, and language models.
- Proficiency in Python and deep learning frameworks like PyTorch.
- Must be able to work on-site in San Francisco, New York, or London.
Culture & Benefits
- Top-tier salary and equity compensation.
- Comprehensive medical, dental, vision, life, and disability insurance.
- Fully paid parental leave and family planning support.
- Daily lunch and dinner provided.
- Regular team off-sites and celebrations.
- Relocation support provided.
Будьте осторожны: если работодатель просит войти в их систему, используя iCloud/Google, прислать код/пароль, запустить код/ПО, не делайте этого - это мошенники. Обязательно жмите "Пожаловаться" или пишите в поддержку. Подробнее в гайде →