ABOUT THE ROLE
You would be working in our pre-training team focused on building out our distributed training and inference of Large Language Models (LLMs). This is a hands-on role that focuses on software reliability and fault tolerance. You will work on cross-platform checkpointing, NCCL recovery, and hardware fault detection. You will make high-level tools. You will not be afraid of debugging Linux kernel modules. You will have access to thousands of GPUs to test changes.
Strong engineering skills are a prerequisite. We assume good knowledge of Torch, NVIDIA GPU architecture, reliability concepts, distributed systems, and best coding practices. A basic understanding of LLM training and inference principles is required. We look for fast learners who are prepared for a steep learning curve and are not afraid to step out of their comfort zone.
YOUR MISSION
To help train the best foundational models for source code generation in the world
RESPONSIBILITIES
SKILLS & EXPERIENCE
Understanding of Large Language Models (LLM)
Basic knowledge of Transformers
Knowledge of deep learning fundamentals
Strong engineering background
Programming experience
Linux API, Linux kernel
Strong algorithmic skills
Python with numpy, PyTorch, or Jax
C/C++
NCCL
Use modern tools and are always looking to improve
Strong critical thinking and ability to question code quality policies when applicable
Distributed systems
Reliability
Observability
Fault-tolerance
K8s stack
Building the world's most capable AI for software development & the applications to unlock the potential of developers.
Salary
$240,000 - $400,000
Location
Remote
Total raised
$626.0M
Last stage
Series B
Investors
No applications, no recruiter spam. Just the intro.
A few questions to make sure this role is the right shape for you. Two minutes.
I write the intro, send it to the founder, and handle the back-and-forth.
If they’re a yes, I book the chat. You show up — that’s the whole job-hunt.