Job Opportunity
We would love to meet you if you :
- Philosophy : You are your own worst critic. You have a high bar for quality and don't rest until the job is done rightno settling for 90%. We want someone who ships fast, with high agency, and who doesn't just voice problems but actively jumps in to fix them.
- Experience : You have deep expertise in Python and PyTorch, with a strong foundation in low-level operating systems concepts including multi-threading, memory management, networking, storage, performance, and scale. You're experienced with modern inference systems like TGI, vLLM, TensorRT-LLM, and Optimum, and comfortable creating custom tooling for testing and optimization.
- Approach : You combine technical expertise with practical problem-solving. You're methodical in debugging complex systems and can rapidly prototype and validate solutions.
The core work will include :
Architecting and implementing robust, scalable inference systems for serving state-of-the-art AI modelsOptimizing model serving infrastructure for high throughput and low latency at scaleDeveloping and integrating advanced inference optimization techniquesWorking closely with our research team to bring cutting-edge capabilities into productionBuilding developer tools and infrastructure to support rapid experimentation and deployment.Bonus points if you :
Have experience with low-level systems programming (CUDA, Triton) and compiler optimizationAre passionate about open-source contributions and staying current with ML infrastructure developmentsBring practical experience with high-performance computing and distributed systemsHave worked in early-stage environments where you helped shape technical directionAre energized by solving complex technical challenges in a collaborative environmentThis is an in person role at our office in SF. We're an early stage company which means that the role requires working hard and moving quickly. Please only apply if that excites you.