Optimize inference throughput for novel model architectures, contribute to frameworks, and train trillion-parameter models on large GPU clusters.
Key Highlights
Technical Skills Required
Benefits & Perks
Job Description
Research Engineer - San Francisco, CA
A company building frontier-scale AI models that automate software engineering and AI research, combining ultra-long context, domain-specific RL, and massive compute infrastructure are looking for a Research Engineer to join their team.
What Will I Be Doing:
- Optimize inference throughput for novel model architectures
- Contribute to frameworks used for our research and production workloads
- Train trillion-parameter models on large GPU clusters
- Curate post-training datasets to improve targeted capabilities
- Build out internet-scale data pipelines and crawlers
- Design, prototype and optimize new model architectures
- Contribute to research across long-context, inference-time compute, RL, and more
What we’re looking for:
- Strong general software engineering skills
- Thorough knowledge of the deep learning literature
- Experience with pre- and post-training of LLMs
- Ability to come up with and evaluate research ideas
- Experience working with large distributed systems
- Ability to handle large ETL workloads
What’s In It for Me:
- Salary of $225K–$550K dependent on experience + significant equity
- Great benefits inc. 401(k) with 6% company match, comprehensive health, unlimited PTO
- Visa sponsorship and SF relocation stipend available
- Well-funded ($465M+) with backing from top investors
Apply now for immediate consideration!