Senior/Staff Engineer for Distributed ML Model Training
We're looking for Senior/Staff engineers with 5+ years of experience in distributed systems and ML large-scale training. You'll be implementing a novel substrate for training distributed ML models that work under consumer grade internet connection. Design and implement large-scale distributed training systems optimized for heterogeneous hardware operating under low-bandwidth, high-latency conditions.
Key Highlights
Key Responsibilities
Technical Skills Required
Benefits & Perks
Job Description
Overview
Pluralis Research carries out foundational research on Protocol Learning: multi-participant training of foundation models where no single participant has, or can ever obtain, a full copy of the model. The purpose of Protocol Learning is to facilitate the creation of community-trained and community-owned frontier models with self-sustaining economics.
We're looking for Senior/Staff engineers with 5+ years of experience in distributed systems and ML large-scale training. You'll be implementing a novel substrate for training distributed ML models that work under consumer grade internet connection.
Responsibilities
Distributed Training Architecture & Optimization
- Design and implement large-scale distributed training systems optimized for heterogeneous hardware operating under low-bandwidth, high-latency conditions.
- Develop and optimize model-parallel training strategies (data, tensor, pipeline parallelism) with custom sharding techniques that minimize communication overhead.
- Optimize GPU utilization, memory efficiency, and compute performance across distributed nodes.
- Implement robust checkpointing, state synchronization, and recovery mechanisms for long-running, fault-prone training jobs.
- Build monitoring and metrics systems to track training progress, model quality, and system bottlenecks.
Searching for Machine Learning & AI roles that provide visa sponsorship? Connect with international employers through Machine Learning & AI Jobs with Visa Sponsorship opportunities actively seeking talented professionals.
- Architect resilient training systems where nodes can fail, networks can partition, and participants can dynamically join or leave.
- Design and optimize peer-to-peer topologies for decentralized coordination across non-co-located nodes.
- Implement NAT traversal, peer discovery, dynamic routing, and connection lifecycle management.
- Profile and optimize communication patterns to reduce latency and bandwidth overhead in multi-participant environments.
Explore our comprehensive directory of visa sponsorship jobs from employers worldwide who are ready to sponsor talented international professionals.
- Strong experience building and operating distributed systems in production.
- Hands-on expertise with distributed training frameworks (FSDP, DeepSpeed, Megatron, or similar).
- Deep understanding of model parallelism (data, tensor, pipeline parallelism).
- Expert-level Python with production experience (concurrency, error handling, retry logic, clean architecture).
- Strong networking fundamentals: P2P systems, gRPC, routing, NAT traversal, distributed coordination.
- Experience optimizing GPU workloads, memory management, and large-scale compute efficiency.
Interested in opportunities specifically in Australia? Discover our dedicated Visa Sponsorship Jobs in Australia page featuring roles from top employers in this location.
- Equity-heavy compensation with meaningful ownership in a mission-driven company
- Competitive base salary for senior engineering roles in Australia
- Visa sponsorship available for exceptional candidates
- Remote-first with optional access to our Melbourne hub
- World-class team — team mates were previously at at Google, Amazon, Microsoft, and leading startups
Similar Jobs
Explore other opportunities that match your interests
synoptix ai
va talks consulting pvt. ltd.