Distributed Training Engineer
Listed on 2025-12-11
-
IT/Tech
AI Engineer, Machine Learning/ ML Engineer
Join to apply for the Distributed Training Engineer role at Periodic Labs
1 month ago Be among the first 25 applicants
About Periodic LabsWe are an AI + physical sciences lab building state of the art models to make novel scientific discoveries. We are well funded and growing rapidly. Team members are owners who identity and solve problems without boundaries or bureaucracy. We eagerly learn new tools and new science to push forward our mission.
AboutThe Role
You will optimize, operate and develop large-scale distributed LLM training systems that power AI scientific research. You will work closely with researchers to bring up, debug, and maintain mid-training and reinforcement learning workflows. You will build tools and directly support frontier-scale experiments to make Periodic Labs the world’s best AI + science lab for physicists, computational materials scientists, AI researchers, and engineers.
You will contribute open-source large scale LLM training frameworks.
- Training on clusters with ≥5,000 GPUs
- 5D parallel LLM training
- Distributed training frameworks such as Megatron-LM, FSDP, Deep Speed, Torch Titan
- Optimizing training throughput for large scale Mixture-of-Expert models
- Entry level
- Full-time
- Engineering and Information Technology
- Technology, Information and Internet
Referrals increase your chances of interviewing at Periodic Labs by 2x
#J-18808-Ljbffr(If this job is in fact in your jurisdiction, then you may be using a Proxy or VPN to access this site, and to progress further, you should change your connectivity to another mobile device or PC).