LLM Inference Engineer

4 Hours ago • All levels
Research Development

Job Description

Periodic Labs is an AI + physical sciences lab focused on state-of-the-art models for scientific discoveries, well-funded and rapidly growing. The LLM Inference Engineer will integrate, optimize, and operate large-scale inference systems, building and maintaining high-performance serving infrastructure for low-latency, high-throughput access to large language models across thousands of GPUs. This role involves close collaboration with researchers to integrate cutting-edge inference into reinforcement learning workloads, building tools, and supporting frontier-scale experiments to advance AI + science research, including contributions to open-source LLM inference software.
Must Have:
  • Optimizing inference for the largest open-source model
  • High-performance model serving frameworks such as TensorRT-LLM, vLLM, SGLang
  • Distributed inference techniques (tensor/expert/pipeline parallelism, speculative decoding, KV cache management)
  • Optimizing GPU utilization and latency for reinforcement learning

Add these skills to join the top 1% applicants for this job

game-texts
model-serving
reinforcement-learning

About Periodic Labs

We are an AI + physical sciences lab building state of the art models to make novel scientific discoveries. We are well funded and growing rapidly. Team members are owners who identity and solve problems without boundaries or bureaucracy. We eagerly learn new tools and new science to push forward our mission.

About the role

You will integrate, optimize, and operate large-scale inference systems to power AI scientific research. You will build and maintain high-performance serving infrastructure that delivers low-latency, high-throughput access to large language models across thousands of GPUs. You will work closely with researchers and engineers to integrate cutting-edge inference into large-scale reinforcement learning workloads. You will build tools and directly support frontier-scale experiments to make Periodic Labs the world’s best AI + science lab. You will make contributions to open-source LLM inference software.

You might thrive in this role if you have experience with:

  • Optimizing inference for the largest open-source model
  • High-performance model serving frameworks such as TensorRT-LLM, vLLM, SGLang
  • Distributed inference techniques (tensor/expert/pipeline parallelism, speculative decoding, KV cache management)
  • Optimizing GPU utilization and latency for reinforcement learning

Set alerts for more jobs like LLM Inference Engineer
Set alerts for new jobs by Periodic Labs
Set alerts for new Research Development jobs in United States
Set alerts for new jobs in United States
Set alerts for Research Development (Remote) jobs

Contact Us
hello@outscal.com
Made in INDIA 💛💙