Member of Technical Staff Mid-Training Infra

Reflection

Not Interested
Bookmark
Report This Job

profile Job Location:

San Francisco, CA - USA

profile Monthly Salary: Not Disclosed
Posted on: 3 days ago
Vacancies: 1 Vacancy

Department:

Engineering

Job Summary

Our Mission

Reflections mission is to build open superintelligence and make it accessible to all.

Were developing open weight models for individuals agents enterprises and even nation states. Our team of AI researchers and company builders come from DeepMind OpenAI Google Brain Meta Anthropic and beyond.

About the Role

  • Design build and operate large-scale GPU infrastructure for high-throughput model inference and mid-training workloads.

  • Develop systems that power synthetic data generation and reinforcement learning pipelines at scale.

  • Build high-performance inference platforms capable of serving and evaluating models across thousands of GPUs.

  • Optimize throughput latency and GPU utilization for large language model inference and rollout workloads.

  • Build infrastructure that supports reinforcement learning pipelines including large-scale rollout generation evaluation and policy improvement loops.

  • Work closely with research teams to support distributed RL workloads and large-scale model evaluation infrastructure.

  • Improve performance of model execution through kernel-level optimization model parallelism strategies and GPU runtime improvements.

  • Develop distributed systems that enable large-scale synthetic data generation and RL-driven training workflows.

  • Diagnose and resolve performance bottlenecks across inference runtimes GPU kernels networking and distributed compute systems.

Ideal Experience

  • Experience deploying and operating large-scale GPU systems for inference or model serving.

  • Several years of hands-on experience building and running production infrastructure.

  • Strong understanding of GPU performance characteristics and optimization techniques.

  • Experience working with modern inference frameworks such as SGLang Megatron or similar high-performance LLM runtimes.

  • Familiarity with distributed reinforcement learning infrastructure or rollout generation systems.

  • Experience optimizing throughput for large-scale model execution workloads.

  • Experience working with GPU kernels or low-level performance optimization.

  • Familiarity with infrastructure used for synthetic data pipelines or RL training workflows.

  • Experience debugging performance issues across GPU networking and distributed execution layers.

What We Offer:

We believe that to build superintelligence that is truly open you need to start at the foundation. Joining Reflection means building from the ground up as part of a small talent-dense team. You will help define our future as a company and help define the frontier of open foundational models.

We want you to do the most impactful work of your career with the confidence that you and the people you care about most are supported.

  • Top-tier compensation: Salary and equity structured to recognize and retain the best talent globally.

  • Health & wellness: Comprehensive medical dental vision life and disability insurance.

  • Life & family: Fully paid parental leave for all new parents including adoptive and surrogate journeys. Financial support for family planning.

  • Benefits & balance: paid time off when you need it relocation support and more perks that optimize your time.

  • Opportunities to connect with teammates: lunch and dinner are provided daily. We have regular off-sites and team celebrations.


Required Experience:

Staff IC

Our MissionReflections mission is to build open superintelligence and make it accessible to all.Were developing open weight models for individuals agents enterprises and even nation states. Our team of AI researchers and company builders come from DeepMind OpenAI Google Brain Meta Anthropic and bey...
View more view more

Key Skills


About Company

Company Logo

Building frontier open intelligence.

View Profile View Profile