Employer Active
Job Alert
You will be updated with latest job alerts via emailJob Alert
You will be updated with latest job alerts via emailAre you ready to join a game-changing open-source AI platform that harnesses the power of hybrid cloud to drive innovation
The Red Hat OpenShift AI (RHOAI) team is looking for a Principal Software Engineer with Kubernetes and MLOps (Machine Learning) experience to join our rapidly growing engineering team. Our focus is to create a platform partner ecosystem and community by which enterprise customers can solve problems to accelerate business success using AI. This is a very exciting opportunity to build and impact the next generation of hybrid cloud MLOps platforms contribute to the development of the RHOAI product participate in open-source communities and be at the forefront of the exciting evolution of AI. Youll join an ecosystem that fosters continuous learning career growth and professional development.
In this role youll be contributing as a model serving and monitoring subject matter expert for the model serving features of the open-source Open Data Hub project by actively participating in KServe TrustyAI Kubeflow HuggingFacevLLM and several other open-source communities. You will work as part of an evolving development team to rapidly design secure build test and release model serving trustworthy AI and model registry capabilities. The role is primarily an individual contributor who will be a key notable contributor to MLOps upstream communities and collaborate closely with the internal cross-functional development teams.
Lead the team strategy and implementation for Kubernetes-native components in Model Serving including Custom Resources Controllers and Operators.
Be an influencer and leader in MLOps-related open source communities to help build an active MLOps open source ecosystem for Open Data Hub and OpenShift AI
Act as an MLOps SME within Red Hat by supporting customer-facing discussions presenting at technical conferences and evangelizing OpenShift AI within the internal community of practices
Architect and design new features for open-source MLOps communities such as KubeFlow and KServe
Provide technical vision and leadership on critical and high-impact projects
Mentor influence and coach a team of distributed engineers
Ensure non-functional requirements including security resiliency and maintainability are met
Write unit and integration tests and work with quality engineers to ensure product quality
Use CI/CD best practices to deliver solutions as productization efforts into RHOAI
Contribute to a culture of continuous improvement by sharing recommendations and technical knowledge with team members
Collaborate with product management other engineering and cross-functional teams to analyze and clarify business requirements
Communicate effectively to stakeholders and team members to ensure proper visibility of development efforts
Give thoughtful and prompt code reviews
Represent RHOAI in external engagements including industry events customer meetings and open-source communities
Proactively utilize AI-assisted development tools (e.g. GitHub Copilot Cursor Claude Code) for code generation auto-completion and intelligent suggestions to accelerate development cycles and enhance code quality.
Explore and experiment with emerging AI technologies relevant to software development proactively identifying opportunities to incorporate new AI capabilities into existing workflows and tooling.
Proven expertise with Kubernetes API development and testing (CRs Operators Controllers) including reconciliation logic.
Strong background with model serving (like KServe vLLM) and distributed inference strategies for LLMs (tensor pipeline data parallelism).
Deep understanding of GPU optimization autoscaling (KEDA/Knative) and low-latency networking (e.g. NVLink P2P GPU).
Experience architecting resilient secure and observable systems for model serving including metrics and tracing.
Advanced skills in Go and Python; ability to design APIs for high-performance inference and streaming.
Excellent system troubleshooting skills in cloud environments and the ability to innovate in fast-paced environments.
Strong communication and leadership skills to mentor teams and represent projects in open-source communities.
Autonomous work ethic and passion for staying at the forefront of AI and open source.
The following will be considered a plus:
An existing contributor in one or more MLOps open source projects such as KubeFlow KServe RayServe and vLLM is a huge plus
Familiarity with optimization techniques for LLMs (quantization TensorRT Hugging Face Accelerate).
Knowledge of end-to-end MLOps workflows including model registry explainability and drift detection.
Bachelors degree in statistics mathematics computer science operations research or a related quantitative field or equivalent expertise; Masters or PhD is a big plus
Understanding of how Open Source and Free Software communities work
Experience with development for public cloud services (AWS GCE Azure)
Experience in engineering consulting or another field related to model serving and monitoring model registry explainable AI deep neural networks in a customer environment or supporting a data science team
Highly experienced in OpenShift
Familiarity with popular Python machine learning libraries such as PyTorch Tensorflow and Hugging Face
#AI-HIRING
#LI-MD2
The salary range for this position is $148540.00 - $245050.00. Actual offer will be based on your qualifications.Pay Transparency
Red Hat determines compensation based on several factors including but not limited to job location experience applicable skills and training external market value and internal pay equity. Annual salary is one component of Red Hats compensation package. This position may also be eligible for bonus commission and/or equity. For positions with Remote-US locations the actual salary range for the position may differ based on location but will be commensurate with job duties and relevant work experience.
About Red Hat
Red Hat is the worlds leading provider of enterprise open source software solutions using a community-powered approach to deliver high-performing Linux cloud container and Kubernetes technologies. Spread across 40 countries our associates work flexibly across work environments from in-office to office-flex to fully remote depending on the requirements of their role. Red Hatters are encouraged to bring their best ideas no matter their title or tenure. Were a leader in open source because of our open and inclusive environment. We hire creative passionate people ready to contribute their ideas help solve complex problems and make an impact.
Benefits
Comprehensive medical dental and vision coverage
Flexible Spending Account - healthcare and dependent care
Health Savings Account - high deductible medical plan
Retirement 401(k) with employer match
Paid time off and holidays
Paid parental leave plans for all new parents
Leave benefits including disability paid family medical leave and paid military leave
Additional benefits including employee stock purchase plan family planning reimbursement tuition reimbursement transportation expense account employee assistance program and more!
Note: These benefits are only applicable to full time permanent associates at Red Hat located in the United States.
Inclusion at Red Hat
Red Hats culture is built on the open source principles of transparency collaboration and inclusion where the best ideas can come from anywhere and anyone. When this is realized it empowers people from different backgrounds perspectives and experiences to come together to share ideas challenge the status quo and drive innovation. Our aspiration is that everyone experiences this culture with equal opportunity and access and that all voices are not only heard but also celebrated. We hope you will join our celebration and we welcome and encourage applicants from all the beautiful dimensions that compose our global village.
Equal Opportunity Policy (EEO)
Red Hat is proud to be an equal opportunity workplace and an affirmative action employer. We review applications for employment without regard to their race color religion sex sexual orientation gender identity national origin ancestry citizenship age veteran status genetic information physical or mental disability medical condition marital status or any other basis prohibited by law.
Required Experience:
Staff IC
Full-Time