Our team is responsible for the AWS Neuron software stack which powers Generative AI and other advanced ML workloads on AWSs custom-built ML accelerators Inferentia and Trainium. These accelerators deliver best-in-class performance and cost-efficiency for ML inference and training in the cloud.
Were building a new core group of engineers in TLV (Tel Aviv) to drive innovation in ML systems performance and software. As a Machine Learning Performance Engineer youll help shape the direction of the team from the ground up and work on:
Optimizing system performance across the entire ML software stack
Analyzing high-performance ML workloads running on Annapurna hardware
Developing high-performance kernels for critical ML operations
Enhancing the Neuron SDK to improve developer experience and system capabilities
Collaborating across Compiler Frameworks and Hardware teams to maximize end-to-end performance
As part of the Performance Engineering Team youll contribute to projects involving instruction scheduling memory management parallelism kernel optimization and compiler enhancements to maximize end-to-end performance.
This is a unique opportunity to be at the intersection of ML and systems within AWS helping to build the future of AI infrastructure right here in Tel Aviv.
Key job responsibilities
basic qualifications
* B.S. or M.S. in computer science or related field
* Proficiency with 1 or more of the following programming languages: Python (preferred) C
* Experience with TensorFlow PyTorch and/or JAX
* 3 years of non-internship professional software development experience
* 3 years of performance optimization experience in LLM Vision or other deep-learning models
Preferred qualifications
* M.S. in computer science or related field
* Experience with developing algorithms for simulation tools
* Experience with VLLM or other inference serving infrastructures
* Experience developing compiler optimization kernel writing or hardware-software co-design
About the team
*Diverse Experiences
Amazon values diverse experiences. Even if you do not meet all of the preferred qualifications and skills listed in the job description we encourage candidates to apply. If your career is just starting hasnt followed a traditional path or includes alternative experiences dont let it stop you from applying.
*Why AWS
Amazon Web Services (AWS) is the worlds most comprehensive and broadly adopted cloud platform. We pioneered cloud computing and never stopped innovating thats why customers from the most successful startups to Global 500 companies trust our robust suite of products and services to power their businesses.
*Work/Life Balance
We value work-life harmony. Achieving success at work should never come at the expense of sacrifices at home which is why flexible work hours and arrangements are part of our culture. When we feel supported in the workplace and at home theres nothing we cant achieve in the cloud.
*Inclusive Team Culture
AWS values curiosity and connection. Our employee-led and company-sponsored affinity groups promote inclusion and empower our people to take pride in what makes us unique. Our inclusion events foster stronger more collaborative teams. Our continual innovation is fueled by the bold ideas fresh perspectives and passionate voices our teams bring to everything we do.
*Mentorship and Career Growth
Were continuously raising our performance bar as we strive to become Earths Best Employer. Thats why youll find endless knowledge-sharing mentorship and other career-advancing resources here to help you develop into a better-rounded professional.
- B.S. or M.S. in computer science or related field
- Proficiency with 1 or more of the following programming languages: C (preferred) Python
- 3 years of non-internship professional software development experience
- 3 years of experience developing compiler optimization graph-theory hardware bring-up FPGA placement and routing algorithms or hardware resource management
- M.S. or Ph.D. in computer science or related field
- Strong knowledge in one or more of the areas of: compiler design instruction scheduling memory allocation data transfer optimization graph partitioning parallel programing code generation Instruction Set Architectures new hardware bring-up and hardware-software co-design
- Experience with LLVM and/or MLIR
- Experience with developing algorithms for simulation tools
- Experience is TensorFlow PyTorch and/or JAX
- Experience in LLM Vision or other deep-learning models
Our inclusive culture empowers Amazonians to deliver the best results for our customers. If you have a disability and need a workplace accommodation or adjustment during the application and hiring process including support for the interview or onboarding process please visit
for more information. If the country/region youre applying in isnt listed please contact your Recruiting Partner.