The AWS Neuron Compiler team is actively seeking skilled compiler engineers to join our efforts in developing a state-of-the-art deep learning compiler stack. This stack is designed to optimize application models across diverse domains including Large Language and Vision originating from leading frameworks such as PyTorch TensorFlow and JAX. Your role will involve working closely with our custom-built Machine Learning accelerators including Inferentia/Trainium which represent the forefront of AWS innovation for advanced ML capabilities powering solutions like Generative AI.
Key job responsibilities
As a ML Compiler Engineer II on the Neuron Compiler Automated Reasoning Group you will develop and maintain tooling for best-in-class technology for raising the bar of the Neuron Compilers accuracy and reliability. You will help lead the efforts building fuzzers and specification synthesis tooling for our LLVM-based compiler. You will work in a team with a science focus and strive to push what we do to the edge of what is known to best deliver our customers.
Strong software development skills using C/Python are critical to this role.
A science background in compiler development is strongly preferred. A background in Machine Learning and AI accelerators is preferred but not required.
In order to be considered for this role candidates must be currently located or willing to relocate to Seattle (Preferred) Cupertino Austin or Toronto.
- 3 years of leading design or architecture (design patterns reliability and scaling) of new and existing systems experience - 2 years of experience in developing compiler features and optimizations - Proficiency in C and Python programming applied to compiler or verification projects - Familiarity with LLVM including knowledge of abstract interpretation and polyhedral domains - Demonstrated scientific approach to software engineering problems
- Masters degree or PhD in computer science or equivalent - Experience with deep learning frameworks like TensorFlow or PyTorch - Understanding of large language model (LLM) training processes - Knowledge of CUDA programming for GPU acceleration
Amazon is an equal opportunity employer and does not discriminate on the basis of protected veteran status disability or other legally protected status.
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.