42dot is seeking an AI Model Optimization and Tool Development Engineer (NPU) to focus on optimizing the autonomous driving stack and ondevice large language models (LLMs). This role involves developing AI model optimization techniques for NPUs and building toolchains to ensure efficient . The engineer will be responsible for optimizing deep learning models for hardware accelerators designing and developing toolchains that enhance performance and supporting the advancement of AI technologies such as autonomous driving and LLMs through hardwareaware optimizations. This position plays a crucial role in bridging AI models with hardware accelerators ensuring seamless integration and optimal efficiency.
Responsibilities
AI Model Porting and Optimization
Port AI models for LLM and autonomous driving stacks to NPU hardware and optimize their performance. Improve inference speed by utilizing techniques such as model compression (quantization pruning etc. operator fusion and memory optimization.
Toolchain Development & Compiler Engineering
Design and implement toolchains for porting AI models to NPUs. Integrate with deep learning frameworks such as TensorFlow and PyTorch to provide an efficient workflow. Develop tools for NPUspecific code generation profiling and debugging.
Optimization of Autonomous Driving and LLM Stacks
Optimize AI modules required for autonomous driving (e.g. object detection path planning) to ensure compatibility and realtime performance. Enhance memory efficiency and speed through LLM inference optimization. Apply model parallelization and distributed techniques in multimodal AI stacks.
Performance Analysis and Improvement
Analyze AI model runtime performance and identify bottlenecks. Implement techniques to maximize hardware utilization.
Research and Adoption of New Technologies
Study the latest advancements in AI model optimization and NPUrelated technologies. Experiment with and adopt new techniques to maximize NPU performance.
Qualifications
Bachelors or Masters degree in Computer Science AI or a related field
At least 3 years of experience in AI model optimization and hardware acceleration
Familiarity with compiler technologies such as LLVM and MLIR
Experience optimizing AI models using NPUs GPUs or ASICs
Proficiency in deep learning frameworks and model conversion tools such as TensorFlow Lite ONNX and PyTorch
Expertise in model compression and optimization techniques including quantization pruning and lazy evaluation
Proficiency in programming languages such as CUDA C and Python with experience in writing hardwareaccelerated code
Strong understanding of memory management and parallel computing techniques
Preferred Qualifications
Experience with autonomous driving stacks including SLAM path planning and object recognition
Optimization experience for ondevice AI/LLM applications
Experience in AI optimization for embedded systems
Contributions to opensource AI optimization projects
Interview Process
Application Review Coding Test First Interview 1 hour) Second Interview 3 hours) Final Selection
The interview processmay vary depending on the position and is subject to change based on the schedule and circumstances.
Applicants will be individually notified of the interview schedule and results via the email provided in their application.
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.