What you'll be doing
In this role, develop compiler optimization algorithms for deep learning workloads. You will optimize inference and training performance for the JAX framework and the OpenXLA compiler on NVIDIA GPUs at scale. You’ll collaborate with our partners in deep learning framework teams and our hardware architecture teams to accelerate the next generation of deep learning software. The scope of these efforts include:
Crafting and implementing compiler optimization techniques for deep learning network graphs.
Designing novel graph partitioning and tensor sharding techniques for distributed training and inference.
Performance tuning and analysis.
Code-generation for NVIDIA GPU backends using open-source compilers such as MLIR, LLVM and OpenAI Triton.
Designing user facing features in JAX and related libraries and other general software engineering work.
Working closely with GPU hardware engineering teams to design AI compiler software features for next-generation GPUs.
What we need to see
Bachelors, Masters or Ph.D. in Computer Science, Computer Engineering, related field (or equivalent experience).
4+ years of relevant work or research experience in performance analysis and compiler optimizations.
Ability to work independently, define project goals and scope, and lead your own development effort adopting clean software engineering and testing practices.
Excellent C/C++ programming and software design skills, including debugging, performance analysis, and test design.
Strong foundation in architecture of CPU, GPUs or other high performance hardware accelerators. Knowledge of high-performance computing and distributed programming.
CUDA or OpenCL programming experience is desired but not required.
Experience with the following technologies is a huge plus: XLA, TVM, MLIR, LLVM, OpenAI Triton, deep learning models and algorithms, and deep learning framework design.
Strong interpersonal skills are required along with the ability to work in a dynamic product-oriented team. A history of mentoring junior engineers and interns is a bonus.
Ways to stand out from the crowd
Experience working deep learning frameworks such as JAX, PyTorch or TensorFlow.
Extensive experience with CUDA or with GPUs in general.
Experience with open-source compilers such as XLA, LLVM, MLIR or TVM.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך