

evaluate and improve state-of-the-art performance techniques in production Large Language Model deployments, and he
What you’ll be doing:
Develop innovative GPU and system architectures to extend the state of the art in AI Inference performance and efficiency
Model, analyze and prototype key deep learning algorithms and applications
Understand and analyze the interplay of hardware and software architectures on future algorithms and applications
Write efficient software for AI Inference, including CUDA kernels, framework level code, and application level code
Collaborate across the company to guide the direction of AI, working with software, research and product teams
What we need to see:
A MS or PhD in a relevant discipline (CS, EE, Math) or equivalent experience, with 5+ years or relevant experience
Strong mathematical foundation in machine learning and deep learning
Expert programming skills in C, C++, and Python
Familiarity with GPU computing (CUDA or similar) and HPC (MPI, OpenMP)
Strong knowledge and coursework in computer architecture
Ways to stand out from the crowd:
Background with systems-level performance modeling, profiling, and analysis
Experience in characterizing and modeling system-level performance, executing comparison studies, and documenting and publishing results
Experience in optimizing AI Inference workloads with CUDA kernel development
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך