

What you’ll be doing:
Working with tech giants to develop and demonstrate solutions based on NVIDIA’s groundbreaking software and hardware technologies.
Partnering with Sales Account Managers and Developer Relations Managers to identify and secure business opportunities for NVIDIA products and solutions.
Serving as the main technical point of contact for customers engaged in the development of intricate AI infrastructure, while also offering support in understanding performance aspects related to tasks like large scale LLM training and inference.
Conducting regular technical customer meetings for project/product details, feature discussions, introductions to new technologies, performance advice, and debugging sessions.
Collaborating with customers to build Proof of Concepts (PoCs) for solutions to address critical business needs and support cloud service integration for NVIDIA technology on hyperscalers.
Analyzing and developing solutions for customer performance issues for both AI and systems performance.
What we need to see:
BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or other Engineering fields or equivalent experience.
4+ years of engineering(performance/system/solution)experience.
Hands-on experience building performance benchmarks for data center systems, including large scale AI training and inference.
Understanding of systems architecture including AI accelerators and networking as it relates to the performance of an overall application.
Effective engineering program management with the capability of balancing multiple tasks.
Ability to communicate ideas clearly through documents, presentations, and in external customer-facing environments.
Ways to stand out from the crowd:
Hands-on experience with Deep Learning frameworks (PyTorch, JAX, etc.), compilers (Triton, XLA, etc.), and NVIDIA libraries (TRTLLM, TensorRT, Nemo, NCCL, RAPIDS, etc.).
Familiarity with deep learning architectures and the latest LLM developments.
Background with NVIDIA hardware and software, performance tuning, and error diagnostics.
Hands-on experience with GPU systems in general including but not limited to performance testing, performance tuning, and benchmarking.
Experience deploying solutions in cloud environments including AWS, GCP, Azure, or OCI as well as knowledge of DevOps/MLOps technologies such as Docker/containers, Kubernetes, data center deployments, etc. Command line proficiency.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What You'll Be Doing:
Working as a key member of our cloud solutions team, you will be the go-to technical expert on NVIDIA's products, helping our clients architect and optimize GPU solutions for AI services.
Collaborating directly with engineering teams to secure design wins, address challenges, usher projects into production, and offer support through the project's lifecycle.
Acting as a trusted advisor to our clients, while developing reference architectures and best practices for running Microsoft AI workloads on NVIDIA infrastructure.
What We Need To See:
4+ years of experience in cloud computing and/or large-scale AI systems.
A BS in EE, CS, Math, or Physics, or equivalent experience.
A proven understanding of cloud computing and large-scale computing systems.
Proficiency in Python, C, or C++ and experience with AI frameworks like Pytorch or TensorFlow.
Passion for machine learning and AI, and the drive to continually learn and apply new technologies.
Excellent interpersonal skills, including the ability to explain complex technical topics to non-experts.
Ways To Stand Out From The Crowd:
Recent projects or contributions (for example, on GitHub) related to large language models and transformer architectures.
Knowledge of Azure cloud and AzureML services.
Experience with CUDA programming and optimization.
Familiarity with NVIDIA networking technologies such as Infiniband.
Proficiency in Linux, Windows Subsystem for Linux, and Windows.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What you'll be doing:
Research, design and implement novel methods for efficient deep learning.
Publish original research.
Collaborate with other team members and teams.
Mentor interns.
Speak at conferences and events.
Work with product groups to transfer technology.
Collaborate with external researchers.
What we need to see:
Completing or recently completed a Ph.D. in Computer Science/Engineering, Electrical Engineering, etc., or have equivalent research experience.
Excellent knowledge of theory and practice of computer vision methods, as well as deep learning.
Background in pruning, quantization, NAS, efficient backbones, and so on, is a plus.
Experience with large language models and large vision-language models is required.
Excellent programming skills in Python and PyTorch; C++ and parallel programming (e.g., CUDA) is a plus.
Hands-on experience with large-scale model training including data preparation and model parallelization (tensor and pipeline) is required.
Outstanding research track record.
Excellent communications skills.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What you’ll be doing:
Collaborating with business development in guiding the customer through the solution adoption process for our Metropolis, Isaac and IGX AI SW platforms, GPU Computing and IGX/Jetson, being responsible for the technical relationship and assisting customers in building creative solutions based on NVIDIA
Be an industry leader with vision on integrating NVIDIA technology into intelligent machines’ architectures
You will engage with customers to develop a keen understanding of their goals, vision and plans, as well as technical needs – and help to define and deliver high-value solutions that meet these needs
Train customers on the adoption of our AI platforms, develop and optimize proof of concepts using the Nvidia robotics and Metropolis platforms as well as the Jetson/IGX SDKs
Establish positive relationships and communication channels with internal teams
What we need to see:
BS or MS in Electrical Engineering or Computer Science or equivalent experience
8+ years of work-related experience in a high-tech electronics industry in a similar role as a systems or solution architect
AI practitioner experience
C, C++, and Python coding
Strong time-management and organization skills for coordinating multiple initiatives, priorities, and implementations of new technology and products into very complex projects
Ways to stand out from the crowd:
NVIDIA GPU development experience
Experience with Omniverse, ISAAC and Metropolis
Experience with generative AI on Jetson or IGX, RIVA, VSS
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What you’ll be doing:
Pre-silicon Power Estimation: Model and estimate CPU power at C-model, RTL, and netlist stages using industry-standard tools.
Power Optimization: Identify inefficiencies and drive design improvements in collaboration with architects, RTL designers, and PD engineers.
Test Development: Create targeted power characterization tests (e.g., peak power, di/dt stress patterns) for both simulation and silicon.
Silicon Validation: Measure CPU power and performance in the lab; correlate silicon results with pre-silicon estimates to refine models.
Cross-functional Collaboration: Partner with multiple engineering disciplines to achieve optimal power efficiency without compromising performance.
What we need to see:
BS/MS in EE, CE, or CS or equivalent experience.
3+ years of experience working in ASIC power measurement and optimization.
Strong understanding of leakage and dynamic power in VLSI circuits
Experience with RTL and netlist power analysis tools such as Power Artist, PrimeTime PX, or equivalent.
Familiarity with CPU microarchitecture (CPU pipeline design, out-of-order execution, cache hierarchy, branch prediction) and understanding of microarchitectural power model.
Ways to stand out from the crowd:
Proficiency in Python for automation and data analysis.
Experience with DVFS, clock gating, power gating, and multi-voltage domain design.
Knowledge of lab instrumentation for power measurement.
Strong communication skills for cross-team technical discussions.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What you’ll be doing:
Develop innovative parallel computing technologies that equip programmers with the compilers, languages, runtime systems, tools, and algorithms needed to solve challenges with modern accelerated systems.
Collaborate with other researchers and engineers to extend the state of the art in parallel computing, machine learning, data analytics, and other technology areas surrounding NVIDIA's business. Deliver your innovations in high-quality software systems, publications, and patents.
Engage with the research community through collaborations, publications, and presentations that produce new technologies and promote education in accelerated computing.
What we need to see:
Completing or recently completed a Doctoral degree (Ph.D.) or equivalent experience in a computational field such as computer science, computer engineering, or scientific computing.
Creativity in developing innovative solutions to the problems faced by parallel programmers and the skill to implement them in software.
Expertise in parallel programming and algorithmic techniques.
Strong programming ability in one or more of the following languages: C, C++, Rust and Python.
Track record of research excellence and publications that demonstrate your body of work.
Relevant research and software development experience.
Ways to stand out from the crowd:
Expertise in applying programming system insights and techniques to problems in machine learning, data science, and distributed computing.
Ability to implement ideas in the CUDA programming model.
Experience with applying AI, such as large language models, to create new ways of solving the problems of software engineering.
Prior success in building software systems used by other developers to solve their own problems.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

In this critical role, you will manage a team to expand NeMo Framework's capabilities, enabling users to develop, train, and optimize models by designing and implementing the latest in distributed training algorithms, model parallel paradigms, model optimizations, defining robust APIs, meticulously analyzing and tuning performance, and expanding our toolkits and libraries to be more comprehensive and coherent. You will collaborate with internal partners, users, and members of the open source community to analyze, design, and implement highly optimized solutions.
What you’ll be doing:
Plan, schedule, mentor, and lead the execution of projects and activities of the team.
Collaborate with internal customers to align priorities across business units.
Coordinate projects across different geographic locations.
Grow and develop a world-class team.
Contribute and advance open source
Solve large-scale, end-to-end AI training challenges, spanning the full model lifecycle from initial orchestration, data pre-processing, running of model training and tuning, to model deployment.
Work at the intersection ofcomputer-architecture,libraries, frameworks, AI applications and the entire software stack.
Innovate and improve model architectures, distributed training algorithms, and model parallel paradigms.
What we need to see:
Excellent understanding of SDLC practices including architecting, testing, continuous integration, and documentation
MS, PhD or equivalent experience in Computer Science, AI, Applied Math, or related field
8+ overall years of industry experience, including 3+ years of management experience.
Proven experience to lead and scale high-performing engineering teams, especially across distributed and functional groups.
Experience with AI Frameworks (e.g. PyTorch, JAX), and/or inference and deployment environments (e.g. TRTLLM, vLLM, SGLang).
Proficient in Python programming, software design, debugging, performance analysis, test design and documentation.
Consistent record of working effectively across multiple engineering initiatives and improving AI libraries with new innovations.
Ways to stand out from the crowd:
Hands-on experience in large-scale AI training, with a deep understanding of core compute system concepts (such as latency/throughput bottlenecks, pipelining, and multiprocessing) and demonstrated excellence in related performance analysis and tuning.
Expertise in distributed computing, model parallelism, and mixed precision training.
Prior experience with Generative AI techniques applied to LLM and Multi-Modal learning (Text, Image, and Video).
Knowledge of GPU/CPU architecture and related numerical software.
Created / contributed to open source deep learning frameworks.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

What you’ll be doing:
Working with tech giants to develop and demonstrate solutions based on NVIDIA’s groundbreaking software and hardware technologies.
Partnering with Sales Account Managers and Developer Relations Managers to identify and secure business opportunities for NVIDIA products and solutions.
Serving as the main technical point of contact for customers engaged in the development of intricate AI infrastructure, while also offering support in understanding performance aspects related to tasks like large scale LLM training and inference.
Conducting regular technical customer meetings for project/product details, feature discussions, introductions to new technologies, performance advice, and debugging sessions.
Collaborating with customers to build Proof of Concepts (PoCs) for solutions to address critical business needs and support cloud service integration for NVIDIA technology on hyperscalers.
Analyzing and developing solutions for customer performance issues for both AI and systems performance.
What we need to see:
BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or other Engineering fields or equivalent experience.
4+ years of engineering(performance/system/solution)experience.
Hands-on experience building performance benchmarks for data center systems, including large scale AI training and inference.
Understanding of systems architecture including AI accelerators and networking as it relates to the performance of an overall application.
Effective engineering program management with the capability of balancing multiple tasks.
Ability to communicate ideas clearly through documents, presentations, and in external customer-facing environments.
Ways to stand out from the crowd:
Hands-on experience with Deep Learning frameworks (PyTorch, JAX, etc.), compilers (Triton, XLA, etc.), and NVIDIA libraries (TRTLLM, TensorRT, Nemo, NCCL, RAPIDS, etc.).
Familiarity with deep learning architectures and the latest LLM developments.
Background with NVIDIA hardware and software, performance tuning, and error diagnostics.
Hands-on experience with GPU systems in general including but not limited to performance testing, performance tuning, and benchmarking.
Experience deploying solutions in cloud environments including AWS, GCP, Azure, or OCI as well as knowledge of DevOps/MLOps technologies such as Docker/containers, Kubernetes, data center deployments, etc. Command line proficiency.
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך