

Share
What you’ll be doing:
Analyze state of the art DL networks (LLM etc.), identify and prototype performance opportunities to influence SW and Architecture team for NVIDIA's current and next gen inference products.
Develop analytical models for the state of the art deep learning networks and algorithm to innovate processor and system architectures design for performance and efficiency.
Specify hardware/software configurations and metrics to analyze performance, power, and accuracy in existing and future uni-processor and multiprocessor configurations.
Collaborate across the company to guide the direction of next-gen deep learning HW/SW by working with architecture, software, and product teams.
What we need to see:
BS or higher degree in a relevant technical field (CS, EE, CE, Math, etc.).
Strong programming skills in Python, C, C++.
Strong background in computer architecture.
Experience with performance modeling, architecture simulation, profiling, and analysis.
Prior experience with LLM or generative AI algorithms.
Ways to stand out from the crowd:
GPU Computing and parallel programming models such as CUDA and OpenCL.
Architecture of or workload analysis on other deep learning accelerators.
Deep neural network training, inference and optimization in leading frameworks (e.g. Pytorch, TensorRT-LLM, vLLM, etc.).
Open-sourceAIcompilers (OpenAI Triton, MLIR, TVM, XLA, etc.).
and proud to be an
These jobs might be a good fit

Share
NVIDIA has been transforming computer graphics, PC gaming, and accelerated computing for more than 25 years.a unique legacy of innovationfueled by great technology—and amazing people.
What you'll be doing:
establishintegrations with NVIDIA Cloud Partners, enabling global developers to easily access GPU-optimized virtual machines.
You will craft and implement IaaS API integrations, collaborating with external engineering teams to ensure reliable, scalable, and consistent connectivity across diverse cloud environments.
Shape integration strategies, develop stateful workflow orchestration, and drive improvements in testing, observability, and automation to ensure high-quality, fault-tolerant solutions.
Be responsible for developing the two-sided marketplace, including the integration ofcomputeproviders and crafting discovery and bidding experiences to match supply with demand.
What we need to see:
5+ years of experience in developing software infrastructure for large-scale AI systems, with a proventrack recordof impact.
Expertise in software engineering withkubernetes, including cluster operations, operator development, node health monitoring, and GPU resource scheduling.
Familiarity with setting up cloud infrastructure environments (VMaaS, VPCs, RDMA, sharedfile-systems).
Proven ability to handle 3rd party API integrations, including communication with external teams, writing API clients, and improving integration reliability.
Comfort in a fast-paced environment, with the ability to collaborate and debug integrations with external engineering teams.
Strong technical knowledge, including proficiencyin a systems programming language (preference for Go) and a solid understanding of software design patterns for stateful workflow orchestration.
BS in Computer Science, Engineering, Physics, Mathematics, or a comparable degree or equivalent experience.
, distributed systems, and API development.
These jobs might be a good fit

Share
Intelligent machines powered by AI computers that can learn, reason and interact with people are no longer science fiction. Today, a self-driving car can meander through a country road at night and find its way. An AI-powered robot can learn motor skills through trial and error. This is truly an extraordinary time — the era of AI has begun. Image recognition and speech recognition — GPU Deep Learning has provided the foundation for machines to learn, perceive, reason and solve problems. The GPU started out as the engine for simulating human creativity, conjuring up the amazing virtual worlds of video games and Hollywood films.
What you will be doing:
Study and develop cutting-edge techniques in deep learning, graphs, machine learning, and data analytics, and perform in-depth analysis and optimization to ensure the best possible performance on current- and next-generation GPU architectures.
Work directly with key customers to understand the current and future problems they are solving and provide the best AI solutions using GPUs.
Collaborate closely with the architecture, research, libraries, tools, and system software teams at NVIDIA to influence the design of next-generation architectures, software platforms, and programming models.
Some travel is required for conferences and for on-site visits with customers.
What we need to see:
A degree from a leading university in an engineering or computer science related discipline (BS; MS or PhD preferred) or equivalent experience
Strong knowledge of C/C++, software design, programming techniques, or AI algorithms
Strong verbal and written communication skills in English and organization skills, with a logical approach to problem solving, time management, and task prioritization skills
These jobs might be a good fit

Share
What you’ll be doing:
Apply Machine Learning, Deep Learning techniques to overcome QA and Automation challenges for different NVIDIA Product lines
Maintain and optimize new and existing automated tests with AI solution
Design, implement and augment existing test frameworks with AI automation tools to reduce manual development efforts and make the current automation framework more productive and smart
Work closely and collaboratively with other development and QA teams to gather automation requirements and review automation design
What we need to see:
Pursuing Master or higher degree in computer science or electrical engineering with strong academics
Good programming skills in python, proficient in applying OOP concepts and Data structures
Good experience with using AI development tools for test plan creation, automation framework and test case development
Have experience on using Cursor, MCP, Coderabbit for script generation/review
Hands on experience in solving complex problems using AI technology or tools such as RAG, LLM/vLLM and AIGC
Experience with large or complex applications and knowledge of code optimization and manipulation
Demonstrates excellent communication skills, maintains a strong sense of initiative and motivation, and consistently upholds high standards for software quality
Ability to work closely & collaboratively with other Development & QA teams especially across geographies
Ways to stand out from the crowd:
Programming skills in C# is a plus
Knowledge of Linux, Mac and Android is a plus
Experience on Machine Learning Frameworks like Pytorch, Keras with TensorFlow, ONNX , and TensorRT is a plus
Background of RL(Reinforce learning), or Meta-learning, or Life-long learning is a plus.
Knowledge of SQL/NoSQL Database is plus
These jobs might be a good fit

Share
What you will be doing:
Develop and optimize the control stack, including locomotion, manipulation, and whole-body control algorithms;
Deploy and evaluate neural network models in physics simulation and on real humanoid hardware;
Design and maintain teleoperation software for controlling humanoid robots with low latency and high precision;
Implement tools and processes for regular robot maintenance, diagnostics, and troubleshooting to ensure system reliability;
Monitor teleoperators at the lab and develop quality assurance workflows to ensure high-quality data collection;
Collaborate with researchers on model training, data processing, and MLOps lifecycle.
What we need to see:
Bachelor’s degree in Computer Science, Robotics, Engineering, or a related field;
3+ years of full-time industry experience in robotics hardware or software full-stack;
Hands-on experience with deploying and debugging neural network models on robotic hardware;
Ability to implement real-time control algorithms, teleoperation stack, and sensor fusion;
Proficiency in languages such as Python, C++, and experience with robotics frames (ROS) and physics simulation (Gazebo, Mujoco, Isaac, etc.).
Experience in maintaining and troubleshooting robotic systems, including mechanical, electrical, and software components.
Physically work on-site on all business days.
Ways to stand out from the crowd:
Master’s or PhD’s degree in Computer Science, Robotics, Engineering, or a related field;
Experience at humanoid robotics companies on real hardware deployment;
Experience in robot hardware design;
Demonstrated Tech Lead experience, coordinating a team of robotics engineers and driving projects from conception to deployment.
These jobs might be a good fit

Share
What you will be doing:
Collaborate closely with our AI/ML researchers to make their ML models more efficient leading to significant productivity improvements and cost savings
Build tools, frameworks, and apply ML techniques to detect & analyze efficiency bottlenecks and deliver productivity improvements for our researchers
Work with researchers working on a variety of innovative ML workloads across Robotics, Autonomous vehicles, LLM’s, Videos and more
Collaborate across the engineering organizations to deliver efficiency in our usage of hardware, software, and infrastructure
Proactively monitor fleet wide utilization patterns, analyze existing inefficiency patterns, or discover new patterns, and deliver scalable solutions to solve them
Keep up to date with the most recent developments in AI/ML technologies, frameworks, and successful strategies, and advocate for their integration within the organization.
What we need to see:
BS or similar background in Computer Science or related area (or equivalent experience)
Minimum 8+ years of experience designing and operating large scale compute infrastructure
Strong understanding of modern ML techniques and tools
Experience investigating, and resolving, training & inference performance end to end
Debugging and optimization experience with NSight Systems and NSight Compute
Experience with debugging large-scale distributed training using NCCL
Proficiency in programming & scripting languages such as Python, Go, Bash, as well as familiarity with cloud computing platforms (e.g., AWS, GCP, Azure) in addition to experience with parallel computing frameworks and paradigms.
Dedication to ongoing learning and staying updated on new technologies and innovative methods in the AI/ML infrastructure sector.
Excellent communication and collaboration skills, with the ability to work effectively with teams and individuals of different backgrounds
Ways to stand out from the crowd:
Background with NVIDIA GPUs, CUDA Programming, NCCL and MLPerf benchmarking
Experience with Machine Learning and Deep Learning concepts, algorithms and models
Familiarity with InfiniBand with IBOP and RDMA
Understanding of fast, distributed storage systems like Lustre and GPFS for AI/HPC workloads
Familiarity with deep learning frameworks like PyTorch and TensorFlow
These jobs might be a good fit

Share
What you’ll be doing:
Analyze state of the art DL networks (LLM etc.), identify and prototype performance opportunities to influence SW and Architecture team for NVIDIA's current and next gen inference products.
Develop analytical models for the state of the art deep learning networks and algorithm to innovate processor and system architectures design for performance and efficiency.
Specify hardware/software configurations and metrics to analyze performance, power, and accuracy in existing and future uni-processor and multiprocessor configurations.
Collaborate across the company to guide the direction of next-gen deep learning HW/SW by working with architecture, software, and product teams.
What we need to see:
BS or higher degree in a relevant technical field (CS, EE, CE, Math, etc.).
Strong programming skills in Python, C, C++.
Strong background in computer architecture.
Experience with performance modeling, architecture simulation, profiling, and analysis.
Prior experience with LLM or generative AI algorithms.
Ways to stand out from the crowd:
GPU Computing and parallel programming models such as CUDA and OpenCL.
Architecture of or workload analysis on other deep learning accelerators.
Deep neural network training, inference and optimization in leading frameworks (e.g. Pytorch, TensorRT-LLM, vLLM, etc.).
Open-sourceAIcompilers (OpenAI Triton, MLIR, TVM, XLA, etc.).
and proud to be an
These jobs might be a good fit

Share
What you’ll be doing:
Analyze state of the art DL networks (LLM etc.), identify and prototype performance opportunities to influence SW and Architecture team for NVIDIA's current and next gen inference products.
Develop analytical models for the state of the art deep learning networks and algorithm to innovate processor and system architectures design for performance and efficiency.
Specify hardware/software configurations and metrics to analyze performance, power, and accuracy in existing and future uni-processor and multiprocessor configurations.
Collaborate across the company to guide the direction of next-gen deep learning HW/SW by working with architecture, software, and product teams.
What we need to see:
BS or higher degree in a relevant technical field (CS, EE, CE, Math, etc.).
Strong programming skills in Python, C, C++.
Strong background in computer architecture.
Experience with performance modeling, architecture simulation, profiling, and analysis.
Prior experience with LLM or generative AI algorithms.
Ways to stand out from the crowd:
GPU Computing and parallel programming models such as CUDA and OpenCL.
Architecture of or workload analysis on other deep learning accelerators.
Deep neural network training, inference and optimization in leading frameworks (e.g. Pytorch, TensorRT-LLM, vLLM, etc.).
Open-sourceAIcompilers (OpenAI Triton, MLIR, TVM, XLA, etc.).
and proud to be an
These jobs might be a good fit