

What you will be doing:
Investigate opportunities to improve communication performance by identifying bottlenecks in today's systems.
Design and implement new communication technologies to accelerate AI and HPC workloads.
Explore innovative solutions in HW and SW for our next generation platforms as part of co-design efforts involving GPU, Networking, and SW architects.
Build proofs-of-concept, conduct experiments, and perform quantitive modeling to evaluate and drive new innovations.
Use simulation to explore performance of large GPU clusters (think scales of 100s of 1000s of GPUs)
What we need to see:
M.S./Ph.D. degree in CS/CE or equivalent experience.
5+ years of relevant experience.
Excellent C/C++ programming and debugging skills.
Experience with parallel programming models (MPI, SHMEM) and at least one communication runtime (MPI, NCCL, NVSHMEM, OpenSHMEM, UCX, UCC).
Deep understanding of operating systems, computer and system architecture.
Solid in fundamentals of network architecture, topology, algorithms, and communication scaling relevant to AI and HPC workloads.
Strong experience with Linux.
Ability and flexibility to work and communicate effectively in a multi-national, multi-time-zone corporate environment.
Ways to stand out from the crowd:
Expertise in related technology and passion for what you do. Experience with CUDA programming and NVIDIA GPUs. Knowledge of high-performance networks like InfiniBand, RoCE, NVLink, etc.
Experience with Deep Learning Frameworks such PyTorch, TensorFlow, etc. Knowledge of deep learning parallelisms and mapping to the communication subsystem. Experience with HPC applications.
Strong collaborative and interpersonal skills and a proven track record of effectively guiding and influencing within a dynamic and multi-functional environment.
משרות נוספות שיכולות לעניין אותך

What you will be doing:
Work with a geographically distributed partner organisation to understand, modify and improve CPU Compiler SW at NVIDIA.
Contribute new features and optimisation techniques.
Develop compiler SW that is optimised for performance.
Be part of a team that is at the centre of AI, HPC and data centre technologies.
What we need to see:
BS or MS degree in Computer Science, Computer Engineering, or related field or equivalent with minimum 3 years of work experience
Experience with compiler development or a related academic or project.
Knowledge of Language Front-Ends or Compiler optimisation techniques and code generation modules.
Strong hands-on C++ programming skills
Excellent verbal and written communications skills
Ways to stand out from the crowd:
Familiarity with CPU architectures such as x86_64, PPC, RISC-V, Arm Architecture (AArch32, AArch64) or DSPs.
Familiarity with industry standard compiler infrastructure such as LLVM and / or GNU Toolchain
Knowledge of AI algorithms, scientific HPC applications and related code optimisations.
A track record of contributing to open source compiler communities.
משרות נוספות שיכולות לעניין אותך

What you'll be doing:
You will play a crucial role in ensuring the success of the Omniverse on DGX Cloud platform by helping to build our deployment infrastructure processes, creating world-class SRE measurement and creating automation tools to improve efficiency of operations, and maintaining a high standard of perfection in service operability and reliability.
Design, build, and implement scalable cloud-based systems for PaaS/IaaS.
Work closely with other teams on new products orfeatures/improvementsof existing products.
Develop, maintain and improve cloud deployment of our software.
Participate in the triage & resolution of complex infra-related issues
Collaborate with developers, QA and Product teams to establish, refine and streamline our software release process, software observability to ensure service operability, reliability, availability.
Maintain services once live by measuring and monitoring availability, latency, and overall system health using metrics, logs, and traces
Develop, maintain and improve automation tools that can help improve efficiency of SRE operations
Practice balanced incident response and blameless postmortems
Be part of an on-call rotation to support production systems
What we need to see:
BS or MS in Computer Science or equivalent program from an accredited University/College.
8+ years of hands-on software engineering or equivalent experience.
Demonstrate understanding of cloud design in the areas of virtualization and global infrastructure, distributed systems, and security.
Expertise in Kubernetes (K8s) & KubeVirt and building RESTful web services.
Understanding of building AI Agentic solutions preferably Nvidia open source AI solutions. Demonstrate working experiences in SRE principles like metrics emission for observability, monitoring, alerting using logs, traces and metrics
Hands on experience working with Docker, Containers and Infrastructure as a Code like terraform deployment CI/CD.
Exhibit knowledge in concepts of working with CSPs, for example: AWS (Fargate, EC2, IAM, ECR, EKS, Route53 etc...), Azure etc.
Ways to stand out from the crowd:
Expertise in technologies such as Stack-storm, OpenStack, Redhat OpenShift, AI DBs like Milvus.
A track record of solving complex problems with elegant solutions.
Prior experience with Go & Python, React.
Demonstrate delivery of complex projects in previous roles.
Showcase ability in developing Frontend application with concepts of SSA, RBAC
משרות נוספות שיכולות לעניין אותך

What you will be doing:
Study and develop cutting-edge techniques in machine learning, graphs, data analytics and deep learning, and perform in-depth analysis and optimization to ensure the best possible performance on current- and next-generation GPU architectures.
Work directly with key customers to understand the current and future problems they are solving and provide the best AI solutions using GPUs.
Collaborate closely with the architecture, research, libraries, tools, and system software teams at NVIDIA to influence the design of next-generation architectures, software platforms, and programming models.
What we need to see:
A Masters degree or PhD in an engineering or computer science related discipline and 3+ years of relevant work or research experience.
Strong knowledge of C/C++, software design, programming techniques, and AI algorithms.
Firsthand work experience with parallel programming, ideally CUDA C/C++.
Strong communication and organization skills, with a logical approach to problem solving, good time management, and task prioritization skills.
Some travel is required for conferences and for on-site visits with developers.
משרות נוספות שיכולות לעניין אותך

What You'll Be Doing:
Contributing to the development of CUDA Quantum by building core infrastructure for inter-device communication and efficient execution across multiple processors
Partnering with architects, product managers, and collaborators to create an extensible toolchain integrating quantum architecture specific components
Solving difficult problems at the intersection of compilers, HPC and quantum computing to enable ground-breaking research and technology
Discussing and refining software designs and implementation strategies with peers
Improving processes and infrastructure to accelerate our development
What We Need To See:
Bachelors Degree in Computer Science, Physics or related engineering field (Ph.D. or Masters preferred), or equivalent experience
5+ years of experience
Ability working on large-scale software projects, and a proven track record of building performant and robust production software
Proficiency in GPU-programming and a solid understanding of performance profiling, multi-processor systems, and compiler fundamentals
Ability to quickly develop expertise in new domains and products, and eagerness to master new challenges
Strong communication and collaboration skills
Extensive knowledge about quantum computing hardware and control systems and/or prior experience implementing optimization and code generation components for various quantum computing architectures
A passion for system designing and a focus on improving extensibility
Familiarity with FPGA programming and HDLs
Deep understanding of compiler toolchains, specifically LLVM/MLIR
משרות נוספות שיכולות לעניין אותך

What you will be doing:
Engage with our partners and customers to root cause functional and performance issues reported with NCCL
Conduct performance characterization and analysis of NCCL and DL applications on groundbreaking GPU clusters
Develop tools and automation to isolate issues on new systems and platforms, including cloud platforms (Azure, AWS, GCP, etc.)
Guide our customers and support teams on HPC knowledge and standard methodologies for running applications on multi-node clusters
Document and conduct trainings/webinars for NCCL
Engage with internal teams in different time zones on networking, GPUs, storage, infrastructure and support.
What we need to see:
B.S./M.S. degree in CS/CE or equivalent experience with 5+ years of relevant experience. Experience with parallel programming and at least one communication runtime (MPI, NCCL, UCX, NVSHMEM)
Excellent C/C++ programming skills, including debugging, profiling, code optimization, performance analysis, and test design
Experience working with engineering or academic research community supporting HPC or AI
Practical experience with high performance networking:Infiniband/RoCE/Ethernetnetworks, RDMA, topologies, congestion control
Expert in Linux fundamentals and a scripting language, preferably Python
Familiar with containers, cloud provisioning and scheduling tools (Docker, Docker Swarm, Kubernetes, SLURM, Ansible)
Adaptability and passion to learn new areas and tools
Flexibility to work and communicate effectively across different teams and timezones
Ways to stand out from the crowd:
Experience conducting performance benchmarking and developing infrastructure on HPC clusters. Prior system administration experience, esp for large clusters. Experience debugging network configuration issues in large scale deployments
Familiarity with CUDA programming and/or GPUs. Good understanding of Machine Learning concepts and experience with Deep Learning Frameworks such PyTorch, TensorFlow
Deep understanding of technology and passionate about what you do
משרות נוספות שיכולות לעניין אותך

As a Senior Robotics Software Engineer on the Isaac Applications Team, you will help build the platform for Physical AI robots — enabling sim-first development, real-world deployment, and continuous learning to make them smarter over time. The ideal candidate will have strong software engineering skills for (soft) realtime robotics applications and real-world experience with multi-body robots, such as humanoids or quadrupeds.
What You Will Be Doing
Bring the latest advancements in Physical AI to simulated and real humanoid robots by building the humanoid reference platform showcasing the power of NVIDIA's technology.
Collaborate across team boundaries to integrate NVIDIA robotics products such as Jetson Thor, Isaac Gr00t, and Isaac Sim/Lab into one solution for humanoid robots.
Take on a variety of challenges, bridging between research and commercial environments.
You will be deploying and testing your software on real robots in the lab.
What we need to see:
Master's degree in Robotics, Computer Science, Engineering, or a related field, or equivalent experience.
5+ years of experience.
Proven experience in designing and building real-time robotics software stacks.
Demonstrated success in bringing robotics solutions from lab to launch, with direct involvement in implementation and deployment.
Proficiency in Python, C++, PyTorch, CUDA, and middleware. Willingness to learn new languages and tools as needed.
Ability to adapt to fast-paced development lifecycles, multi-functional organizations, new technologies and platforms.
Strong problem-solving skills and the ability to determine the best approach to complex challenges.
Excellent communication skills and a collaborative approach to working effectively with diverse teams.
Ways to Stand out from The Crowd:
Familiarity with NVIDIA robotics products such as Jetson, Isaac Gr00t, Isaac ROS and Isaac Sim/Lab.
Experience with Machine Learning / Reinforcement Learning.
Experience with Bazel.
משרות נוספות שיכולות לעניין אותך

What you will be doing:
Investigate opportunities to improve communication performance by identifying bottlenecks in today's systems.
Design and implement new communication technologies to accelerate AI and HPC workloads.
Explore innovative solutions in HW and SW for our next generation platforms as part of co-design efforts involving GPU, Networking, and SW architects.
Build proofs-of-concept, conduct experiments, and perform quantitive modeling to evaluate and drive new innovations.
Use simulation to explore performance of large GPU clusters (think scales of 100s of 1000s of GPUs)
What we need to see:
M.S./Ph.D. degree in CS/CE or equivalent experience.
5+ years of relevant experience.
Excellent C/C++ programming and debugging skills.
Experience with parallel programming models (MPI, SHMEM) and at least one communication runtime (MPI, NCCL, NVSHMEM, OpenSHMEM, UCX, UCC).
Deep understanding of operating systems, computer and system architecture.
Solid in fundamentals of network architecture, topology, algorithms, and communication scaling relevant to AI and HPC workloads.
Strong experience with Linux.
Ability and flexibility to work and communicate effectively in a multi-national, multi-time-zone corporate environment.
Ways to stand out from the crowd:
Expertise in related technology and passion for what you do. Experience with CUDA programming and NVIDIA GPUs. Knowledge of high-performance networks like InfiniBand, RoCE, NVLink, etc.
Experience with Deep Learning Frameworks such PyTorch, TensorFlow, etc. Knowledge of deep learning parallelisms and mapping to the communication subsystem. Experience with HPC applications.
Strong collaborative and interpersonal skills and a proven track record of effectively guiding and influencing within a dynamic and multi-functional environment.
משרות נוספות שיכולות לעניין אותך