

What you will be doing:
You will work and develop state of the art techniques in deep learning, graphs, machine learning, and data analytics, and perform in-depth analysis and optimization to ensure the best possible performance on current- and next-generation GPU architectures
You will provide the best AI solutions using GPUs working directly with key customers
Collaborate closely with the architecture, research, libraries, tools, and system software teams to influence the design of next-generation architectures, software platforms, and programming models
What we need to see:
Pursuing MS or PhD from a leading University in an engineering or Computer Science related discipline
Strong knowledge of C/C++, software design, programming techniques, and AI algorithms
Experience with parallel programming, ideally CUDA C/C++
Good communication and organization skills, with a logical approach to problem solving, time management, and task prioritization skills
Preferred internship duration: 4+ months
משרות נוספות שיכולות לעניין אותך

We’re working on the next generation of recommendation tools and pushing the boundaries of accelerating model training and inference on GPU. You’ll join a team of ML, HPC and Software Engineers and Applied Researcher developing a framework designed to make the productization of GPU-based recommender systems as simple and fast as possible.
What you’ll be doing
In your role as CUDA Engineer Intern you will be profiling and investigating the performance of optimized code together within our HPC team. Part of this job will be to perform tests, unit tests and validate the numerical performance and correctness of the code. You will discuss your approach and results together with our CUDA engineers.
What we need to see:
Experience with c++, CUDA, python and Linux.
Bachelor or Master degree in software engineering or technical field such as mathematics or applied science.
communication skills
ambitious to grow and learn about building machine learning applications, optimization and software engineering.

What you will be doing:
Work and develop state of the art techniques in deep learning, graphs, machine learning, and data analytics, and perform in-depth analysis and optimization to ensure the best possible performance on current- and next-generation GPU architectures
You will provide the best AI solutions using GPUs working directly with key customers
Collaborate closely with the architecture, research, libraries, tools, and system software teams to influence the design of next-generation architectures, software platforms, and programming models
What we need to see:
Available to work as intern for at least 3 months+, 3 days per week
Pursuing MS or PhD from a leading University in an Engineering or Computer Science related discipline and will graduate in 2027
Strong knowledge of C/C++, software design, programming techniques, and AI algorithms
Experience with parallel programming, ideally CUDA C/C++
Good communication and organization skills, with a logical approach to problem solving, time management, and task prioritization skills

What you'll be doing:
Designing, developing, and testing software components for the NVOS operating system.
Collaborating with firmware, hardware, and QA teams on feature integration and validation.
Debugging, profiling, and optimizing existing code for performance and reliability.
Supporting software releases and maintaining product stability through continuous improvements.
Support internal and external customers for any NVOS related items.
What we need to see:
B.Sc. in Computer Science, Software Engineering, or Electrical Engineering.
2+ years of hands-on experience programming in C++ and Python.
Working knowledge of networking and operating system fundamentals.
Experience using AI-assisted development tools.
Proficiency working in a Linux environment.
Proactive, open-minded, and a quick learner.
Well-organized, agile, and capable of leading your own tasks.
Collaborative personality with a love for teamwork.
Ways to stand out from the crowd:
Background in switch and/or networking environment – advantage.
Experience with SONiC.
Exposure to kernel or driver-level development.
Experience working with customers – advantage.
Acted as a tech lead or a scrum master – advantage.

What you’ll be doing:
Drive the implementation and deployment of NVIDIA Inference Microservice (NIM) solutions
Use NVIDIA NIM Factory Pipeline to package optimized models (including LLM, VLM, Retriever, CV, OCR, etc.) into containers providing standardized API access
Refine NIM tools for the community, help the community to build their performant NIMs
Design and implement agentic AI tailored to customer business scenarios using NIMs
Deliver technical projects, demos and customer support tasks
Provide technical support and guidance to customers, facilitating the adoption and implementation of NVIDIA technologies and products
Collaborate with cross-functional teams to enhance and expand our AI solutions
What we need to see:
Pursuing Bachelor or Master in Computer Science, AI, or a related field; Or PhD candidates in ML Infra or data systems for ML.
Proficiency in at least one inference framework (e.g., TensorRT, ONNX Runtime, PyTorch)
Strong programming skills in Python or C++
Excellent problem-solving skills and ability to troubleshoot complex technical issues
Demonstrated ability to collaborate effectively across diverse, global teams, adapting communication styles while maintaining clear, constructive professional interactions
Ways to stand out from the crowd:
Expertise in model optimization techniques, particularly using TensorRT
Familiarity with disaggregated LLM Inference
CUDA optimization experience, extensive experience designing and deploying large scale HPC and enterprise computing systems
Familiarity with main stream inference engines (e.g., vLLM, SGLang)
Experience with DevOps/MLOps such as Docker, Git, and CI/CD practices

What you'll be doing:
Define the strategy and tactics for successfully integrating complex cloud software stacks, fromevaluation/proof-of-concepttodesign/development/qualification/validation,through customer acceptance
Execute the plan you craft by: scoping the project, identifying and assembling the core support team, formally kicking off the project with customers, developing a joint project timeline, supporting the design-in effort by providing technical collateral/tools, leading engineering design reviews, ensuring the implementation of certain customer-required features, driving resolution of any blocking issues, and ensuring timely customer acceptance and production deployment
Translate customer requirements into specific tasks and assign those tasks to relevant functional areas with ease. Collecting, prioritizing, tracking, and driving customer-reported issues to resolution/closure is a key focus of this role.
Clearly and intelligently communicate issue status to the customer and internal (including executive) teams
Work closely with engineering, product management, and sales teams to resolve technical issues, set correct expectations and keep the scope of the projects on a track to successful execution
Effectively organize and lead complex technicalengineering-to-engineeringdiscussions
Correctly represent the urgency of issues and raise them (internally and/or externally) as appropriate
Regularly communicate program status and key issues to management
Develop a close working relationship with customer development teams and use creativity to find solutions to their issues
What we need to see:
BS/MS Engineering or Computer Science or equivalent experience
8+ years in a high-tech electronics industry with a focus on program/project management
Experience and therefore deep customer savvy born of running significant and complex software engagement programs, especially with strategic customers with large datacenter deployments
Strong technical savvy, including software qualification and deployment for datacenter
Supreme leadership skills across broad and diverse functional teams
Validated history of 5+ years of project/program management of software product solutions
Strong technical background and experience working with customers in a clouddevelopment/deploymentand/or OEM environment
Experience leading global projects
Ways to stand out from the crowd:
Graphics, and/or cloud datacenterdevelopment/deploymentexperience desirable
Comfortable defining and managing complex program schedules with a deep understanding of dependencies, critical paths, key and critical landmarks, and work you're doing
Consistent record of handling multiple customers and customer programs across worldwide geographies and time zones

What you'll be doing:
Driving the infrastructure support for the open-source-first strategy of software development and deployment
Building scalable automation for build, test, integration, and release processes for publicly distributed deep learning libraries
Developing throughout the software stack, from the user experience and user interfaces down to the cluster and database layers
Configuring, maintaining, and building upon deployments of industry-standard tools (e.g. Kubernetes, Jenkins, Docker, CMake, Gitlab, Jira, etc.)
Advancing the state of the art in those industry-standard tools
What we need to see:
A Masters Degree in Computer Science or Computer Engineering or equivalent experience.
3+ years of relevant experience
Strong programming skills in Python (or similar) and familiarity with C/C++ development
Experience setting up, maintaining, and automating continuous integration systems (e.g. Jenkins, GitHub Actions, GitLab pipelines, Azure DevOps)
Fluency in SCM (e.g. Git, Perforce) and build systems (e.g. Make, CMake, Bazel)
Ways to stand out from the crowd:
Experience with infrastructure support for Github open-source projects
Track record of active involvement in open-source community
Experience designing and developing automation in Jenkins with Groovy (or similar)
Background with distributed systems and cluster/cloud computing, especially with Kubernetes
Experience with mobile/embedded platforms and multiple operating systems (Ubuntu, RedHat, Windows, QNX, or similar)

What you will be doing:
You will work and develop state of the art techniques in deep learning, graphs, machine learning, and data analytics, and perform in-depth analysis and optimization to ensure the best possible performance on current- and next-generation GPU architectures
You will provide the best AI solutions using GPUs working directly with key customers
Collaborate closely with the architecture, research, libraries, tools, and system software teams to influence the design of next-generation architectures, software platforms, and programming models
What we need to see:
Pursuing MS or PhD from a leading University in an engineering or Computer Science related discipline
Strong knowledge of C/C++, software design, programming techniques, and AI algorithms
Experience with parallel programming, ideally CUDA C/C++
Good communication and organization skills, with a logical approach to problem solving, time management, and task prioritization skills
Preferred internship duration: 4+ months
משרות נוספות שיכולות לעניין אותך