

What you’ll be doing:
Working with Cloud Service Providers to develop and demonstrate solutions based on NVIDIA’s groundbreaking software and hardware technologies.
Build and deploy solutions at scale using NVIDIA's AI software on cloud-based GPU platforms.
Build custom PoCs for solutions that address customer's critical business needs while applying NVIDIA's hardware and software technologies.
Partner with Sales Account Managers or Developer Relations Managers to identify and secure business opportunities for NVIDIA products and solutions.
Conduct regular technical customer meetings for project/product details, feature discussions, intro to new technologies, and debugging sessions.
Prepare and deliver technical content to customers including presentations about purpose-built solutions, workshops about NVIDIA products and solutions, etc.
What we need to see:
BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or other Engineering fields or equivalent experience.
7+ years of Solutions Engineering (or similar Sales Engineering roles) experience.
Established track record of deploying AI/ ML solutions in cloud environments including AWS, GCP, Azure or OCI
Knowledge of DevOps/MLOps technologies such as Docker/containers, Kubernetes, data center deployments, etc.
Effective time management and capable of balancing multiple tasks.
Ability to communicate ideas clearly through documents, presentation, etc.
Ways to stand out from the crowd:
AWS, GCP, Azure or OCI Professional Solution Architect Certifications
Hands-on experience with NVIDIA GPUs and SDKs (i.e. CUDA, Triton, TensorRT-LLM, etc.)
Deep understanding of the full software development lifecycle, including best practices for system design, architectural patterns, and comprehensive testing.
Solid working knowledge of Python
System-level experience, specifically GPU-based systems
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך

You will collaborate closely with researchers to design and scale agents - enabling them to reason, plan, call tools and code just like human engineers. You will work on building and maintaining the core infrastructure for deploying and running these agents in production, powering all our agentic tools and applications and ensuring their seamless and efficient performance. If you're passionate about the latest research and cutting-edge technologies shaping generative AI, this role and team offer an exciting opportunity to be at the forefront of innovation.
What you'll be doing:
Design, develop, and improve scalable infrastructure to support the next generation of AI applications, including copilots and agentic tools.
Drive improvements in architecture, performance, and reliability, enabling teams to bring to bear LLMs and advanced agent frameworks at scale.
Collaborate across hardware, software, and research teams, mentoring and supporting peers while encouraging best engineering practices and a culture of technical excellence.
Stay informed of the latest advancements in AI infrastructure and contribute to continuous innovation across the organization.
What we need to see:
Master or PhD or equivalent experience in Computer Science or related field, with a minimum of 5 years in large-scale distributed systems or AIinfrastructure.
Advanced expertise in Python (required), strong experience with JavaScript, and deep knowledge of software engineering principles, OOP/functional programming, and writing high-performance, maintainable code.
Demonstrated expertise in crafting scalable microservices, web apps, SQL, and NoSQL databases (especially MongoDB and Redis) in production with containers, Kubernetes, and CI/CD.
Solid experience with distributed messaging systems (e.g., Kafka), and integrating event-driven or decoupled architectures into robust enterprise solutions.
Practical experience integrating and fine-tuning LLMs or agent frameworks (e.g., LangChain, LangGraph, AutoGen, OpenAI Functions, RAG, vector databases, timely engineering).
Demonstrated end-to-end ownership of engineering solutions, from architecture and development to deployment, integration, and ongoingoperations/support.
Excellent communication skills and a collaborative, proactive approach.
You will also be eligible for equity and .

What you’ll be doing:
Develop and test sample applications for chemistry and materials discovery using artificial intelligence.
Help develop AI first workflows using NVIDIA technology and popular deep learning frameworks.
Create clear, practical examples and documentation for developers and researchers.
What we need to see:
Pursuing a PhD in Chemistry, Materials Science, Computer Science, or a related field.
Familiarity with AI/ML concepts and experience with at least one deep learning framework (e.g., PyTorch, TensorFlow).
Basic understanding of chemistry or materials science principles.
Ways to stand out from the crowd:
Experience with GPU programming or CUDA and machine learning frameworks such as PyTorch.
Contributions to open-source projects related to AI or scientific computing.
Coursework or projects involving AI for scientific applications.
You will also be eligible for Intern
Applications for this job will be accepted at least until November 14,2025.NVIDIA
What you'll be doing:
As a senior member in our team, you will work with pre-silicon and post-silicon data analytics - visualization, insights and modeling.
Design and uphold sturdy data pipelines and ETL processes for the ingestion and processing of DFX Engineering data from various origins
Lead engineering efforts by collaborating with cross-functional teams (execution, analytics, data science, product) to define data requirements and ensure data quality and consistency
You will work on hard-to-solve problems in the Design For Test space which will involve application of algorithm design, using statistical tools to analyze and interpret complex datasets and explorations using Applied AI methods.
In addition, you will help develop and deploy DFT methodologies for our next generation products using Gen AI solutions.
You will also help mentor junior engineers on test designs and trade-offs including cost and quality.
What we need to see:
BSEE (or equivalent experience) with 5+, MSEE with 3+, or PhD with 1+ years of experience in low-power DFT, Data Visualization, Applied Machine Learning or Database Management.
Experience with SQL, ETL, and data modeling is crucial
Hands-on experience with cloud platforms (AWS, Azure, GCP)
Design and implement highly scalable, fault tolerant distributed database solutions
Lead data modeling, performance tuning, and capacity planning for large-scale, mission-critical storage workloads
Excellent knowledge in using statistical tools for data analysis & insights.
Strong programming and scripting skills in Perl, Python, C++ or Tcl is expected
Outstanding written and oral communication skills with the curiosity to work on rare challenges.
Ways to stand out from the crowd:
Experience in data pipeline and database architecture for real-world systems
Experience in application of AI for EDA-related problem-solving
Good understanding of technology and passionate about what you do
Strong collaborative and interpersonal skills, specifically a proven ability to effectively guide and influence within a dynamic environment
You will also be eligible for equity and .

NVIDIA is searching for a world-class researcher in generative AI to join our research team. You will be conducting original research for generative AI applications, including image generation, video generation, 3D generation, and audio generation. You will be working with a team of world-class researchers eager to make great impacts with generative AI models. You will be building research prototypes and scaling them with large datasets and compute. After building prototypes that demonstrate the promise of your research, you will work with product teams to help them integrate your ideas into products.
What you'll be doing:
Conduct original research in the space of generative AI
Implement and train large-scale generative AI models for various content creation applications
Collaborate with other research team members, a diverse set of internal product teams, and external researchers
Have a broader impact through the transfer of the technology you've developed to relevant product groups
What we need to see:
Ph.D. in Computer Science/Engineering, Electrical Engineering, or a related field (or equivalent experience).
5+ years of relevant research experience.
Excellent collaboration and interpersonal skills
Excellent python/C++ programming skills
Great knowledge of common deep-learning frameworks
Experience in processing or curating large-scale datasets
Excellent knowledge of theory and practice of deep learning, computer vision, natural language processing, or computer graphics
Track record of research excellence or significant product development
You will also be eligible for equity and .

What you’ll be doing:
Contribute features to vLLM that empower the newest models with the latest NVIDIA GPU hardware features; profile and optimize the inference framework (vLLM) with methods like speculative decoding,data/tensor/expert/pipeline-parallelism,prefill-decode disaggregation.
Develop, optimize, and benchmark GPU kernels (hand-tuned and compiler-generated) using techniques such as fusion, autotuning, and memory/layout optimization; build and extend high-level DSLs and compiler infrastructure to boost kernel developer productivity while approaching peak hardware utilization.
Define and build inference benchmarking methodologies and tools; contribute both new benchmark and NVIDIA’s submissions to the industry-leading MLPerf Inference benchmarking suite.
Architect the scheduling and orchestration of containerized large-scale inference deployments on GPU clusters across clouds.
Conduct and publish original research that pushes the pareto frontier for the field of ML Systems; survey recent publications and find a way to integrate research ideas and prototypes into NVIDIA’s software products.
What we need to see:
Bachelor’s degree (or equivalent expeience) in Computer Science (CS), Computer Engineering (CE) or Software Engineering (SE) with 7+ years of experience; alternatively, Master’s degree in CS/CE/SE with 5+ years of experience; or PhD degree with the thesis and top-tier publications in ML Systems, GPU architecture, or high-performance computing.
Strong programming skills in Python and C/C++; experience with Go or Rust is a plus; solid CS fundamentals: algorithms & data structures, operating systems, computer architecture, parallel programming, distributed systems, deep learning theories.
Knowledgeable and passionate about performance engineering in ML frameworks (e.g., PyTorch) and inference engines (e.g., vLLM and SGLang).
Familiarity with GPU programming and performance: CUDA, memory hierarchy, streams, NCCL; proficiency with profiling/debug tools (e.g., Nsight Systems/Compute).
Experience with containers and orchestration (Docker, Kubernetes, Slurm); familiarity with Linux namespaces and cgroups.
Excellent debugging, problem-solving, and communication skills; ability to excel in a fast-paced, multi-functional setting.
Ways to stand out from the crowd
Experience building and optimizing LLM inference engines (e.g., vLLM, SGLang).
Hands-on work with ML compilers and DSLs (e.g., Triton,TorchDynamo/Inductor,MLIR/LLVM, XLA), GPU libraries (e.g., CUTLASS) and features (e.g., CUDA Graph, Tensor Cores).
Experience contributing tocontainerization/virtualizationtechnologies such ascontainerd/CRI-O/CRIU.
Experience with cloud platforms (AWS/GCP/Azure), infrastructure as code, CI/CD, and production observability.
Contributions to open-source projects and/or publications; please include links to GitHub pull requests, published papers and artifacts.
You will also be eligible for equity and .

What you'll be doing:
Working with NVIDIA AI Native customers on data center GPU server and networking infrastructure deployments.
Guiding customer discussions on network topologies, compute/storage, and supporting the bring-up ofserver/network/clusterdeployments.
Identifying new project opportunities for NVIDIA products and technology solutions in data center and AI applications.
Conducting regular technical meetings with customers as a trusted advisor, discussing product roadmaps, cluster debugging, and new technology introductions.
Building custom demonstrations and proofs of concept to address critical business needs.
Analyzing and debugging compute/network performance issues.
What we need to see:
BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or related fields, or equivalent experience.
5+ years of experience in Solution Engineering or similar roles.
System-level understanding of server architecture, NICs, Linux, system software, and kernel drivers.
Practical knowledge of networking - switching & routing for Ethernet/Infiniband, and data center infrastructure (power/cooling).
Familiarity with DevOps/MLOps technologies such as Docker/containers and Kubernetes.
Effective time management and ability to balance multiple tasks.
Excellent communication skills for articulating ideas and code clearly through documents and presentations.
Ways to stand out from the crowd:
External customer-facing skills and experience.
Experience with the bring-up and deployment of large clusters.
Proficiency in systems engineering, coding, and debugging, including C/C++, Linux kernel, and drivers.
Hands-on experience with NVIDIA systems/SDKs (e.g., CUDA), NVIDIA networking technologies (e.g., DPU or equivalent experience, RoCE, InfiniBand), and/or ARM CPU solutions.
Familiarity with virtualization technology concepts.
You will also be eligible for equity and .

What you’ll be doing:
Working with Cloud Service Providers to develop and demonstrate solutions based on NVIDIA’s groundbreaking software and hardware technologies.
Build and deploy solutions at scale using NVIDIA's AI software on cloud-based GPU platforms.
Build custom PoCs for solutions that address customer's critical business needs while applying NVIDIA's hardware and software technologies.
Partner with Sales Account Managers or Developer Relations Managers to identify and secure business opportunities for NVIDIA products and solutions.
Conduct regular technical customer meetings for project/product details, feature discussions, intro to new technologies, and debugging sessions.
Prepare and deliver technical content to customers including presentations about purpose-built solutions, workshops about NVIDIA products and solutions, etc.
What we need to see:
BS/MS/PhD in Electrical/Computer Engineering, Computer Science, Physics, or other Engineering fields or equivalent experience.
7+ years of Solutions Engineering (or similar Sales Engineering roles) experience.
Established track record of deploying AI/ ML solutions in cloud environments including AWS, GCP, Azure or OCI
Knowledge of DevOps/MLOps technologies such as Docker/containers, Kubernetes, data center deployments, etc.
Effective time management and capable of balancing multiple tasks.
Ability to communicate ideas clearly through documents, presentation, etc.
Ways to stand out from the crowd:
AWS, GCP, Azure or OCI Professional Solution Architect Certifications
Hands-on experience with NVIDIA GPUs and SDKs (i.e. CUDA, Triton, TensorRT-LLM, etc.)
Deep understanding of the full software development lifecycle, including best practices for system design, architectural patterns, and comprehensive testing.
Solid working knowledge of Python
System-level experience, specifically GPU-based systems
You will also be eligible for equity and .
משרות נוספות שיכולות לעניין אותך