Share
What You’ll Be Doing:
Help drive NVIDIA’s inference platform technical go-to-market efforts
Work closely with engineering and product management teams to understand key technical capabilities of our inference stack from GPUs, CPUs, networking, CUDA libraries, model architectures and deployment techniques (e.g.parallelisms, configurations, etc.)
Diligently review and remain up to date on model architectures, frameworks, arxiv papers, whitepapers deployment techniques (e.g.disaggregated serving, KV cache implementations) and identify intersection points between the latest AI models and NVIDIA’s platform to maximize performance and minimize TCO
Develop crisp clear positioning, messaging and assets to highlight NVIDIA’s leadership position in inference. Assets (blogs, whitepapers, presentations, analyst briefings, seminars at developer conferences)
Closely follow competitive inference announcements and prepare appropriate responses for business and technical/developer audiences
Assist on building keynote slides for executives for areas that you’re a subject matter expert
What We Need to See:
A BS Degree in Computer Science or Engineering or related field or equivalent experience in a technical product marketing role; Masters Degree preferred.
6+ years of experience in LLM, AI/ML development in an engineering role followed by 5+ years of experience in product management or technical product marketing of AI/ML products
Deep understanding of modern data center architectures, accelerated computing, distributed inference, deep learning frameworks (PyTorch, TensorFlow, JAX), and inference-specific frameworks & optimizations (Dynamo, Triton Inference Server, TensorRT-LLM, vLLM, SGLang)
Market Awareness – Experience conducting technical competitive analysis and synthesizing key insights
Collaboration & Influence – Proven ability to work cross-functionally across engineering, product management, sales, and marketing teams
Strong Communication, Asset Creation & Storytelling – Ability to translate sophisticated technical concepts into clear, compelling narratives for both technical and business audiences
Ability to present to executive audiences
Ways to Stand Out from the crowd:
Hands-on experience with AI inferencing workflows using NVIDIA or open-source serving frameworks running on accelerated computing in the data center
Experience developing LLM models
Experience working with hyperscale cloud providers
Hands-on Technical Competence – Background in software development, AI infrastructure, data center silicon
Demonstrated ability to engage with executive leadership and external partners
Published technical content or speaking experience at industry events
Have a portfolio of published marketing/launch assets
You will also be eligible for equity and .
These jobs might be a good fit