What you'll be doing:
Architect and build robust, scalable frameworks and APIs to integrate AI agents with Nvidia's suite of developer tools, simulation platforms, and enterprise software.
Collaborate with research, applied AI, and product teams to understand agent capabilities and translate them into production-ready services and features.
Design and implement CI/CD pipelines and MLOps practices tailored for the AI agent lifecycle, including testing, deployment, monitoring, and continuous improvement.
Optimize the performance, latency, and resource consumption of deployed AI agents, ensuring they operate efficiently within our compute infrastructure.
Develop internal tooling and automation to streamline the onboarding of new AI agents and enhance the observability of agent fleets.
Champion best practices for secure and reliable agent-based systems, including data handling, access control, and interaction protocols.
Serve as a key technical resource for solving sophisticated integration issues between AI agents and target applications.
What we need to see:
BSc or above in Computer Science, Computer Engineering, or a related field, or equivalent experience.
7+ years of hands-on experience in software engineering, with a focus on backend systems, cloud services, or infrastructure.
Proven experience with AI/ML frameworks (e.g., PyTorch, TensorFlow) and a strong understanding of large language models (LLMs), transformers, and agent-based architectures (e.g., LangChain, LlamaIndex).
Demonstrated experience in architecting, building and deploying complex integrations between AI agents and external tools, APIs, or software services.
Expert-level programming skills in Python. Experience with C++ is a strong plus.
Experience designing, building, and maintaining RESTful APIs, gRPC, and other service-to-service communication protocols.
Excellent problem-solving skills and the ability to navigate complex, ambiguous technical challenges.
Strong communication and interpersonal skills, with a proven ability to collaborate effectively across multidisciplinary teams.
Ways to stand out from the crowd:
Hands-on experience building or fine-tuning LLMs or other generative models.
Prior experience with MLOps, and agentic infrastructure.
Contributions to open-source AI/ML projects. Experience with Infrastructure as Code (Terraform, Ansible). Prior experience in developing platforms for internal developer communities.
knowledge of cloud platforms (AWS, GCP, Azure), container orchestration (Kubernetes, Docker), and building scalable microservices.
Familiarity with vector databases (e.g., Milvus, Pinecone) and model serving infrastructure (e.g., Triton Inference Server).
משרות נוספות שיכולות לעניין אותך