מציאת משרת הייטק בחברות הטובות ביותר מעולם לא הייתה קלה יותר
What you'll be doing:
In this role you will be building and improving our ecosystem around GPU-accelerated computing including developing large scale automation solutions. You will also be maintaining and building deep learning AI-HPC GPU clusters at scale and supporting our researchers to run their flows on our clusters including performance analysis and optimizations of deep learning workflows. You will design, implement and support operational and reliability aspects of large scale distributed systems with focus on performance at scale, real time monitoring, logging, and alerting.
Design and implement state-of-the-art GPU compute clusters.
Optimize cluster operations for maximum reliability, efficiency, and performance.
Drive foundational improvements and automation to enhance researcher productivity.
Troubleshoot, diagnose, and root cause of system failures and isolate the components/failure scenarios while working with internal & external partners.
Scale systems sustainably through mechanisms like automation, and evolve systems by pushing for changes that improve reliability and velocity.
Practice sustainable incident response and blameless postmortems and Be part of an on-call rotation to support production systems
Write and review code, develop documentation and capacity plans, debug the hardest problems, live, on some of the largest and most complex systems in the world.
Implement remediations across software and hardware stack according to plan, while keeping a thorough procedural record and data log and Manage upgrades and automated rollbacks across all clusters.
What we need to see:
Bachelor’s degree in computer science, Electrical Engineering or related field or equivalent experience with a minimum5+ yearsof experience designing and operating large scale compute infrastructure.
Proven experience in site reliability engineering for high-performance computing environments with operational experience of at least2K GPUscluster.
Deep understanding of GPU computing and AI infrastructure.
Passion for solving complex technical challenges and optimizing system performance.
Experience with AI/HPC advanced job schedulers, and ideally familiarity with schedulers such as Slurm.
Working knowledge of cluster configuration management tools such as BCM or Ansible and infrastructure level applications, such as Kubernetes, Terraform, MySQL, etc.
In depth understating of container technologies like Docker, Enroot, etc.
Experience programming in Python and Bash scripting.
Ways to stand out from the crowd:
Interest in crafting, analyzing, and fixing large-scale distributed systems.
Familiarity with NVIDIA GPUs, Cuda Programming, NCCL, MLPerf benchmarking, InfiniBand with IBoIP and RDMA.
Experience with Cloud Deployment, BCM, Terraform.
Understanding of fast, distributed storage systems like Lustre and GPFS for AI/HPC workloads.
Multi-cloud experience.
משרות נוספות שיכולות לעניין אותך