Provide deep insights of on-device ML model performance, as well as explore optimizations where appropriate.
Drive new capabilities for ML benchmarking service.
Play a key role in maintaining the health and performance of the service, including debugging failures and addressing user questions / requests.
Collaborate extensively with ML and hardware teams across Apple.
Strong ML fundamentals across training, evaluation and inference, and knowledge of modern model architectures such as Transformers, CNNs or Stable Diffusion;
Programming and software design skills (proficiency in Python and/or C/C++);
A passion for edge / on-device ML;
Understanding about performance modeling, analysis and profiling of computer systems, and how to optimize code run time and throughput for a given platform;
Collaboration, product-focus and excellent interpersonal skills.
Masters or PhDs in Computer Science or relevant disciplines;
On-device ML frameworks such as CoreML, TFLite or ExecuTorch;
Experience with any ML authoring framework (PyTorch, TensorFlow, JAX, etc.) is a strong plus;
Experience in software architecture, APIs, high performance extensible software and scalable software systems;
Understanding of how to optimize code run time and throughput for a given platform;
Interest and experience in power and/or hardware accelerators is a plus;
Back-end system skills including containers (docker), cloud orchestration (Kubernetes), database (SQL, Postgres).
Note: Apple benefit, compensation and employee stock programs are subject to eligibility requirements and other terms of the applicable plan or program.