In this role, you will be responsible for operationalizing machine learning models—from building real-time and batch inference pipelines to optimizing system performance, reliability, and experimentation velocity. You’ll help bridge the gap between research and production by developing the infrastructure, tooling, and monitoring required to ship ML-driven features safely and efficiently. Key Responsibilities* Partner with ML researchers and product teams to transition models into production, ensuring reliability, scalability, and low latency.* Design and implement robust inference services using object-oriented languages (e.g., Java, Scala, C++) that operate at scale across Apple platforms.* Build and manage data pipelines and model execution frameworks to support both batch and streaming use cases.* Develop tooling and infrastructure for model deployment, versioning, rollback, and online evaluation.* Lead A/B testing efforts, including integration, metric tracking, experiment validation, and performance analysis.* Collaborate with infrastructure teams to improve observability, alerting, and model health monitoring.* Drive continuous improvement in latency, throughput, fault tolerance, and overall system reliability.