We are looking for a candidate with a proven track record in applied ML research. Responsibilities in the role will include training large scale multimodal (2D/3D vision-language) models on distributed backends, deployment of compact neural architectures efficiently on device, address growing set of safety challenges to make the model robust and aligned with human values. Ensuring quality in the wild, with an emphasis on model safety, fairness and robustness would constitute a meaningful part of the role. You will be interacting very closely with a variety of ML researchers, software engineers, hardware and design teams cross functionally. The primary responsibilities of the role would center on enriching multimodal capabilities of large language models. The user experience initiative would focus on aligning image/video content to the space of LMs for visual actions & multi-turn interactions.