Bachelor's Degree in Computer Science, Electrical or Computer Engineering, or related field practical experience working in a highly performant team in an academic or industrial research environment
Advanced experience with PyTorch and Python
Experience with building large language models
Familiarity with Language models, transformers like BERT, GPT-3, Llama
Experience with Parameter-Efficient Fine-Tuning methods for foundational models, LoRA, DoRA etc.
Excellent communication and presentation skills
Preferred Qualifications
Master’s or PhD in Computer Science, Electrical or Computer Engineering, or related field and practical experience working in a highly performant team in an academic or industrial research environment.
Experience with model quantization such as GPTQ, AWQ etc.
Experience with distributed training libraries like DeepSpeed is a plus.
Record of publications in top-tier conferences or journals (ICLR, ACL, ICML, CVPR, ICCV, ECCV, NeurIPS, TPAMI, etc.)
Familiarity with AzureML, Azure DevOps and CI/CD
Responsibilities
You will utilize the latest research to optimize, fine-tune, and customize LLMs for edge device inferencing.
You will contribute to the technical design, architecture, development, and evaluation of foundation models.
You will document ongoing work and share findings to foster innovation within the group. You will also adhere to ethics and privacy policies during research processes and information collection.
You will collaborate with senior team members and integrate cutting-edge research. Additionally, you will gain a deep understanding of community methods and develop expertise in a specialized area.
You will work closely with engineering and product development teams.