Responsibilities
Research quantization and compression techniques for large language models (LLMs) and text-to-image generation models;Track and explore cutting-edge directions in efficient model deployment and inference acceleration.
Qualifications:Requirements
- Bachelor’s or Master’s degree candidate in Computer Science, Software Engineering, or related fields.
- Solid understanding of deep learning or large language model (LLM) fundamentals and applications;
- Familiarity with model compression techniques such as quantization and pruning;
- Proficiency in Python or other programming languages commonly used for deep learning development.
- Preferred Qualifications
- Strong self-motivation and problem-solving skills;
- Passion for technological innovation and practical engineering, with a drive for continuous exploration and improvement;
- Experience in model fine-tuning, inference optimization, or related tool development is a plus.
Student / InternShift 1 (China)PRC, Shanghai
This role will require an on-site presence. * Job posting details (such as work model, location or time type) are subject to change.