Responsible for building multimodal large models, including but not limited to data engineering design, algorithm design, and model development, to create industry-leading multimodal large language models.
Responsible for deploying multimodal large models in relevant products to enhance competitiveness and support business success.
Responsible for continuously tracking the latest technological advancements in multimodal large models from both industry and academia, and planning technical roadmaps based on product needs.
Responsible for strengthening the algorithmic capabilities of multimodal engines by improving core competencies in algorithm theory, pre-trained models, and multi-source feature representation.
Requirements & Skills:
Master’s or PhD Degree in Computer Science or related fields such as Mathematics, Computer Engineering.
5+ years of work experience in the related fields such as LLM, Machine Learning, and NLP.
Familiar with the technological evolution and current industry applications of multimodal large models.
High-quality research publications and high-impact open-source projects are assets.
Solid understanding of both computer vision and natural language processing, with the ability to design multimodal solutions that integrate multiple modalities.
Proficient in designing AI solutions for multimodal scenarios and capable of algorithm development and implementation.