This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
At Luma, the Pre-Training / Scaling team is responsible for building the core multimodal AI systems that power our entire platform. Working at the forefront of generative AI research, this team develops the fundamental architectures and training methodologies that enable our models to see, hear, understand, and interact with the world across video, image, text, and audio modalities.
Job Responsibility:
Lead cutting-edge research in multimodal foundation models spanning video, image, text, and audio
Design and implement novel algorithms, architectures, and techniques for large-scale generative AI models
Develop training methodologies for foundation models across thousands of GPUs
Research and implement state-of-the-art techniques in Autoregressive LLMs, Vision Language Models, and / or Diffusion Models
Collaborate with cross-functional teams to transition research into production systems
Requirements:
Expertise in Python and PyTorch with experience building ML models from scratch
Deep understanding of multimodal generative models and deep learning architectures
(Preferred) Strong research track record in generative AI with published work in top-tier venues preferred
(Preferred) Experience with large-scale distributed training systems