This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
The Microsoft AI Superintelligence Post Training team is dedicated to advancing post-training methods for both OpenAI and open-source models. Their work encompasses continual pre-training, large-scale deep reinforcement learning running on extensive GPU resources, and significant efforts to curate and synthesize training data. In addition, the team employs various fine-tuning approaches to support both research and product development. The team also develops advanced AI technologies that integrate language and multi-modality for a range of Microsoft products. The team is particularly active in developing code-specific models, including those used in Github Copilot and Visual Studio Code, such as code completion model and the software engineering (SWE) agent models. The team has also produced publications as by-products, including work such as LoRA, DeBerTa, Oscar, Rho-1, Florence, and the open-source Phi models.
Job Responsibility:
Perform large-scale model training - Especially with LLMs, SLMs, multimodal, or code-specific models
Perform data curation and synthesis - Creating and refining datasets to optimize training outcomes
Hands-on coding- write efficient, production-quality code and debug complex training jobs
Work on both proprietary and open-source frameworks - Demonstrated proficiency in training pipelines and architecture
Full-stack modeling responsibility - From data ingestion and training to evaluation and inference management
Contribute to or build on existing innovations like technical report of the well-known models
Develop novel AI solutions that bridge language, vision, and code understanding
Help develop models powering tools like GitHub Copilot, Cursor, and VS Code suggestions
Embody our Culture and Values
Requirements:
Doctorate in relevant field AND 3+ years related research experience OR Master's Degree in relevant field AND 4+ years related research experience OR Bachelor's Degree in relevant field AND 6+ years related research experience OR equivalent experience
5+ years of coding experience in Python and experience with ML frameworks such as PyTorch and Triton
3+ years of experience in data curation and synthesis, creating and refining datasets to optimize training outcomes
3+ years of proven ability to design and scale training infrastructure and pipelines in production environments
3+ years of large-scale model training - especially with LLMs, SLMs, multimodal, or code-specific models
Prior research publication record with over 3000 citations
Ability to meet Microsoft, customer and/or government security screening requirements are required for this role. These requirements include, but are not limited to the following specialized security screenings: Microsoft Cloud Background Check: This position will be required to pass the Microsoft Cloud background check upon hire/transfer and every two years thereafter.
Nice to have:
Agile, solution-oriented, and able to operate with minimal overhead within a startup style mindset
Self-driven and organized with the ability to take ownership of projects and document findings clearly and effectively
Proven track record of impactful research - Preferably at leading research labs, with published work or real-world deployments
Extensive experience with foundation models, including large-scale training, model inference, reinforcement learning, reasoning models, vision-language integration, and audio-visual modeling
Hands-on experience with large-scale distributed training or serving, and systems of thinking
Experience working with large, complex datasets and developing data pipelines for LLM training
Demonstrated ability to collaborate within interdisciplinary teams and communicate complex, multimodal research concepts effectively