This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
Meta's Training and Inference Accelerators (MTIA) team is developing novel HW to enable efficient execution of AI training and inference workloads. In this role, you will have end-to-end responsibility for the performance of in-production AI models in their transition from stock HW to MTIA chips, with a focus on models that require multi-node compute.
Job Responsibility:
Identifying bottlenecks and quantifying opportunities for improving performance
In-depth, end-to-end performance analysis and reporting
Developing optimizations to address identified bottlenecks
Optimizing compute/communication overlap
Work closely with other compiler teams as well as client teams (Recommendation Systems, Generative AI, etc)
Requirements:
Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience
Experience developing and deploying optimizations at the level of PyTorch/Aten or comparable stacks
A Masters degree and 4+ years in-domain experience