This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
We internally call this team MBMB (More Big More Better). You will own optimizations on both the training and on-robot inference stacks. We are still in a regime of step-function, not incremental, gains.
Job Responsibility:
Making GPUs go brrrrr
Implementing ML, hardware, and software changes that lead to step-function gains
Optimizing both the inference and training stacks
Requirements:
Proficient and stay current with the latest ML techniques for training and inference optimizations in transformer and diffusion based architectures
Will chase ML optimizations anywhere: From the CUDA kernels, to ML architecture, to frontend or backend network bottlenecks, CPU bottlenecks, NVLink and comms, to torch, numpy, and Python inefficiencies