This list contains only the countries for which job offers have been published in the selected language (e.g., in the French version, only job offers written in French are displayed, and in the English version, only those in English).
At AMD, we are enabling the next generation of AI innovation by leveraging the power of our GPU architecture to accelerate AI workloads. As part of the AI/ML Framework Department, you will lead multiple teams focused on the enablement, optimization, and integration of AI inference models across AMD’s GPU platform. This critical leadership role is responsible for ensuring that AI inference models run efficiently and at scale on AMD GPUs, driving high performance across training and real-time inference tasks. In this role, you will oversee the optimization of AI Inference teams ensuring seamless compatibility with AMD's ROCm platform. You will work closely with AMD’s hardware and software teams to deliver best-in-class AI solutions, ensuring that AMD’s GPUs continue to provide outstanding performance for industry-leading AI Inference Software Development.
Job Responsibility:
Lead and develop teams responsible for AI inference model enablement and optimization
Direct efforts to optimize AI frameworks for seamless compatibility and performance on AMD GPUs (Instinct, Navi)
Oversee benchmarking, performance tuning, and optimization of AI inference models to improve latency, throughput, and efficiency on AMD hardware
Partner with hardware, software, and QA teams to ensure tight integration of AI frameworks with ROCm for maximum performance
Drive AI model optimization innovations, enhancing the speed, efficiency, and scalability of AI workloads
Lead the vision and strategy for optimizing AI inference on AMD GPUs
Collaborate with customers and open-source communities to ensure that AMD’s AI solutions meet industry needs, fostering contributions to MIGraphX, vLLM, and other AMD AI Framework Inference teams
Oversee automation frameworks to streamline model integration and performance testing, ensuring scalability across diverse AI workloads
Requirements:
10+ years in AI/ML software development
5+ years in leadership roles managing AI model enablement or optimization teams
Expertise in optimizing real-time AI models for deep learning applications (computer vision, NLP, etc.)
Proficiency with AI frameworks (TensorFlow, PyTorch, ONNX Runtime, JAX, Triton) and their optimization for GPU architectures
Strong background in optimizing software for AMD GPUs or similar high-performance platforms
Familiarity with ROCm is a plus
Proven experience with performance optimization, benchmarking, and scaling AI models on GPUs
Exceptional ability to collaborate cross-functionally and define long-term strategies for AI/ML innovation
Strong verbal and written communication skills, with experience presenting to senior leadership and working with customers and partners
Advanced degree (Master’s or PhD) in Computer Science, Electrical Engineering, AI/ML, or related field
Contributions to open-source AI projects or community-driven initiatives are a plus
Nice to have:
Contributions to open-source AI projects or community-driven initiatives are a plus