Director Software Development - AI Inference at Advanced Micro Devices Inc
Bellevue, WA 98007, USA -
Full Time


Start Date

Immediate

Expiry Date

21 Jul, 25

Salary

0.0

Posted On

21 Apr, 25

Experience

0 year(s) or above

Remote Job

Yes

Telecommute

Yes

Sponsor Visa

No

Skills

Leadership, Software Development, Benchmarking, Nlp, Strategy, Software, Collaboration, Communication Skills, Computer Vision, Tuning, Triton

Industry

Information Technology/IT

Description

PREFERRED EXPERIENCE:

  • Leadership: AI/ML software development in leadership roles managing AI model enablement or optimization teams.
  • AI Inference Optimization: Expertise in optimizing real-time AI models for deep learning applications (computer vision, NLP, etc.).
  • AI Frameworks: Proficiency with AI frameworks (TensorFlow, PyTorch, ONNX Runtime, JAX, Triton) and their optimization for GPU architectures.
  • GPU Optimization: Strong background in optimizing software for AMD GPUs or similar high-performance platforms. Familiarity with ROCm is a plus.
  • Benchmarking & Performance Tuning: Proven experience with performance optimization, benchmarking, and scaling AI models on GPUs.
  • Collaboration & Strategy: Exceptional ability to collaborate cross-functionally and define long-term strategies for AI/ML innovation.
  • Communication: Strong verbal and written communication skills, with experience presenting to senior leadership and working with customers and partners.
Responsibilities

WHAT YOU DO AT AMD CHANGES EVERYTHING

We care deeply about transforming lives with AMD technology to enrich our industry, our communities, and the world. Our mission is to build great products that accelerate next-generation computing experiences – the building blocks for the data center, artificial intelligence, PCs, gaming and embedded. Underpinning our mission is the AMD culture. We push the limits of innovation to solve the world’s most important challenges. We strive for execution excellence while being direct, humble, collaborative, and inclusive of diverse perspectives.
AMD together we advance_
Responsibilities:

THE ROLE:

At AMD, we are enabling the next generation of AI innovation by leveraging the power of our GPU architecture to accelerate AI workloads. As part of the AI/ML Framework Department, you will lead multiple teams focused on the enablement, optimization, and integration of AI inference models across AMD’s GPU platform. This critical leadership role is responsible for ensuring that AI inference models run efficiently and at scale on AMD GPUs, driving high performance across training and real-time inference tasks.
In this role, you will oversee the optimization of AI Inference teams ensuring seamless compatibility with AMD’s ROCm platform. You will work closely with AMD’s hardware and software teams to deliver best-in-class AI solutions, ensuring that AMD’s GPUs continue to provide outstanding performance for industry-leading AI Inference Software Development.

KEY RESPONSIBILITIES:

  • Leadership & Team Management: Lead and develop teams responsible for AI inference model enablement and optimization, driving growth and success within the department.
  • AI Inference Optimization: Direct efforts to optimize AI frameworks for seamless compatibility and performance on AMD GPUs (Instinct, Navi).
  • Model Tuning & Performance: Oversee benchmarking, performance tuning, and optimization of AI inference models to improve latency, throughput, and efficiency on AMD hardware.
  • Cross-functional Collaboration: Partner with hardware, software, and QA teams to ensure tight integration of AI frameworks with ROCm for maximum performance.
  • Innovation & Strategy: Drive AI model optimization innovations, enhancing the speed, efficiency, and scalability of AI workloads. Lead the vision and strategy for optimizing AI inference on AMD GPUs.
  • Customer & Community Engagement: Collaborate with customers and open-source communities to ensure that AMD’s AI solutions meet industry needs, fostering contributions to MIGraphX, vLLM, and other AMD AI Framework Inference teams.
  • Automation & Scaling: Oversee automation frameworks to streamline model integration and performance testing, ensuring scalability across diverse AI workloads.
Loading...