Skip to content

Add AMD and Intel GPU/NPU Support #86

@korgano

Description

@korgano

Problem Statement:

  • Many users possess AMD or Intel hardware without NVIDIA GPUs.
  • Current ONNX Runtime acceleration is limited to CUDA/DirectML.
  • Lack of AMD/Intel support forces fallback to CPU inference—higher latency and lower throughput.

Proposal:

  1. Integrate onnxruntime-rocm

  2. Integrate onnxruntime-openvino

  3. Update build scripts & documentation

    • Add EP-specific install instructions (pip wheels, environment setup)
    • Provide sample configs for AMD and Intel pipelines

Benefits:

  • Reduced Latency

    • Hardware acceleration on AMD GPUs (ROCm) and Intel NPUs (OpenVINO) cuts inference time by 2–10× compared to CPU only.
  • Broader Hardware Coverage

    • Supports a wider user base: gaming PCs (AMD/Intel), AI PCs with AMD Strix Point/Gorgon Point/Strix Halo APUs/NPUs and Intel Core Ultra NPUs.

References:


Please consider this enhancement to unlock the full potential of local hardware acceleration across AMD and Intel platforms!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions