-
Notifications
You must be signed in to change notification settings - Fork 237
Closed
Description
Problem Statement:
- Many users possess AMD or Intel hardware without NVIDIA GPUs.
- Current ONNX Runtime acceleration is limited to CUDA/DirectML.
- Lack of AMD/Intel support forces fallback to CPU inference—higher latency and lower throughput.
Proposal:
-
Integrate
onnxruntime-rocm- Enable ROCm EP for AMD GPUs via ONNX Runtime
- Optionally include MIGraphX EP for deep-graph optimization (ONNX Runtime
-
Integrate
onnxruntime-openvino- Leverage OpenVINO EP to offload inference to Intel CPUs, integrated/discrete GPUs, and NPUs (ONNX Runtime, OpenVINO Documentation)
-
Update build scripts & documentation
- Add EP-specific install instructions (pip wheels, environment setup)
- Provide sample configs for AMD and Intel pipelines
Benefits:
-
Reduced Latency
- Hardware acceleration on AMD GPUs (ROCm) and Intel NPUs (OpenVINO) cuts inference time by 2–10× compared to CPU only.
-
Broader Hardware Coverage
- Supports a wider user base: gaming PCs (AMD/Intel), AI PCs with AMD Strix Point/Gorgon Point/Strix Halo APUs/NPUs and Intel Core Ultra NPUs.
References:
- ROCm Execution Provider (AMD) — ONNX Runtime
- MIGraphX Execution Provider (AMD) — ONNX Runtime
- OpenVINO Execution Provider (Intel) — ONNX Runtime
- Intel NPU Support — OpenVINO Documentation
Please consider this enhancement to unlock the full potential of local hardware acceleration across AMD and Intel platforms!
davidhop11, earendildev, Lustrewilkin, Halldrix, clemlesne and 1 more
Metadata
Metadata
Assignees
Labels
No labels