Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Allow Usage of Intel oneDNN EP For ONNX Backend #133

Open
narolski opened this issue Jul 27, 2022 · 2 comments
Open

Allow Usage of Intel oneDNN EP For ONNX Backend #133

narolski opened this issue Jul 27, 2022 · 2 comments

Comments

@narolski
Copy link

narolski commented Jul 27, 2022

Is your feature request related to a problem? Please describe.
I would like to use the Intel oneDNN Execution Provider (EP) in ONNX Runtime built for Triton Inference Server ONNX Backend.

Describe the solution you'd like
Ideally, the oneDNN EP should be enabled the same way we can enable the usage of OpenVino EP in model configuration:

optimization {
  execution_accelerators {
    cpu_execution_accelerator : [ {
      name : "openvino"
    } ]
  }
}

Describe alternatives you've considered
I've tried to pass dnnl under cpu_execution_accelerator, but this is not supported.

oneDNN might yield greater performance improvements for CPU inference than OpenVino, that is why it would be great to be able to use it within the Triton Inference Server.

Update: Furthermore, it seems that onednn is enabled by default for ONNX Runtime wheel built with onednn over the default ONNX Runtime CPU Execution Provider:

When using the python wheel from the ONNX Runtime built with DNNL execution provider, it will be automatically prioritized over the CPU execution provider. Python APIs details are here.

Additional context
ONNX Runtime documentation: https://fs-eire.github.io/onnxruntime/docs/execution-providers/oneDNN-ExecutionProvider.html

@narolski
Copy link
Author

@pranavsharma Do you think it will be possible to implement this configuration option?

@pranavsharma
Copy link
Contributor

We've not planned for it yet. Would you like to contribute?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants