-
Notifications
You must be signed in to change notification settings - Fork 58
Issues: triton-inference-server/onnxruntime_backend
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
In Dockerfile gen script, CUDNN_VERSION should be obtained from docker image
#52
opened Jul 13, 2021 by
GuanLuo
Expose all string key/value configs instead of doing it piecemeal.
enhancement
New feature or request
#107
opened Mar 17, 2022 by
pranavsharma
Improve autocomplete to make it more robust against partial model configuration
#113
opened Apr 20, 2022 by
tanmayv25
default-max-batch-size doesn't cooperate well with preferred_batch_size
#148
opened Sep 28, 2022 by
OvervCW
GPT2 performance degradation with higher sequence length on ONNX Runtime
#157
opened Nov 9, 2022 by
rgallardone
Possible to enable dynamic batch dimension only on one some input tensors?
#165
opened Dec 30, 2022 by
kgu3
Can I build the Onnxruntime backend for Windows without Docker??
#175
opened Mar 15, 2023 by
victorsoyvictor
InvalidArgumentError: The tensor Input (Input) of Slice op is not initialized.
#191
opened May 25, 2023 by
qiu-pinggaizi
Add
enable_dynamic_shapes
To Model Config To Resolve CNN Memory Leaks With OpenVino EP
#194
opened Jun 2, 2023 by
narolski
Previous Next
ProTip!
Follow long discussions with comments:>50.