-
Notifications
You must be signed in to change notification settings - Fork 98
Issues: huggingface/optimum-nvidia
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Supporting Bert / Roberta - tags: enchancement / new_model
model
#21
opened Dec 6, 2023 by
michaelfeil
Enhancing Compatibility and Extending Support for Optimum-NVIDIA Across Diverse Workloads
#34
opened Dec 11, 2023 by
yihong1120
Segmentation fault: address not mapped to object at address 0xb1fe8
#44
opened Dec 14, 2023 by
SinanAkkoyun
FileNotFoundError: [Errno 2] No such file or directory: '/data/Dilip/models/llama-2-7b-chat-hf/build.json'
bug
Something isn't working
#47
opened Dec 15, 2023 by
dilip467
Not able to run 'Generate' from QuickStart section
bug
Something isn't working
#61
opened Jan 3, 2024 by
harikrishnaapc
Error when Running LLAMA with tensor parallelism = 2
bug
Something isn't working
#68
opened Jan 23, 2024 by
TheCodeWrangler
llama.py with fp8 is broken (inference produces garbage results)
#71
opened Feb 10, 2024 by
urimerhav
How do you use the library in your scripts after pulling and running the Docker image?
#72
opened Feb 15, 2024 by
jddunn
Incorrect tensorrt_llm config class initialization
bug
Something isn't working
#90
opened Mar 7, 2024 by
Wojx
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.