-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DeepSeek-R1-Distill-Qwen-32B-GGUF needs the deepseek-r1-qwen tokenizer #1900
Comments
I guess it can be solved by simply change a commit id? Check c9dfad4. |
Then we will need to wait for updates. I guess it won't be too hard to match up api, perhaps we can just pass the code to GPT and expect it to solve the issue. Would make a try if I get some time (busy week). |
Thank you! |
It is not just a commit synchronization issue. There are many LLAMA_API parts in llama_cpp.py that need to be updated and synchronized to the new version refactored in llama.h of llama.cpp master. |
I am eager to receive the update as well. The Deepseek R1 model is truly remarkable. Hope this adaptation of API could be more automatized through AI coder in future |
does llama.cpp have such problem? |
I guess using llama.cpp directly works fine as bartowsky did the quant with it, check the model url. |
"Deepseek R1 is one of the most amazing and impressive breakthroughs I’ve ever seen — and as open source, a profound gift to the world." - Marc Andreessen Is it a big ask to add support for this model? |
Current Behavior
Can't run DeepSeek-R1-Distill-Qwen-32B-GGUF, I get a tokenizer error:
"llama_model_load: error loading model: error loading model vocabulary: unknown pre-tokenizer type: 'deepseek-r1-qwen'"
Model url: https://huggingface.co/bartowski/DeepSeek-R1-Distill-Qwen-32B-GGUF
Environment and Context
Hardware: VPS with 32 GB RAM
OS: Debian 12
Using the current version of llama-cpp-python
Suggestion
Upgrade to the required version of llama.cpp:
https://github.com/ggerganov/llama.cpp/releases/tag/b4514
The text was updated successfully, but these errors were encountered: