-
Notifications
You must be signed in to change notification settings - Fork 12.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG]Load model #542
Comments
The errors you’re seeing occur because the model’s configuration isn’t aligned with what your current tooling (either VLLM or Hugging Face’s Transformers with quantization support) expects. Here’s what’s happening and some suggestions on how to proceed:
DeepseekForCausalLM How to solve it: Check for a Supported Version: If you’re tied to DeepSeek models, see if there’s a version of the model (for example, one with the architecture DeepseekForCausalLM or DeepseekV2ForCausalLM) that is officially supported by VLLM.
ValueError: Unknown quantization type, got fp8 - supported types are: ['awq', 'bitsandbytes_4bit', 'bitsandbytes_8bit', 'gptq', ...] How to solve it: Modify the Configuration: Wait for Official Support: Consult Documentation/Community: |
I got below error when I try to laod deepseek model for testing: Traceback (most recent call last): def download_file(url, save_path): def download_deepseek_model_files(): def generate_poem():
if name == "main": |
Same here. Use AutoModelForCausalLM to load model and get the error. Did you fix it? |
When i try to load deepseek model:
Attemp 1:
ERROR 1: ValidationError: 1 validation error for VLLM
Attemp 2:
ERROR 2:
Why it happends and how to solve
The text was updated successfully, but these errors were encountered: