We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hello, may I ask why the performance of my pruned model has decreased after lora fine-tuning with the following parameters
/h3cstore_ns/jcxie/condaenv/bin/python /h3cstore_ns/jcxie/LISA/wanda-main/lora_ft/finetune_lm.py --model_name_or_path /h3cstore_ns/jcxie/LISA/wanda-main/ckpt/lm7b --config_name "/h3cstore_ns/jcxie/hf_weights/llama-2-7b-hf" --dataset_name c4 --num_train_epochs 2 --block_size 1024 --per_device_train_batch_size 8 --per_device_eval_batch_size 8 --do_train --do_eval --max_train_samples 30000 --max_eval_samples 128 --learning_rate 1e-4 --overwrite_output_dir --output_dir /h3cstore_ns/jcxie/LISA/wanda-main/ckpt/lm7b_lora &&
wait
I used 8 * 3090 and DDP for fine-tuning
Performance after only pruning: perplexity on wikitext 6.94
Performance of fine-tuning after pruning: 10.3
The text was updated successfully, but these errors were encountered:
No branches or pull requests
Hello, may I ask why the performance of my pruned model has decreased after lora fine-tuning with the following parameters
/h3cstore_ns/jcxie/condaenv/bin/python /h3cstore_ns/jcxie/LISA/wanda-main/lora_ft/finetune_lm.py
--model_name_or_path /h3cstore_ns/jcxie/LISA/wanda-main/ckpt/lm7b
--config_name "/h3cstore_ns/jcxie/hf_weights/llama-2-7b-hf"
--dataset_name c4
--num_train_epochs 2
--block_size 1024
--per_device_train_batch_size 8
--per_device_eval_batch_size 8
--do_train
--do_eval
--max_train_samples 30000
--max_eval_samples 128
--learning_rate 1e-4
--overwrite_output_dir
--output_dir /h3cstore_ns/jcxie/LISA/wanda-main/ckpt/lm7b_lora &&
wait
I used 8 * 3090 and DDP for fine-tuning
Performance after only pruning: perplexity on wikitext 6.94
Performance of fine-tuning after pruning: 10.3
The text was updated successfully, but these errors were encountered: