-
Notifications
You must be signed in to change notification settings - Fork 106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama3 reports shape error after pruning #69
Comments
+1 |
+1 got same error |
1 similar comment
+1 got same error |
Has anyone solved this problem? |
The line attn_output = attn_output.reshape(bsz, q_len, self.hidden_size) can be modified to attn_output = attn_output.reshape(bsz, q_len, -1) to resolve the issue. |
got same error |
I tried this method, but it doesn't work. In the modeling_llama.py file I downloaded, the source code is: attn_output = attn_output.reshape(bsz, q_len, -1) |
The command I run:
'''
python llama3.py --pruning_ratio 0.25
--device cuda --eval_device cuda
--base_model home/Meta-Llama-3-8B
--block_wise --block_mlp_layer_start 4 --block_mlp_layer_end 30
--block_attention_layer_start 4 --block_attention_layer_end 30
--save_ckpt_log_name llama3_prune
--pruner_type taylor --taylor param_first
--max_seq_len 2048
--test_after_train --test_before_train --save_model
'''
When running to line 259 of the code, an error occurs:
How to solve this problem? Looking forward to your reply!
The text was updated successfully, but these errors were encountered: