You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
But some how the output shape after pruning had different size compare to the target:
### The attention Q, K, V shape mismatch
size mismatch for model.embed_tokens.weight: copying a param with shape torch.Size([32000, 2560]) from checkpoint, the shape in
current model is torch.Size([32000, 2048]).
size mismatch for model.layers.0.self_attn.q_proj.weight: copying a param with shape torch.Size([2048, 2560]) from checkpoint,
the shape in current model is torch.Size([2048, 2048]).
size mismatch for model.layers.0.self_attn.k_proj.weight: copying a param with shape torch.Size([2048, 2560]) from checkpoint,
the shape in current model is torch.Size([2048, 2048]).
....
Is there any problem in prune_params() methods when apply l0_module?
The text was updated successfully, but these errors were encountered:
I had used your sample scripts to pruned
princeton-nlp/Sheared-LLaMA-2.7B
into 1.3B size. This is my model config trainingNote that i just use
github
proportion to check the pipeline is completed or not. After pruning, I convert model by this scriptsBut some how the output shape after pruning had different size compare to the target:
Is there any problem in
prune_params()
methods when applyl0_module
?The text was updated successfully, but these errors were encountered: