Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cannot load vicuna-7b-delta-v0 #14

Closed
LianxinGao opened this issue Aug 29, 2023 · 5 comments
Closed

Cannot load vicuna-7b-delta-v0 #14

LianxinGao opened this issue Aug 29, 2023 · 5 comments

Comments

@LianxinGao
Copy link
Contributor

  1. Use aspen.load_llama_tf_weight to load vicuna-7b-delta-v0 model use more than 30GB memory then caused OOM.

  2. Use utils.convert_hf_to_pth to transfer vicuna-7b-delta-v0 to .pth model, then use aspen.load_llama_7b_weight to load .pth model, an error is reported:

Not use layer model.embed_tokens.weight.
Traceback (most recent call last):
  File "/data/glx/code/multi_lora/legacy.py", line 43, in <module>
    aspen.load_llama_7b_weight(llama_model, config["base_model"], config["device"])
  File "/data/glx/code/multi_lora/aspen/modelloader.py", line 21, in load_llama_7b_weight
    layer_id = int(layer_name[:layer_name.find(".")])
ValueError: invalid literal for int() with base 10: 'ayers'
@github-actions
Copy link

welcome to the Aspen' first issue

@mikecovlee
Copy link
Member

mikecovlee commented Aug 29, 2023 via email

@mikecovlee
Copy link
Member

mikecovlee commented Aug 29, 2023 via email

@mikecovlee
Copy link
Member

噢,我知道咋回事了,明天我推个patch

@mikecovlee
Copy link
Member

Plz close this issue if problem solved.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants