-
Notifications
You must be signed in to change notification settings - Fork 38
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to Reproduce ALMA-7b-LoRA Performance, Seeking Assistance #58
Comments
Thanks for your interest! Could you please provide the |
It looks like the results you got are very close to the checkpoint we released under the same virtual env. I suspect the main issue could come from the version mismatch. Please try uninstall
Hope this is helpful |
Thank you for your suggestion! I will try it out and provide feedback in this issue. |
By the way, could you please provide the versions of the |
Thank you for your excellent work.
While fine-tuning the ALMA-7b-Pretrain model and testing with the checkpoint you provided, I was unable to reproduce the performance of ALMA-7b-LoRA as described in the paper. I would appreciate any guidance or suggestions you could offer.
I used the code, data, and scripts provided in this repository (including runs/parallel_ft_lora.sh and evals/alma_7b_lora.sh), with a training batch size of 256 and four V100 GPUs.
Please feel free to ask if you need more details about my experiments.
The text was updated successfully, but these errors were encountered: