-
Notifications
You must be signed in to change notification settings - Fork 907
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segmentation fault (only) with 13B model. #45
Comments
I have the same result, I also ran it under WSL on windows, works with 7B model, not with 13B model. Same md5sum. Same result btw for ggerganov/llama.cpp, from which this project is forked. It gives a more detailed error message:
I did not use the 7B model from the torrent, but from the download url in this repo, did you do the same? Perhaps the 13B model has to be transformed to an appropriate format before it can be used in this project? |
Yes, I downloaded the 7B file from the direct link on the readme. Also used the magnet link on the readme for the 13B file. Haven’t seen anything about converting the 13B downloaded file. |
It has nothing to do with converting. Main.cpp thinks this is a multi-part file. Usually the 13B model is splitted into two files. But here, we have only one file. |
Make sure to compile it again, using the latest version of the source code |
@PriNova I tried that, changed the equivalent line on chat.cpp and compiled again, unfortunately didn't help :( @antimatter15 I compiled the latest master, unfortunately didn't help :( |
Well, just by doing some basic
No idea how to proceed from here though :( |
i used this |
Try to provide a full path to model like ./chat -m D:/alpaca/13b/ggml-alpaca-13b-q4.bin |
In my case, the ggml-alpaca-13b-q4.bin works. I referred to this tweet. This includes several corrections in chat.cpp |
Changing |
How can you change that line if there isn't a main.cpp file in the alpaca.cpp folder |
It is already fixed in the chat.cpp file at line 34 in the model paramteres. |
I fixed this issue by troubleshooting on my own machine! I had the same issue running on WSL. The segmentation fault is due to not enough RAM. I have 32GB RAM and was able to run up to the 13B model but not the 30B under WSL. I tried building then running it under Windows and not WSL and it worked. Seems like WSL cannot use the page file properly at least for this project. 30B is slow but works for me now! |
You can increase your page file globally or locally. https://learn.microsoft.com/en-us/windows/wsl/wsl-config#configuration-setting-for-wslconfig |
Good to know you can do it for WSL! I found it easier to just do it on Windows and compile/run it outside of WSL. |
This should add a better error handling when memory buffer allocation fails. |
hi, where did you get the ggml-alpaca-13b-q4.bin download file from? I can't seem to find it anywhere to download. |
I just downloaded the 13B model from the torrent (
ggml-alpaca-13b-q4.bin
), pulled the latest master and compiled. It works absolutely fine with the 7B model, but I just get the Segmentation fault with 13B model.Checksum of the 13B model;
66f3554e700bd06104a4a5753e5f3b5b
I'm running Ubuntu under WSL on Windows.
The text was updated successfully, but these errors were encountered: