Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Segmentation Fault with 7b on Raspberry Pi 3 #213

Open
leafyus opened this issue Apr 10, 2023 · 5 comments
Open

Segmentation Fault with 7b on Raspberry Pi 3 #213

leafyus opened this issue Apr 10, 2023 · 5 comments

Comments

@leafyus
Copy link

leafyus commented Apr 10, 2023

leafy@raspberrypi:~/alpaca.cpp $ ./chat
main: seed = 1681116282
llama_model_load: loading model from 'ggml-alpaca-7b-q4.bin' - please wait ...
llama_model_load: ggml ctx size = 6065.34 MB
Segmentation fault

Tried to run the 7b Alpaca Model on my Raspberry Pi 3, but getting a Segmentation Fault every time. Compiled from source code. The RP3 has 4 GBs of RAM, is that the problem?

@themanyone
Copy link

themanyone commented Apr 10, 2023

Same here, segmentation fault, but on an old Linux x86_64 elitebook laptop. In this case it was -mavx causing the error.
This compiles, but it runs pretty slow without mavx: ggerganov#107

gcc -I. -O3 -DNDEBUG -std=c11 -fPIC -pthread -msse3 -c ggml.c -o ggml.o
g++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread chat.cpp ggml.o utils.o -o chat
g++ -I. -I./examples -O3 -DNDEBUG -std=c++11 -fPIC -pthread quantize.cpp ggml.o utils.o -o quantize

@BlueveryPi
Copy link

same problem here. working with a aws ec2. having the problem even without building from source. The file size is 4.0GB? may be the problem...

@yaozakai
Copy link

Hello, I also have the same issue with a Raspberry Pi 4, 4GB

I'm not sure the file size is the issue, as this blog got it working for a RPi 5 also with the same 4GB of RAM.

@themanyone
Copy link

Might want to trim down any running processes, also. Running non-graphical session, perhaps?

@yaozakai
Copy link

Yeah so I run headless. I ended up using a 2-bit model instead that was 2+GB in size and it seemed to work.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants