-
Notifications
You must be signed in to change notification settings - Fork 3.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Crash with multiple whisper states running at the same time CUDA #2177
Comments
Doesn't look like it. I downloaded the master branch and used that... [Current thread is 1 (Thread 0x7fef323f1000 (LWP 3420))] |
It's not merged, so you need to use the |
My mistake. The branch is good. No crash. |
I didn't have this issue with 1.5.5 but with 1.6.1 (haven't tried 1.6.0), running multiple whisper_full_with_state ends up with some issues in the cuda back end with freeing up memory?
Relevant part of the stack trace below. I was testing the library with CUDA 12.4 with a new L4 card. Was previously testing with a T4 card. I don't know if that bares any relevance however. I haven't tried 1.6.1 on the T4 test build. Running a single thread has no issue.
#0 0x00007f85d6f88b8f in raise () from /lib64/libc.so.6
#1 0x00007f85d6f5bea5 in abort () from /lib64/libc.so.6
#2 0x00007f8600ad947a in ggml_cuda_pool_vmm::free(void*, unsigned long) () from /opt1/resource/lib/libwhisper.so
#3 0x00007f8600ad1f00 in ggml_cuda_op_mul_mat_cublas(ggml_backend_cuda_context&, ggml_tensor const*, ggml_tensor const*, ggml_tensor*, char const*, float const*, char const*, float*, long, long, long, long, CUstream_st*) () from /opt1/resource/lib/libwhisper.so
#4 0x00007f8600ad4bdc in ggml_cuda_op_mul_mat(ggml_backend_cuda_context&, ggml_tensor const*, ggml_tensor const*, ggml_tensor*, void ()(ggml_backend_cuda_context&, ggml_tensor const, ggml_tensor const*, ggml_tensor*, char const*, float const*, char const*, float*, long, long, long, long, CUstream_st*), bool) ()
from /opt1/resource/lib/libwhisper.so
#5 0x00007f8600ad5c79 in ggml_cuda_mul_mat(ggml_backend_cuda_context&, ggml_tensor const*, ggml_tensor const*, ggml_tensor*) () from /opt1/resource/lib/libwhisper.so
#6 0x00007f8600ad7eee in ggml_backend_cuda_graph_compute(ggml_backend*, ggml_cgraph*) () from /opt1/resource/lib/libwhisper.so
#7 0x00007f8600bd2b39 in ggml_backend_graph_compute () from /opt1/resource/lib/libwhisper.so
#8 0x00007f8600c217e1 in whisper_encode_internal(whisper_context&, whisper_state&, int, int, bool ()(void), void*) () from /opt1/resource/lib/libwhisper.so
#9 0x00007f8600c2194f in whisper_encode_with_state () from /opt1/resource/lib/libwhisper.so
#10 0x00007f8600c267f3 in whisper_lang_auto_detect_with_state () from /opt1/resource/lib/libwhisper.so
#11 0x00007f8600c345d5 in whisper_full_with_state () from /opt1/resource/lib/libwhisper.so
The text was updated successfully, but these errors were encountered: