-
Notifications
You must be signed in to change notification settings - Fork 43
Java tests failed when CUDA enabled on version 3.0.0 #54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. Weβll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Here is the log file: |
Damn, I didn't test thoroughly enough with CUDA, but I can reproduce the problems, thanks for reporting. It seems to be related to java-llama.cpp/src/main/cpp/server.hpp Line 2266 in 6d500b5
|
It turns out this is a bug in llama.cpp after all and I've created an issue there (see ggml-org/llama.cpp#6672) It didn't produce a crash for you because the The problem only seems to occur with models that don't support infilling, which unfortunately is the case for the model used in the unit tests. However, everything works correctly with models that support infilling (e.g. codellama). |
I changed the model that is used for testing to codellama, so there shouldn't be a segmentation fault anymore. However, I'm still leaving this issue open until the underlying issue is fixed within llama.cpp. |
I think I've diagnosed the issue & pointed to the tag that fixed it in the related thread. ggml-org/llama.cpp#6672 |
Hello!
I really appreciate that you have upgraded this project!
However, there are still 2 tests that cannot pass.
testGenerateInfill
andtestCompleteInfillCustom
. The outputs would be something like this:I have built with the command
cmake .. -DBUILD_SHARED_LIBS=ON -DLLAMA_CUDA=ON -DLLAMA_CURL=ON
.Also, I have tested vanilla llama.cpp of tag b2619, with the same build args above and the same inference args (shown below), and it worked without crash:
and
Anyway, other java tests have been passed.
Thanks!
The text was updated successfully, but these errors were encountered: