-
Notifications
You must be signed in to change notification settings - Fork 43
Upgrade llama.cpp to b3265, support gemma2, remove beam parameter[ https://github.com/ggerganov/llama.cpp/pull/7985 ] #71
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Hey @zdl010 thanks for the pull request! I'll update the C++ code to b3265 and merge after that. |
@kherud I see there's a lot of update in server.cpp in llama.cpp. Is that would take long to adjust in server.hpp? |
Hey @ardinursyamsu yeah it's a challenge to keep up with the rapid development of llama.cpp. Sometimes there are bugs, where it isn't obvious if they come from llama.cpp or the Java binding. I'll have another try to update to the latest version later today, though. |
Ok, there was a change in llama.cpp to no longer statically link the ggml library (see ggml-org/llama.cpp#8166), which caused the previous Windows builds here to fail. Not sure why it worked for Linux/MacOS (probably because of rpath). I'll look for a solution tomorrow and release a new version then. |
Upgrade llama.cpp to b3265, support gemma2, remove beam parameter[ https://github.com/ggml-org/llama.cpp/pull/7985 ]