-
Notifications
You must be signed in to change notification settings - Fork 1.1k
_create_completion cache_item.input_ids.tolist() - AttributeError: 'NoneType' object has no attribute 'input_ids' #348
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Sadly, there's no obvious error there that can point to where the issue is within Any chance you can reproduce the issue with a an identical |
It seems the actual error was cut out of my output provided, please give me a moment to capture the issue |
@gjmulder hopefully this output is more helpful
The error seems to reference: https://github.com/abetlen/llama-cpp-python/blob/main/llama_cpp/llama.py#LL814C18-L814C18 |
Yup, looks like a bug with the new diskcache implementation, working on a fix. |
Thank you @abetlen Much respect :) |
@snxraven after resolving the issue with the cache (LlamaCache was changed to an abstract base class, I changed this back to point to LlamaRAMCache to avoid breaking changes) it looks like there might be a bug with the |
@abetlen Thanks for the heads up, ill stay updated here and within the repo for updates. Good looking out |
Is this the related issue? I applied the fix at the bottom of the thread, and the cache seems to be working. |
@AlphaAtlas yes that looks like the issue. Did you also remove this line? |
Yeah, I removed all three |
The cache issue is fixed upstream: ggml-org/llama.cpp#1699 (comment) |
@AlphaAtlas @snxraven should be fixed now in v0.1.62 |
@abetlen looks perfect :) All seems well you can most likely close this |
…n#348) * cmdline option for custom amount of model parts (--n_parts N) * Update main.cpp --------- Co-authored-by: Georgi Gerganov <[email protected]>
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
The chat completion endpoint should be processing chat and responding accordingly
Current Behavior
I see the following crash:
Environment and Context
This is a docker-compose project, feel free to check how I have the installation configured:
https://git.ssh.surf/snxraven/llama-cpp-python-djs-bot
The text was updated successfully, but these errors were encountered: