-
Notifications
You must be signed in to change notification settings - Fork 11.6k
Error converting new stablelm-2-12b-chat #6553
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
The 12b parameter model is not supported yet, I'll take a look later today. |
It looks like they are using a per head layer norm in their implementation. It's not supported in llama.cpp afaik. I'm not planning to implement it, since the model doesn't seem that good. |
Got exactly the same error, cannot make GGUF from it (using convert-hf-to-gguf.py).
@Galunid from https://stability.ai/news/introducing-stable-lm-2-12b - it doesn't look bad for its weight class: |
This may be the same that command r plus does. llama.cpp supports it, but you have to be careful to reshape q/k to 3d before doing the norm, and to export the norm as f32. |
+1, same issue here, just doing exactly the same conversion attempt. Edit : |
@bartowski1182 @IzzyHibbert I added a PR with a working solution: #6635. I've tested with 12B and 12B chat. Feel free to try, not you have to use this branch of the model for now: https://huggingface.co/stabilityai/stablelm-2-12b/tree/stack-per-head-qk-norm @ggerganov Can you please assign this issue to me so I can track, thanks! |
closed in #6635 |
Using version b2589
Attempting convert-hf-to-gguf.py on
https://huggingface.co/stabilityai/stablelm-2-12b-chat
Results in error:
Can not map tensor 'model.layers.0.self_attn.k_layernorm.norms.0.weight'
The text was updated successfully, but these errors were encountered: