-
Notifications
You must be signed in to change notification settings - Fork 84
mistral3 #1561
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
in the meantime if anyone else is looking for a gptq mistral small jeffcookio/Mistral-Small-3.1-24B-Instruct-2503-HF-gptqmodel-4b-128g worked for me |
@ewof What error do you get with Mistral3? |
|
@wemoveon2 Please check out PR/branch #1563 and recompile gptqmodel using
and check if mistra3 is fixed. |
im on transformers 4.51.3 |
@ewof Is MIstra3 a visual (hybrid) model that contains both text and visual input? |
yea and vision_config.model_type is pixtral from the models config.json |
@ewof Ugh... hybrid models needs manual quantization support as many hybrid model has standard, and non-standarard way of defining how the secondary model (multiple models inside one model config) is defined. We will try to tackle this with a manual MIstral3 support first, then create a generic code so that all future multi-modal models can work without too much integration work. Right now, the hybrid models are a pain since everyone has not yet decided on how the modeling code (preprocessing) and the forwarding hand-offs should work internally. Wild wild west. |
can someone reply here when/if mistral3 support is added (not sure how ISTA-DASLab/Mistral-Small-3.1-24B-Instruct-2503-GPTQ-4b-128g was made)
The text was updated successfully, but these errors were encountered: