-
-
Notifications
You must be signed in to change notification settings - Fork 12.5k
Closed
Labels
new-modelRequests to new modelsRequests to new models
Description
The model to consider.
vLLM supports mistral's "consolidated" format for the Pixtral model found at: https://huggingface.co/mistral-community/pixtral-12b-240910
However when HF implemented Pixtral in Transformers, they use a different format leveraging the existing Llava model structure. Model example: https://huggingface.co/mistral-community/pixtral-12b
HF PR reference: huggingface/transformers#33449
Supporting the HF version means we can produce quantized versions of the model with LLM Compressor
The closest model vllm already supports.
No response
What's your difficulty of supporting the model you want?
Easy to moderate, all operations should already be implemented inside of vLLM
Before submitting a new issue...
- Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
ScrattlebeardKohakuBlueleaf and ReichenbachianKohakuBlueleaf and stt-anth
Metadata
Metadata
Assignees
Labels
new-modelRequests to new modelsRequests to new models