You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is --outtype supposed to convert to the type, eg: --outtype f16 on a f32 LoRA adapter in PEFT format is supposed to create an f16 LoRA in GGUF format (ie: the same command semantics as convert_hf_to_gguf_update.py)?
or is it just a "type hint" incase it can't infer the type of the PEFT format LoRA?
I also tried running llama-quant on the resulting LoRA in f32 GGUF format, but it doesn't like being given these files...
Are non-f32 LoRAs even allowed?
If it is a bug then I'll do some more investigation and see if I can find out what is happening.