Skip to content

(possible) Misc. bug: convert_lora_to_gguf.py ignores --outtype #15890

@jukofyork

Description

@jukofyork

Name and Version

Tested on latest pull.

Problem description & steps to reproduce

I'm not sure if this is a bug or intended:

  • Is --outtype supposed to convert to the type, eg: --outtype f16 on a f32 LoRA adapter in PEFT format is supposed to create an f16 LoRA in GGUF format (ie: the same command semantics as convert_hf_to_gguf_update.py)?
  • or is it just a "type hint" incase it can't infer the type of the PEFT format LoRA?

I also tried running llama-quant on the resulting LoRA in f32 GGUF format, but it doesn't like being given these files...

Are non-f32 LoRAs even allowed?

If it is a bug then I'll do some more investigation and see if I can find out what is happening.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions