@@ -1202,6 +1202,7 @@ def set_gguf_parameters(self):
12021202 self .gguf_writer .add_head_count_kv (hparams ["num_key_value_heads" ])
12031203 self .gguf_writer .add_parallel_residual (hparams ["use_parallel_residual" ] if "use_parallel_residual" in hparams else True )
12041204 self .gguf_writer .add_layer_norm_eps (self .find_hparam (["layer_norm_eps" , "norm_eps" ]))
1205+ self .gguf_writer .add_file_type (self .ftype )
12051206
12061207 _q_norms : list [dict [str , Tensor ]] | None = None
12071208 _k_norms : list [dict [str , Tensor ]] | None = None
@@ -1578,6 +1579,7 @@ def set_gguf_parameters(self):
15781579 self .gguf_writer .add_rope_dimension_count (self .hparams ["hidden_size" ] // self .hparams ["num_attention_heads" ])
15791580 self .gguf_writer .add_head_count (self .hparams ["num_attention_heads" ])
15801581 self .gguf_writer .add_layer_norm_rms_eps (self .hparams ["layer_norm_epsilon" ])
1582+ self .gguf_writer .add_file_type (self .ftype )
15811583
15821584
15831585@Model .register ("Qwen2ForCausalLM" )
@@ -1815,6 +1817,7 @@ def set_gguf_parameters(self):
18151817 self .gguf_writer .add_head_count (hparams ["num_attention_heads" ])
18161818 self .gguf_writer .add_head_count_kv (5 ) # hparams["num_key_value_heads"]) is wrong
18171819 self .gguf_writer .add_layer_norm_rms_eps (hparams ["rms_norm_eps" ])
1820+ self .gguf_writer .add_file_type (self .ftype )
18181821
18191822 def shuffle_attn_q_weight (self , data_torch ):
18201823 assert data_torch .size () == (5120 , 5120 )
@@ -1994,6 +1997,7 @@ def set_gguf_parameters(self):
19941997 self .gguf_writer .add_head_count (self .hparams ["num_attention_heads" ])
19951998 self .gguf_writer .add_layer_norm_rms_eps (self .hparams ["rms_norm_eps" ])
19961999 self .gguf_writer .add_head_count_kv (self .hparams ["num_key_value_heads" ])
2000+ self .gguf_writer .add_file_type (self .ftype )
19972001
19982002 def modify_tensors (self , data_torch : Tensor , name : str , bid : int | None ) -> Iterable [tuple [str , Tensor ]]:
19992003 num_heads = self .hparams ["num_attention_heads" ]
0 commit comments