Skip to content

Commit 78eb487

Browse files
authored
llama : fix qs.n_attention_wv for DeepSeek-V2 (#9156)
1 parent a77feb5 commit 78eb487

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

src/llama.cpp

+2-1
Original file line numberDiff line numberDiff line change
@@ -16822,7 +16822,8 @@ static void llama_model_quantize_internal(const std::string & fname_inp, const s
1682216822

1682316823
// TODO: avoid hardcoded tensor names - use the TN_* constants
1682416824
if (name.find("attn_v.weight") != std::string::npos ||
16825-
name.find("attn_qkv.weight") != std::string::npos) {
16825+
name.find("attn_qkv.weight") != std::string::npos ||
16826+
name.find("attn_kv_b.weight")!= std::string::npos) {
1682616827
++qs.n_attention_wv;
1682716828
} else if (name == LLM_TN(model.arch)(LLM_TENSOR_OUTPUT, "weight")) {
1682816829
qs.has_output = true;

0 commit comments

Comments
 (0)