Skip to content

Commit 9640fe9

Browse files
authored
Update llama.cpp - Fix non-fa ppl
Suggested by @ubergarm - ikawrakow#662 (comment)
1 parent 0040a42 commit 9640fe9

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

src/llama.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -16254,7 +16254,7 @@ struct llm_build_context {
1625416254
// reshape for multi-head
1625516255
Qcur = ggml_reshape_3d(ctx0, Qcur, n_embd_head, n_head, n_tokens);
1625616256
Kcur = ggml_reshape_3d(ctx0, Kcur, n_embd_head, n_head_kv, n_tokens);
16257-
Vcur = ggml_reshape_3d(ctx0, Vcur, n_embd_head, n_head_kv, n_tokens);
16257+
// Vcur = ggml_reshape_3d(ctx0, Vcur, n_embd_head, n_head_kv, n_tokens);
1625816258

1625916259
// Apply Q/K norm if available (GLM-4.5 355B variant)
1626016260
if (model.layers[il].attn_q_norm) {

0 commit comments

Comments
 (0)