@@ -6179,9 +6179,9 @@ static bool llm_load_tensors(
61796179 layer.ffn_up = ml.create_tensor(ctx_split, tn(LLM_TENSOR_FFN_UP, "weight", i), {n_embd, n_ff});
61806180
61816181 // optional MLP bias
6182- layer.ffn_gate_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_GATE, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6183- layer.ffn_down_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_DOWN, "bias", i), {n_embd}, llama_model_loader::TENSOR_NOT_REQUIRED);
6184- layer.ffn_up_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_UP, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6182+ layer.ffn_gate_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_GATE, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6183+ layer.ffn_down_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_DOWN, "bias", i), {n_embd}, llama_model_loader::TENSOR_NOT_REQUIRED);
6184+ layer.ffn_up_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_UP, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
61856185 } else {
61866186 layer.ffn_gate_inp = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_FFN_GATE_INP, "weight", i), {n_embd, n_expert});
61876187
@@ -6505,7 +6505,7 @@ static bool llm_load_tensors(
65056505 layer.bv = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_V, "bias", i), {n_embd_gqa});
65066506
65076507 layer.wo = ml.create_tensor(ctx_split, tn(LLM_TENSOR_ATTN_OUT, "weight", i), {n_embd, n_embd}); //output_dens
6508- layer.bo = ml.create_tensor(ctx_split , tn(LLM_TENSOR_ATTN_OUT, "bias", i), {n_embd}); //output_dens
6508+ layer.bo = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_ATTN_OUT, "bias", i), {n_embd}); //output_dens
65096509
65106510 layer.attn_out_norm = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_OUT_NORM, "weight", i), {n_embd}); //output_norm
65116511 layer.attn_out_norm_b = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_OUT_NORM, "bias", i), {n_embd});
0 commit comments