@@ -1733,7 +1733,7 @@ void llama_model::load_hparams(llama_model_loader & ml) {
17331733 type = LLM_TYPE_UNKNOWN;
17341734 }
17351735 } break;
1736- case LLM_ARCH_HUNYUAN_MOE :
1736+ case LLM_ARCH_HUNYUAN_V1_MOE :
17371737 {
17381738 ml.get_key(LLM_KV_ATTENTION_LAYERNORM_RMS_EPS, hparams.f_norm_rms_eps);
17391739 ml.get_key(LLM_KV_EXPERT_FEED_FORWARD_LENGTH, hparams.n_ff_exp);
@@ -5078,7 +5078,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
50785078 layer.ffn_up_b = create_tensor(tn(LLM_TENSOR_FFN_UP, "bias", i), {ffn_intermediate_size}, TENSOR_NOT_REQUIRED);
50795079 }
50805080 } break;
5081- case LLM_ARCH_HUNYUAN_MOE :
5081+ case LLM_ARCH_HUNYUAN_V1_MOE :
50825082 {
50835083 tok_embd = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, 0);
50845084
@@ -16580,8 +16580,8 @@ struct llm_build_arcee : public llm_graph_context {
1658016580 }
1658116581};
1658216582
16583- struct llm_build_hunyuan_moe : public llm_graph_context {
16584- llm_build_hunyuan_moe (const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
16583+ struct llm_build_hunyuan_v1_moe : public llm_graph_context {
16584+ llm_build_hunyuan_v1_moe (const llama_model & model, const llm_graph_params & params) : llm_graph_context(params) {
1658516585 const int64_t n_embd_head = hparams.n_embd_head_v;
1658616586
1658716587 GGML_ASSERT(n_embd_head == hparams.n_embd_head_k);
@@ -17615,9 +17615,9 @@ ggml_cgraph * llama_model::build_graph(const llm_graph_params & params) const {
1761517615 {
1761617616 llm = std::make_unique<llm_build_ernie4_5_moe>(*this, params);
1761717617 } break;
17618- case LLM_ARCH_HUNYUAN_MOE :
17618+ case LLM_ARCH_HUNYUAN_V1_MOE :
1761917619 {
17620- llm = std::make_unique<llm_build_hunyuan_moe >(*this, params);
17620+ llm = std::make_unique<llm_build_hunyuan_v1_moe >(*this, params);
1762117621 } break;
1762217622 case LLM_ARCH_HUNYUAN_V1_DENSE:
1762317623 {
@@ -17831,7 +17831,7 @@ llama_rope_type llama_model_rope_type(const llama_model * model) {
1783117831 case LLM_ARCH_EXAONE4:
1783217832 case LLM_ARCH_MINICPM3:
1783317833 case LLM_ARCH_DOTS1:
17834- case LLM_ARCH_HUNYUAN_MOE :
17834+ case LLM_ARCH_HUNYUAN_V1_MOE :
1783517835 case LLM_ARCH_HUNYUAN_V1_DENSE:
1783617836 case LLM_ARCH_LFM2:
1783717837 return LLAMA_ROPE_TYPE_NEOX;
0 commit comments