@@ -1777,26 +1777,26 @@ static const std::map<llm_arch, std::map<llm_tensor, const char *>> LLM_TENSOR_N
1777
1777
},
1778
1778
},
1779
1779
{
1780
- LLM_ARCH_UNKNOWN ,
1780
+ LLM_ARCH_SMOLLM3 ,
1781
1781
{
1782
- { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1782
+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1783
+ { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1784
+ { LLM_TENSOR_OUTPUT, " output" },
1785
+ { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1786
+ { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1787
+ { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1788
+ { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1789
+ { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1790
+ { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1791
+ { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1792
+ { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1793
+ { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
1783
1794
},
1784
1795
},
1785
1796
{
1786
- LLM_ARCH_SMOLLM3 ,
1797
+ LLM_ARCH_UNKNOWN ,
1787
1798
{
1788
- { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1789
- { LLM_TENSOR_OUTPUT_NORM, " output_norm" },
1790
- { LLM_TENSOR_OUTPUT, " output" },
1791
- { LLM_TENSOR_ATTN_NORM, " blk.%d.attn_norm" },
1792
- { LLM_TENSOR_ATTN_Q, " blk.%d.attn_q" },
1793
- { LLM_TENSOR_ATTN_K, " blk.%d.attn_k" },
1794
- { LLM_TENSOR_ATTN_V, " blk.%d.attn_v" },
1795
- { LLM_TENSOR_ATTN_OUT, " blk.%d.attn_output" },
1796
- { LLM_TENSOR_FFN_NORM, " blk.%d.ffn_norm" },
1797
- { LLM_TENSOR_FFN_GATE, " blk.%d.ffn_gate" },
1798
- { LLM_TENSOR_FFN_DOWN, " blk.%d.ffn_down" },
1799
- { LLM_TENSOR_FFN_UP, " blk.%d.ffn_up" },
1799
+ { LLM_TENSOR_TOKEN_EMBD, " token_embd" },
1800
1800
},
1801
1801
},
1802
1802
};
0 commit comments