{ "model.layers.3.feed_forward.routed_experts.ffn.w2._layer.weight": "part1.safetensors", "lm_head.weight": "part2.safetensors", "model.tok_embeddings.embedding_weight": "part3.safetensors", "model.layers.2.feed_forward.w2._layer.weight": "part4.safetensors", "model.layers.2.attention.wo._layer.weight": "part4.safetensors", "model.layers.0.attention.l2q_proj._layer.weight": "part4.safetensors", "model.layers.1.attention.l2q_proj._layer.weight": "part4.safetensors", "model.layers.2.attention.l2q_proj._layer.weight": "part4.safetensors", "model.layers.0.attention.lkv2kv_v.weight": "part4.safetensors", "model.layers.1.attention.lkv2kv_v.weight": "part4.safetensors", "model.layers.2.attention.lkv2kv_v.weight": "part4.safetensors", "model.layers.3.attention.lkv2kv_v.weight": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.ffn.w2._layer.matmul.weight_scale": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.router.dense.weight": "part4.safetensors", "model.layers.0.attention.l2q_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.0.attention.l2q_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.1.attention.l2q_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.1.attention.l2q_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.2.attention.l2q_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.2.attention.l2q_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.3.attention.l2q_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.3.attention.l2q_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.0.attention.wo.quant_op.input_scale": "part4.safetensors", "model.layers.1.attention.wo.quant_op.input_scale": "part4.safetensors", "model.layers.2.attention.wo.quant_op.input_scale": "part4.safetensors", "model.layers.3.attention.wo.quant_op.input_scale": "part4.safetensors", "model.layers.0.attention.wo._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.0.attention.wo._layer.matmul.quant_bias": "part4.safetensors", "model.layers.1.attention.wo._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.1.attention.wo._layer.matmul.quant_bias": "part4.safetensors", "model.layers.2.attention.wo._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.2.attention.wo._layer.matmul.quant_bias": "part4.safetensors", "model.layers.3.attention.wo._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.3.attention.wo._layer.matmul.quant_bias": "part4.safetensors", "model.layers.0.attention.wo.quant_op.input_zp": "part4.safetensors", "model.layers.1.attention.wo.quant_op.input_zp": "part4.safetensors", "model.layers.2.attention.wo.quant_op.input_zp": "part4.safetensors", "model.layers.3.attention.wo.quant_op.input_zp": "part4.safetensors", "model.layers.0.attention_norm.weight": "part4.safetensors", "model.layers.0.feed_forward.w2._layer.matmul.weight_scale": "part4.safetensors", "model.layers.0.ffn_norm.weight": "part4.safetensors", "model.layers.1.attention_norm.weight": "part4.safetensors", "model.layers.1.feed_forward.w2._layer.matmul.weight_scale": "part4.safetensors", "model.layers.1.ffn_norm.weight": "part4.safetensors", "model.layers.2.attention_norm.weight": "part4.safetensors", "model.layers.2.feed_forward.w2._layer.matmul.weight_scale": "part4.safetensors", "model.layers.2.ffn_norm.weight": "part4.safetensors", "model.layers.3.attention_norm.weight": "part4.safetensors", "model.layers.3.feed_forward.shared_experts.w2._layer.matmul.weight_scale": "part4.safetensors", "model.layers.3.ffn_norm.weight": "part4.safetensors", "model.norm_out.weight": "part4.safetensors", "model.layers.0.attention.l2q_proj.quant_op.input_scale": "part4.safetensors", "model.layers.0.attention.lq_norm.weight": "part4.safetensors", "model.layers.1.attention.l2q_proj.quant_op.input_scale": "part4.safetensors", "model.layers.1.attention.lq_norm.weight": "part4.safetensors", "model.layers.2.attention.l2q_proj.quant_op.input_scale": "part4.safetensors", "model.layers.2.attention.lq_norm.weight": "part4.safetensors", "model.layers.3.attention.l2q_proj.quant_op.input_scale": "part4.safetensors", "model.layers.3.attention.lq_norm.weight": "part4.safetensors", "model.layers.0.attention.l2q_proj.quant_op.input_zp": "part4.safetensors", "model.layers.1.attention.l2q_proj.quant_op.input_zp": "part4.safetensors", "model.layers.2.attention.l2q_proj.quant_op.input_zp": "part4.safetensors", "model.layers.3.attention.l2q_proj.quant_op.input_zp": "part4.safetensors", "model.layers.0.attention.lkv_norm.weight": "part4.safetensors", "model.layers.1.attention.lkv_norm.weight": "part4.safetensors", "model.layers.2.attention.lkv_norm.weight": "part4.safetensors", "model.layers.3.attention.lkv_norm.weight": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.router.e_score_correction_bias": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.router.router.topk_bias": "part4.safetensors", "model.layers.0.feed_forward.w2._layer.weight": "part5.safetensors", "model.layers.1.feed_forward.w2._layer.weight": "part5.safetensors", "model.layers.0.attention.wo._layer.weight": "part5.safetensors", "model.layers.1.attention.wo._layer.weight": "part5.safetensors", "model.layers.3.attention.wo._layer.weight": "part5.safetensors", "model.layers.3.attention.l2q_proj._layer.weight": "part5.safetensors", "model.layers.0.attention.lkv2kv_k_nope.weight": "part5.safetensors", "model.layers.1.attention.lkv2kv_k_nope.weight": "part5.safetensors", "model.layers.2.attention.lkv2kv_k_nope.weight": "part5.safetensors", "model.layers.3.attention.lkv2kv_k_nope.weight": "part5.safetensors", "model.layers.3.feed_forward.shared_experts.w2._layer.weight": "part5.safetensors", "model.layers.3.feed_forward.routed_experts.ffn.w1._layer.weight": "part0.safetensors", "model.layers.3.feed_forward.routed_experts.ffn.w3._layer.weight": "part0.safetensors", "model.layers.0.feed_forward.w1._layer.weight": "part4.safetensors", "model.layers.0.feed_forward.w3._layer.weight": "part4.safetensors", "model.layers.2.feed_forward.w1._layer.weight": "part4.safetensors", "model.layers.2.feed_forward.w3._layer.weight": "part4.safetensors", "model.layers.3.feed_forward.shared_experts.w1._layer.weight": "part4.safetensors", "model.layers.3.feed_forward.shared_experts.w3._layer.weight": "part4.safetensors", "model.layers.1.attention.q2l_proj._layer.weight": "part4.safetensors", "model.layers.1.attention.kv2l._layer.weight": "part4.safetensors", "model.layers.3.attention.q2l_proj._layer.weight": "part4.safetensors", "model.layers.3.attention.kv2l._layer.weight": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.ffn.w1._layer.matmul.weight_scale": "part4.safetensors", "model.layers.3.feed_forward.routed_experts.ffn.w3._layer.matmul.weight_scale": "part4.safetensors", "model.layers.0.feed_forward.w1._layer.matmul.weight_scale": "part4.safetensors", "model.layers.0.feed_forward.w3._layer.matmul.weight_scale": "part4.safetensors", "model.layers.1.feed_forward.w1._layer.matmul.weight_scale": "part4.safetensors", "model.layers.1.feed_forward.w3._layer.matmul.weight_scale": "part4.safetensors", "model.layers.2.feed_forward.w1._layer.matmul.weight_scale": "part4.safetensors", "model.layers.2.feed_forward.w3._layer.matmul.weight_scale": "part4.safetensors", "model.layers.0.attention.q2l_proj.quant_op.input_scale": "part4.safetensors", "model.layers.0.attention.kv2l.quant_op.input_scale": "part4.safetensors", "model.layers.1.attention.q2l_proj.quant_op.input_scale": "part4.safetensors", "model.layers.1.attention.kv2l.quant_op.input_scale": "part4.safetensors", "model.layers.2.attention.q2l_proj.quant_op.input_scale": "part4.safetensors", "model.layers.2.attention.kv2l.quant_op.input_scale": "part4.safetensors", "model.layers.3.attention.q2l_proj.quant_op.input_scale": "part4.safetensors", "model.layers.3.attention.kv2l.quant_op.input_scale": "part4.safetensors", "model.layers.0.attention.q2l_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.0.attention.kv2l._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.0.attention.q2l_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.0.attention.kv2l._layer.matmul.quant_bias": "part4.safetensors", "model.layers.1.attention.q2l_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.1.attention.kv2l._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.1.attention.q2l_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.1.attention.kv2l._layer.matmul.quant_bias": "part4.safetensors", "model.layers.2.attention.q2l_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.2.attention.kv2l._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.2.attention.q2l_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.2.attention.kv2l._layer.matmul.quant_bias": "part4.safetensors", "model.layers.3.attention.q2l_proj._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.3.attention.kv2l._layer.matmul.dequant_scale": "part4.safetensors", "model.layers.3.attention.q2l_proj._layer.matmul.quant_bias": "part4.safetensors", "model.layers.3.attention.kv2l._layer.matmul.quant_bias": "part4.safetensors", "model.layers.3.feed_forward.shared_experts.w1._layer.matmul.weight_scale": "part4.safetensors", "model.layers.3.feed_forward.shared_experts.w3._layer.matmul.weight_scale": "part4.safetensors", "model.layers.0.attention.q2l_proj.quant_op.input_zp": "part4.safetensors", "model.layers.0.attention.kv2l.quant_op.input_zp": "part4.safetensors", "model.layers.1.attention.q2l_proj.quant_op.input_zp": "part4.safetensors", "model.layers.1.attention.kv2l.quant_op.input_zp": "part4.safetensors", "model.layers.2.attention.q2l_proj.quant_op.input_zp": "part4.safetensors", "model.layers.2.attention.kv2l.quant_op.input_zp": "part4.safetensors", "model.layers.3.attention.q2l_proj.quant_op.input_zp": "part4.safetensors", "model.layers.3.attention.kv2l.quant_op.input_zp": "part4.safetensors", "model.layers.1.feed_forward.w1._layer.weight": "part5.safetensors", "model.layers.1.feed_forward.w3._layer.weight": "part5.safetensors", "model.layers.0.attention.q2l_proj._layer.weight": "part5.safetensors", "model.layers.0.attention.kv2l._layer.weight": "part5.safetensors", "model.layers.2.attention.q2l_proj._layer.weight": "part5.safetensors", "model.layers.2.attention.kv2l._layer.weight": "part5.safetensors" }