7767517 188 353 Input in0 0 1 in0 Split splitncnn_0 1 2 in0 1 2 Input in1 0 1 in1 Split splitncnn_1 1 24 in1 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 Input in2 0 1 in2 Split splitncnn_2 1 12 in2 29 30 31 32 33 34 35 36 37 38 39 40 Input kv_cache 0 48 cache_k0 cache_v0 cache_k1 cache_v1 cache_k2 cache_v2 cache_k3 cache_v3 cache_k4 cache_v4 cache_k5 cache_v5 cache_k6 cache_v6 cache_k7 cache_v7 cache_k8 cache_v8 cache_k9 cache_v9 cache_k10 cache_v10 cache_k11 cache_v11 cache_k12 cache_v12 cache_k13 cache_v13 cache_k14 cache_v14 cache_k15 cache_v15 cache_k16 cache_v16 cache_k17 cache_v17 cache_k18 cache_v18 cache_k19 cache_v19 cache_k20 cache_v20 cache_k21 cache_v21 cache_k22 cache_v22 cache_k23 cache_v23 LayerNorm ln_37 1 1 2 41 0=1024 1=1.000000e-05 2=1 Reshape reshape_98 1 1 40 42 0=1 1=1 MultiHeadAttention attention_25 4 3 41 42 cache_k0 cache_v0 43 out_cache_k0 out_cache_v0 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_0 2 1 1 43 44 0=0 Split splitncnn_3 1 2 44 45 46 LayerNorm ln_38 1 1 46 47 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_74 5 3 47 26 27 cache_k1 cache_v1 48 out_cache_k1 out_cache_v1 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_1 2 1 45 48 49 0=0 Split splitncnn_4 1 2 49 50 51 LayerNorm ln_39 1 1 51 52 0=1024 1=1.000000e-05 2=1 Gemm gemm_0 1 1 52 53 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_86 1 1 53 54 Gemm gemm_1 1 1 54 55 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_2 2 1 50 55 56 0=0 Split splitncnn_5 1 2 56 57 58 LayerNorm ln_40 1 1 58 59 0=1024 1=1.000000e-05 2=1 Reshape reshape_99 1 1 39 60 0=1 1=1 MultiHeadAttention attention_26 4 3 59 60 cache_k2 cache_v2 61 out_cache_k2 out_cache_v2 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_3 2 1 57 61 62 0=0 Split splitncnn_6 1 2 62 63 64 LayerNorm ln_41 1 1 64 65 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_75 5 3 65 24 25 cache_k3 cache_v3 66 out_cache_k3 out_cache_v3 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_4 2 1 63 66 67 0=0 Split splitncnn_7 1 2 67 68 69 LayerNorm ln_42 1 1 69 70 0=1024 1=1.000000e-05 2=1 Gemm gemm_2 1 1 70 71 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_87 1 1 71 72 Gemm gemm_3 1 1 72 73 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_5 2 1 68 73 74 0=0 Split splitncnn_8 1 2 74 75 76 LayerNorm ln_43 1 1 76 77 0=1024 1=1.000000e-05 2=1 Reshape reshape_100 1 1 38 78 0=1 1=1 MultiHeadAttention attention_27 4 3 77 78 cache_k4 cache_v4 79 out_cache_k4 out_cache_v4 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_6 2 1 75 79 80 0=0 Split splitncnn_9 1 2 80 81 82 LayerNorm ln_44 1 1 82 83 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_76 5 3 83 22 23 cache_k5 cache_v5 84 out_cache_k5 out_cache_v5 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_7 2 1 81 84 85 0=0 Split splitncnn_10 1 2 85 86 87 LayerNorm ln_45 1 1 87 88 0=1024 1=1.000000e-05 2=1 Gemm gemm_4 1 1 88 89 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_88 1 1 89 90 Gemm gemm_5 1 1 90 91 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_8 2 1 86 91 92 0=0 Split splitncnn_11 1 2 92 93 94 LayerNorm ln_46 1 1 94 95 0=1024 1=1.000000e-05 2=1 Reshape reshape_101 1 1 37 96 0=1 1=1 MultiHeadAttention attention_28 4 3 95 96 cache_k6 cache_v6 97 out_cache_k6 out_cache_v6 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_9 2 1 93 97 98 0=0 Split splitncnn_12 1 2 98 99 100 LayerNorm ln_47 1 1 100 101 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_77 5 3 101 20 21 cache_k7 cache_v7 102 out_cache_k7 out_cache_v7 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_10 2 1 99 102 103 0=0 Split splitncnn_13 1 2 103 104 105 LayerNorm ln_48 1 1 105 106 0=1024 1=1.000000e-05 2=1 Gemm gemm_6 1 1 106 107 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_89 1 1 107 108 Gemm gemm_7 1 1 108 109 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_11 2 1 104 109 110 0=0 Split splitncnn_14 1 2 110 111 112 LayerNorm ln_49 1 1 112 113 0=1024 1=1.000000e-05 2=1 Reshape reshape_102 1 1 36 114 0=1 1=1 MultiHeadAttention attention_29 4 3 113 114 cache_k8 cache_v8 115 out_cache_k8 out_cache_v8 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_12 2 1 111 115 116 0=0 Split splitncnn_15 1 2 116 117 118 LayerNorm ln_50 1 1 118 119 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_78 5 3 119 18 19 cache_k9 cache_v9 120 out_cache_k9 out_cache_v9 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_13 2 1 117 120 121 0=0 Split splitncnn_16 1 2 121 122 123 LayerNorm ln_51 1 1 123 124 0=1024 1=1.000000e-05 2=1 Gemm gemm_8 1 1 124 125 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_90 1 1 125 126 Gemm gemm_9 1 1 126 127 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_14 2 1 122 127 128 0=0 Split splitncnn_17 1 2 128 129 130 LayerNorm ln_52 1 1 130 131 0=1024 1=1.000000e-05 2=1 Reshape reshape_103 1 1 35 132 0=1 1=1 MultiHeadAttention attention_30 4 3 131 132 cache_k10 cache_v10 133 out_cache_k10 out_cache_v10 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_15 2 1 129 133 134 0=0 Split splitncnn_18 1 2 134 135 136 LayerNorm ln_53 1 1 136 137 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_79 5 3 137 16 17 cache_k11 cache_v11 138 out_cache_k11 out_cache_v11 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_16 2 1 135 138 139 0=0 Split splitncnn_19 1 2 139 140 141 LayerNorm ln_54 1 1 141 142 0=1024 1=1.000000e-05 2=1 Gemm gemm_10 1 1 142 143 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_91 1 1 143 144 Gemm gemm_11 1 1 144 145 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_17 2 1 140 145 146 0=0 Split splitncnn_20 1 2 146 147 148 LayerNorm ln_55 1 1 148 149 0=1024 1=1.000000e-05 2=1 Reshape reshape_104 1 1 34 150 0=1 1=1 MultiHeadAttention attention_31 4 3 149 150 cache_k12 cache_v12 151 out_cache_k12 out_cache_v12 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_18 2 1 147 151 152 0=0 Split splitncnn_21 1 2 152 153 154 LayerNorm ln_56 1 1 154 155 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_80 5 3 155 14 15 cache_k13 cache_v13 156 out_cache_k13 out_cache_v13 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_19 2 1 153 156 157 0=0 Split splitncnn_22 1 2 157 158 159 LayerNorm ln_57 1 1 159 160 0=1024 1=1.000000e-05 2=1 Gemm gemm_12 1 1 160 161 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_92 1 1 161 162 Gemm gemm_13 1 1 162 163 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_20 2 1 158 163 164 0=0 Split splitncnn_23 1 2 164 165 166 LayerNorm ln_58 1 1 166 167 0=1024 1=1.000000e-05 2=1 Reshape reshape_105 1 1 33 168 0=1 1=1 MultiHeadAttention attention_32 4 3 167 168 cache_k14 cache_v14 169 out_cache_k14 out_cache_v14 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_21 2 1 165 169 170 0=0 Split splitncnn_24 1 2 170 171 172 LayerNorm ln_59 1 1 172 173 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_81 5 3 173 12 13 cache_k15 cache_v15 174 out_cache_k15 out_cache_v15 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_22 2 1 171 174 175 0=0 Split splitncnn_25 1 2 175 176 177 LayerNorm ln_60 1 1 177 178 0=1024 1=1.000000e-05 2=1 Gemm gemm_14 1 1 178 179 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_93 1 1 179 180 Gemm gemm_15 1 1 180 181 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_23 2 1 176 181 182 0=0 Split splitncnn_26 1 2 182 183 184 LayerNorm ln_61 1 1 184 185 0=1024 1=1.000000e-05 2=1 Reshape reshape_106 1 1 32 186 0=1 1=1 MultiHeadAttention attention_33 4 3 185 186 cache_k16 cache_v16 187 out_cache_k16 out_cache_v16 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_24 2 1 183 187 188 0=0 Split splitncnn_27 1 2 188 189 190 LayerNorm ln_62 1 1 190 191 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_82 5 3 191 10 11 cache_k17 cache_v17 192 out_cache_k17 out_cache_v17 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_25 2 1 189 192 193 0=0 Split splitncnn_28 1 2 193 194 195 LayerNorm ln_63 1 1 195 196 0=1024 1=1.000000e-05 2=1 Gemm gemm_16 1 1 196 197 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_94 1 1 197 198 Gemm gemm_17 1 1 198 199 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_26 2 1 194 199 200 0=0 Split splitncnn_29 1 2 200 201 202 LayerNorm ln_64 1 1 202 203 0=1024 1=1.000000e-05 2=1 Reshape reshape_107 1 1 31 204 0=1 1=1 MultiHeadAttention attention_34 4 3 203 204 cache_k18 cache_v18 205 out_cache_k18 out_cache_v18 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_27 2 1 201 205 206 0=0 Split splitncnn_30 1 2 206 207 208 LayerNorm ln_65 1 1 208 209 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_83 5 3 209 8 9 cache_k19 cache_v19 210 out_cache_k19 out_cache_v19 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_28 2 1 207 210 211 0=0 Split splitncnn_31 1 2 211 212 213 LayerNorm ln_66 1 1 213 214 0=1024 1=1.000000e-05 2=1 Gemm gemm_18 1 1 214 215 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_95 1 1 215 216 Gemm gemm_19 1 1 216 217 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_29 2 1 212 217 218 0=0 Split splitncnn_32 1 2 218 219 220 LayerNorm ln_67 1 1 220 221 0=1024 1=1.000000e-05 2=1 Reshape reshape_108 1 1 30 222 0=1 1=1 MultiHeadAttention attention_35 4 3 221 222 cache_k20 cache_v20 223 out_cache_k20 out_cache_v20 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_30 2 1 219 223 224 0=0 Split splitncnn_33 1 2 224 225 226 LayerNorm ln_68 1 1 226 227 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_84 5 3 227 6 7 cache_k21 cache_v21 228 out_cache_k21 out_cache_v21 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_31 2 1 225 228 229 0=0 Split splitncnn_34 1 2 229 230 231 LayerNorm ln_69 1 1 231 232 0=1024 1=1.000000e-05 2=1 Gemm gemm_20 1 1 232 233 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_96 1 1 233 234 Gemm gemm_21 1 1 234 235 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_32 2 1 230 235 236 0=0 Split splitncnn_35 1 2 236 237 238 LayerNorm ln_70 1 1 238 239 0=1024 1=1.000000e-05 2=1 Reshape reshape_109 1 1 29 240 0=1 1=1 MultiHeadAttention attention_36 4 3 239 240 cache_k22 cache_v22 241 out_cache_k22 out_cache_v22 0=1024 1=16 2=1048576 3=1024 4=1024 5=1 7=1 BinaryOp add_33 2 1 237 241 242 0=0 Split splitncnn_36 1 2 242 243 244 LayerNorm ln_71 1 1 244 245 0=1024 1=1.000000e-05 2=1 MultiHeadAttention attention_85 5 3 245 4 5 cache_k23 cache_v23 246 out_cache_k23 out_cache_v23 0=1024 1=16 2=1048576 3=1024 4=1024 7=1 BinaryOp add_34 2 1 243 246 247 0=0 Split splitncnn_37 1 2 247 248 249 LayerNorm ln_72 1 1 249 250 0=1024 1=1.000000e-05 2=1 Gemm gemm_22 1 1 250 251 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=4096 9=1024 ReLU relu_97 1 1 251 252 Gemm gemm_23 1 1 252 253 10=4 2=0 3=1 4=0 5=1 6=1 7=0 8=1024 9=4096 BinaryOp add_35 2 1 248 253 254 0=0 LayerNorm ln_73 1 1 254 255 0=1024 1=1.000000e-05 2=1 Gemm gemm_24 1 1 255 out0 10=-1 2=0 3=1 4=0 5=1 6=1 7=0 8=256206 9=1024