root
commited on
Commit
·
f1bc1f0
1
Parent(s):
e8fd46a
update
Browse filesSigned-off-by: root <[email protected]>
- model-00001-of-00014.safetensors +2 -2
- model-00002-of-00014.safetensors +2 -2
- model-00003-of-00014.safetensors +2 -2
- model-00004-of-00014.safetensors +2 -2
- model-00005-of-00014.safetensors +2 -2
- model-00006-of-00014.safetensors +2 -2
- model-00007-of-00014.safetensors +2 -2
- model-00008-of-00014.safetensors +2 -2
- model-00009-of-00014.safetensors +2 -2
- model-00010-of-00014.safetensors +2 -2
- model-00011-of-00014.safetensors +2 -2
- model-00012-of-00014.safetensors +2 -2
- model-00013-of-00014.safetensors +2 -2
- model-00014-of-00014.safetensors +2 -2
- model.safetensors.index.json +26 -26
model-00001-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c8245d49c65c1be03de3d97310f58951b25d867353e0f8409d2e25d5308b0a5
|
| 3 |
+
size 4932307584
|
model-00002-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:025ccf9d44b268e397cb49caf5fd8836ddc0861c6920c8223ce84bd4dde5d07a
|
| 3 |
+
size 4875989696
|
model-00003-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c0a2b1d7d09da8e1f2007877ef78c6d60cfefee1f181a7a519a3a184405fc725
|
| 3 |
+
size 4875989720
|
model-00004-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e56acc07cb1c87d2d8576ecd0435e302951de7fbfca41f4f65ab32b367a167c
|
| 3 |
+
size 4875989752
|
model-00005-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:33d43d09da1b62b5fdad2f9134c11a7598573b39d4cc6daa62feaf6ce0d5146a
|
| 3 |
+
size 4875989752
|
model-00006-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b180d65bfda25d31931a5887bec20e59cb8a3fd24d866bc1bfb6edf4fcd8283
|
| 3 |
+
size 4875989752
|
model-00007-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba71af62d1e7f8d3ff2fbcc7537adfe97223501b45bfb8802f3d40221123b9d3
|
| 3 |
+
size 4875989752
|
model-00008-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3791cc769a06c3dd965b594d0ebed36026f97f3cf59ae859aa92ba3fd6781e6
|
| 3 |
+
size 4875989752
|
model-00009-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:03556cfb7159e2f3999ef1610c5eb44456a29bb6328af26889d785c4d2f7f6de
|
| 3 |
+
size 4875989752
|
model-00010-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5149f2d34a2e3bc4248ae89c9a046491066395254b9c35765b95257f0e1c2bdf
|
| 3 |
+
size 4875989752
|
model-00011-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb2b02a41b63208152df53fa5b9d55ab17a052b604051047d3058c3d54940dc9
|
| 3 |
+
size 4875989752
|
model-00012-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9104fdfb1cccaf14d2114a8b767578630c7cbe218e0fd1af14638bdbc9737f1
|
| 3 |
+
size 4875989752
|
model-00013-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:01b02c17fd7ea0ff6ea9d836a601ab38a74b1c8593c4800dfdbe7c8392828cd1
|
| 3 |
+
size 4875989752
|
model-00014-of-00014.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4e83c27b30b91e83a2c0984e6ddd57baac2b0a2ce30ade1386ed2e534e8ff6c
|
| 3 |
+
size 2080144040
|
model.safetensors.index.json
CHANGED
|
@@ -60,11 +60,11 @@
|
|
| 60 |
"model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
|
| 61 |
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
|
| 62 |
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
|
| 63 |
-
"model.layers.13.input_layernorm.weight": "model-
|
| 64 |
"model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
|
| 65 |
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
|
| 66 |
"model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
|
| 67 |
-
"model.layers.13.post_attention_layernorm.weight": "model-
|
| 68 |
"model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
|
| 69 |
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
|
| 70 |
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
|
|
@@ -115,11 +115,11 @@
|
|
| 115 |
"model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
|
| 116 |
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
|
| 117 |
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
|
| 118 |
-
"model.layers.18.input_layernorm.weight": "model-
|
| 119 |
"model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
|
| 120 |
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
|
| 121 |
"model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
|
| 122 |
-
"model.layers.18.post_attention_layernorm.weight": "model-
|
| 123 |
"model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
|
| 124 |
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
|
| 125 |
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
|
|
@@ -181,11 +181,11 @@
|
|
| 181 |
"model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
|
| 182 |
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
|
| 183 |
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
|
| 184 |
-
"model.layers.23.input_layernorm.weight": "model-
|
| 185 |
"model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
|
| 186 |
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
|
| 187 |
"model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
|
| 188 |
-
"model.layers.23.post_attention_layernorm.weight": "model-
|
| 189 |
"model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
|
| 190 |
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
|
| 191 |
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
|
|
@@ -236,11 +236,11 @@
|
|
| 236 |
"model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
|
| 237 |
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
|
| 238 |
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
|
| 239 |
-
"model.layers.28.input_layernorm.weight": "model-
|
| 240 |
"model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
|
| 241 |
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
|
| 242 |
"model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
|
| 243 |
-
"model.layers.28.post_attention_layernorm.weight": "model-
|
| 244 |
"model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
|
| 245 |
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
|
| 246 |
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
|
|
@@ -258,11 +258,11 @@
|
|
| 258 |
"model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
| 259 |
"model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
| 260 |
"model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
| 261 |
-
"model.layers.3.input_layernorm.weight": "model-
|
| 262 |
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
|
| 263 |
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
|
| 264 |
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
|
| 265 |
-
"model.layers.3.post_attention_layernorm.weight": "model-
|
| 266 |
"model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
|
| 267 |
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
|
| 268 |
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
|
|
@@ -302,11 +302,11 @@
|
|
| 302 |
"model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
| 303 |
"model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
| 304 |
"model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
| 305 |
-
"model.layers.33.input_layernorm.weight": "model-
|
| 306 |
"model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
|
| 307 |
"model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
|
| 308 |
"model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
|
| 309 |
-
"model.layers.33.post_attention_layernorm.weight": "model-
|
| 310 |
"model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
|
| 311 |
"model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
|
| 312 |
"model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
|
|
@@ -357,11 +357,11 @@
|
|
| 357 |
"model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
|
| 358 |
"model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
|
| 359 |
"model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
|
| 360 |
-
"model.layers.38.input_layernorm.weight": "model-
|
| 361 |
"model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
|
| 362 |
"model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
|
| 363 |
"model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
|
| 364 |
-
"model.layers.38.post_attention_layernorm.weight": "model-
|
| 365 |
"model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
|
| 366 |
"model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
|
| 367 |
"model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
|
|
@@ -423,11 +423,11 @@
|
|
| 423 |
"model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
|
| 424 |
"model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
|
| 425 |
"model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
|
| 426 |
-
"model.layers.43.input_layernorm.weight": "model-
|
| 427 |
"model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
|
| 428 |
"model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
|
| 429 |
"model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
|
| 430 |
-
"model.layers.43.post_attention_layernorm.weight": "model-
|
| 431 |
"model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
|
| 432 |
"model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
|
| 433 |
"model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
|
|
@@ -478,11 +478,11 @@
|
|
| 478 |
"model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
|
| 479 |
"model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
|
| 480 |
"model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
|
| 481 |
-
"model.layers.48.input_layernorm.weight": "model-
|
| 482 |
"model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
|
| 483 |
"model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
|
| 484 |
"model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
|
| 485 |
-
"model.layers.48.post_attention_layernorm.weight": "model-
|
| 486 |
"model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
|
| 487 |
"model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
|
| 488 |
"model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
|
|
@@ -544,11 +544,11 @@
|
|
| 544 |
"model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
|
| 545 |
"model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
|
| 546 |
"model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
|
| 547 |
-
"model.layers.53.input_layernorm.weight": "model-
|
| 548 |
"model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
|
| 549 |
"model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
|
| 550 |
"model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
|
| 551 |
-
"model.layers.53.post_attention_layernorm.weight": "model-
|
| 552 |
"model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
|
| 553 |
"model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
|
| 554 |
"model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
|
|
@@ -599,11 +599,11 @@
|
|
| 599 |
"model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
|
| 600 |
"model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
|
| 601 |
"model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
|
| 602 |
-
"model.layers.58.input_layernorm.weight": "model-
|
| 603 |
"model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
|
| 604 |
"model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
|
| 605 |
"model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
|
| 606 |
-
"model.layers.58.post_attention_layernorm.weight": "model-
|
| 607 |
"model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
|
| 608 |
"model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
|
| 609 |
"model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
|
|
@@ -665,11 +665,11 @@
|
|
| 665 |
"model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
|
| 666 |
"model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
|
| 667 |
"model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
|
| 668 |
-
"model.layers.63.input_layernorm.weight": "model-
|
| 669 |
"model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
|
| 670 |
"model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
|
| 671 |
"model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
|
| 672 |
-
"model.layers.63.post_attention_layernorm.weight": "model-
|
| 673 |
"model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
|
| 674 |
"model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
|
| 675 |
"model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
|
|
@@ -687,11 +687,11 @@
|
|
| 687 |
"model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
|
| 688 |
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
|
| 689 |
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
|
| 690 |
-
"model.layers.8.input_layernorm.weight": "model-
|
| 691 |
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
|
| 692 |
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
|
| 693 |
"model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
|
| 694 |
-
"model.layers.8.post_attention_layernorm.weight": "model-
|
| 695 |
"model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
|
| 696 |
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
|
| 697 |
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",
|
|
|
|
| 60 |
"model.layers.12.self_attn.q_norm.weight": "model-00003-of-00014.safetensors",
|
| 61 |
"model.layers.12.self_attn.q_proj.weight": "model-00003-of-00014.safetensors",
|
| 62 |
"model.layers.12.self_attn.v_proj.weight": "model-00003-of-00014.safetensors",
|
| 63 |
+
"model.layers.13.input_layernorm.weight": "model-00004-of-00014.safetensors",
|
| 64 |
"model.layers.13.mlp.down_proj.weight": "model-00004-of-00014.safetensors",
|
| 65 |
"model.layers.13.mlp.gate_proj.weight": "model-00003-of-00014.safetensors",
|
| 66 |
"model.layers.13.mlp.up_proj.weight": "model-00004-of-00014.safetensors",
|
| 67 |
+
"model.layers.13.post_attention_layernorm.weight": "model-00004-of-00014.safetensors",
|
| 68 |
"model.layers.13.self_attn.k_norm.weight": "model-00003-of-00014.safetensors",
|
| 69 |
"model.layers.13.self_attn.k_proj.weight": "model-00003-of-00014.safetensors",
|
| 70 |
"model.layers.13.self_attn.o_proj.weight": "model-00003-of-00014.safetensors",
|
|
|
|
| 115 |
"model.layers.17.self_attn.q_norm.weight": "model-00004-of-00014.safetensors",
|
| 116 |
"model.layers.17.self_attn.q_proj.weight": "model-00004-of-00014.safetensors",
|
| 117 |
"model.layers.17.self_attn.v_proj.weight": "model-00004-of-00014.safetensors",
|
| 118 |
+
"model.layers.18.input_layernorm.weight": "model-00005-of-00014.safetensors",
|
| 119 |
"model.layers.18.mlp.down_proj.weight": "model-00005-of-00014.safetensors",
|
| 120 |
"model.layers.18.mlp.gate_proj.weight": "model-00004-of-00014.safetensors",
|
| 121 |
"model.layers.18.mlp.up_proj.weight": "model-00005-of-00014.safetensors",
|
| 122 |
+
"model.layers.18.post_attention_layernorm.weight": "model-00005-of-00014.safetensors",
|
| 123 |
"model.layers.18.self_attn.k_norm.weight": "model-00004-of-00014.safetensors",
|
| 124 |
"model.layers.18.self_attn.k_proj.weight": "model-00004-of-00014.safetensors",
|
| 125 |
"model.layers.18.self_attn.o_proj.weight": "model-00004-of-00014.safetensors",
|
|
|
|
| 181 |
"model.layers.22.self_attn.q_norm.weight": "model-00005-of-00014.safetensors",
|
| 182 |
"model.layers.22.self_attn.q_proj.weight": "model-00005-of-00014.safetensors",
|
| 183 |
"model.layers.22.self_attn.v_proj.weight": "model-00005-of-00014.safetensors",
|
| 184 |
+
"model.layers.23.input_layernorm.weight": "model-00006-of-00014.safetensors",
|
| 185 |
"model.layers.23.mlp.down_proj.weight": "model-00006-of-00014.safetensors",
|
| 186 |
"model.layers.23.mlp.gate_proj.weight": "model-00005-of-00014.safetensors",
|
| 187 |
"model.layers.23.mlp.up_proj.weight": "model-00006-of-00014.safetensors",
|
| 188 |
+
"model.layers.23.post_attention_layernorm.weight": "model-00006-of-00014.safetensors",
|
| 189 |
"model.layers.23.self_attn.k_norm.weight": "model-00005-of-00014.safetensors",
|
| 190 |
"model.layers.23.self_attn.k_proj.weight": "model-00005-of-00014.safetensors",
|
| 191 |
"model.layers.23.self_attn.o_proj.weight": "model-00005-of-00014.safetensors",
|
|
|
|
| 236 |
"model.layers.27.self_attn.q_norm.weight": "model-00006-of-00014.safetensors",
|
| 237 |
"model.layers.27.self_attn.q_proj.weight": "model-00006-of-00014.safetensors",
|
| 238 |
"model.layers.27.self_attn.v_proj.weight": "model-00006-of-00014.safetensors",
|
| 239 |
+
"model.layers.28.input_layernorm.weight": "model-00007-of-00014.safetensors",
|
| 240 |
"model.layers.28.mlp.down_proj.weight": "model-00007-of-00014.safetensors",
|
| 241 |
"model.layers.28.mlp.gate_proj.weight": "model-00006-of-00014.safetensors",
|
| 242 |
"model.layers.28.mlp.up_proj.weight": "model-00007-of-00014.safetensors",
|
| 243 |
+
"model.layers.28.post_attention_layernorm.weight": "model-00007-of-00014.safetensors",
|
| 244 |
"model.layers.28.self_attn.k_norm.weight": "model-00006-of-00014.safetensors",
|
| 245 |
"model.layers.28.self_attn.k_proj.weight": "model-00006-of-00014.safetensors",
|
| 246 |
"model.layers.28.self_attn.o_proj.weight": "model-00006-of-00014.safetensors",
|
|
|
|
| 258 |
"model.layers.29.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
| 259 |
"model.layers.29.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
| 260 |
"model.layers.29.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
| 261 |
+
"model.layers.3.input_layernorm.weight": "model-00002-of-00014.safetensors",
|
| 262 |
"model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors",
|
| 263 |
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00014.safetensors",
|
| 264 |
"model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors",
|
| 265 |
+
"model.layers.3.post_attention_layernorm.weight": "model-00002-of-00014.safetensors",
|
| 266 |
"model.layers.3.self_attn.k_norm.weight": "model-00001-of-00014.safetensors",
|
| 267 |
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00014.safetensors",
|
| 268 |
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00014.safetensors",
|
|
|
|
| 302 |
"model.layers.32.self_attn.q_norm.weight": "model-00007-of-00014.safetensors",
|
| 303 |
"model.layers.32.self_attn.q_proj.weight": "model-00007-of-00014.safetensors",
|
| 304 |
"model.layers.32.self_attn.v_proj.weight": "model-00007-of-00014.safetensors",
|
| 305 |
+
"model.layers.33.input_layernorm.weight": "model-00008-of-00014.safetensors",
|
| 306 |
"model.layers.33.mlp.down_proj.weight": "model-00008-of-00014.safetensors",
|
| 307 |
"model.layers.33.mlp.gate_proj.weight": "model-00007-of-00014.safetensors",
|
| 308 |
"model.layers.33.mlp.up_proj.weight": "model-00008-of-00014.safetensors",
|
| 309 |
+
"model.layers.33.post_attention_layernorm.weight": "model-00008-of-00014.safetensors",
|
| 310 |
"model.layers.33.self_attn.k_norm.weight": "model-00007-of-00014.safetensors",
|
| 311 |
"model.layers.33.self_attn.k_proj.weight": "model-00007-of-00014.safetensors",
|
| 312 |
"model.layers.33.self_attn.o_proj.weight": "model-00007-of-00014.safetensors",
|
|
|
|
| 357 |
"model.layers.37.self_attn.q_norm.weight": "model-00008-of-00014.safetensors",
|
| 358 |
"model.layers.37.self_attn.q_proj.weight": "model-00008-of-00014.safetensors",
|
| 359 |
"model.layers.37.self_attn.v_proj.weight": "model-00008-of-00014.safetensors",
|
| 360 |
+
"model.layers.38.input_layernorm.weight": "model-00009-of-00014.safetensors",
|
| 361 |
"model.layers.38.mlp.down_proj.weight": "model-00009-of-00014.safetensors",
|
| 362 |
"model.layers.38.mlp.gate_proj.weight": "model-00008-of-00014.safetensors",
|
| 363 |
"model.layers.38.mlp.up_proj.weight": "model-00009-of-00014.safetensors",
|
| 364 |
+
"model.layers.38.post_attention_layernorm.weight": "model-00009-of-00014.safetensors",
|
| 365 |
"model.layers.38.self_attn.k_norm.weight": "model-00008-of-00014.safetensors",
|
| 366 |
"model.layers.38.self_attn.k_proj.weight": "model-00008-of-00014.safetensors",
|
| 367 |
"model.layers.38.self_attn.o_proj.weight": "model-00008-of-00014.safetensors",
|
|
|
|
| 423 |
"model.layers.42.self_attn.q_norm.weight": "model-00009-of-00014.safetensors",
|
| 424 |
"model.layers.42.self_attn.q_proj.weight": "model-00009-of-00014.safetensors",
|
| 425 |
"model.layers.42.self_attn.v_proj.weight": "model-00009-of-00014.safetensors",
|
| 426 |
+
"model.layers.43.input_layernorm.weight": "model-00010-of-00014.safetensors",
|
| 427 |
"model.layers.43.mlp.down_proj.weight": "model-00010-of-00014.safetensors",
|
| 428 |
"model.layers.43.mlp.gate_proj.weight": "model-00009-of-00014.safetensors",
|
| 429 |
"model.layers.43.mlp.up_proj.weight": "model-00010-of-00014.safetensors",
|
| 430 |
+
"model.layers.43.post_attention_layernorm.weight": "model-00010-of-00014.safetensors",
|
| 431 |
"model.layers.43.self_attn.k_norm.weight": "model-00009-of-00014.safetensors",
|
| 432 |
"model.layers.43.self_attn.k_proj.weight": "model-00009-of-00014.safetensors",
|
| 433 |
"model.layers.43.self_attn.o_proj.weight": "model-00009-of-00014.safetensors",
|
|
|
|
| 478 |
"model.layers.47.self_attn.q_norm.weight": "model-00010-of-00014.safetensors",
|
| 479 |
"model.layers.47.self_attn.q_proj.weight": "model-00010-of-00014.safetensors",
|
| 480 |
"model.layers.47.self_attn.v_proj.weight": "model-00010-of-00014.safetensors",
|
| 481 |
+
"model.layers.48.input_layernorm.weight": "model-00011-of-00014.safetensors",
|
| 482 |
"model.layers.48.mlp.down_proj.weight": "model-00011-of-00014.safetensors",
|
| 483 |
"model.layers.48.mlp.gate_proj.weight": "model-00010-of-00014.safetensors",
|
| 484 |
"model.layers.48.mlp.up_proj.weight": "model-00011-of-00014.safetensors",
|
| 485 |
+
"model.layers.48.post_attention_layernorm.weight": "model-00011-of-00014.safetensors",
|
| 486 |
"model.layers.48.self_attn.k_norm.weight": "model-00010-of-00014.safetensors",
|
| 487 |
"model.layers.48.self_attn.k_proj.weight": "model-00010-of-00014.safetensors",
|
| 488 |
"model.layers.48.self_attn.o_proj.weight": "model-00010-of-00014.safetensors",
|
|
|
|
| 544 |
"model.layers.52.self_attn.q_norm.weight": "model-00011-of-00014.safetensors",
|
| 545 |
"model.layers.52.self_attn.q_proj.weight": "model-00011-of-00014.safetensors",
|
| 546 |
"model.layers.52.self_attn.v_proj.weight": "model-00011-of-00014.safetensors",
|
| 547 |
+
"model.layers.53.input_layernorm.weight": "model-00012-of-00014.safetensors",
|
| 548 |
"model.layers.53.mlp.down_proj.weight": "model-00012-of-00014.safetensors",
|
| 549 |
"model.layers.53.mlp.gate_proj.weight": "model-00011-of-00014.safetensors",
|
| 550 |
"model.layers.53.mlp.up_proj.weight": "model-00012-of-00014.safetensors",
|
| 551 |
+
"model.layers.53.post_attention_layernorm.weight": "model-00012-of-00014.safetensors",
|
| 552 |
"model.layers.53.self_attn.k_norm.weight": "model-00011-of-00014.safetensors",
|
| 553 |
"model.layers.53.self_attn.k_proj.weight": "model-00011-of-00014.safetensors",
|
| 554 |
"model.layers.53.self_attn.o_proj.weight": "model-00011-of-00014.safetensors",
|
|
|
|
| 599 |
"model.layers.57.self_attn.q_norm.weight": "model-00012-of-00014.safetensors",
|
| 600 |
"model.layers.57.self_attn.q_proj.weight": "model-00012-of-00014.safetensors",
|
| 601 |
"model.layers.57.self_attn.v_proj.weight": "model-00012-of-00014.safetensors",
|
| 602 |
+
"model.layers.58.input_layernorm.weight": "model-00013-of-00014.safetensors",
|
| 603 |
"model.layers.58.mlp.down_proj.weight": "model-00013-of-00014.safetensors",
|
| 604 |
"model.layers.58.mlp.gate_proj.weight": "model-00012-of-00014.safetensors",
|
| 605 |
"model.layers.58.mlp.up_proj.weight": "model-00013-of-00014.safetensors",
|
| 606 |
+
"model.layers.58.post_attention_layernorm.weight": "model-00013-of-00014.safetensors",
|
| 607 |
"model.layers.58.self_attn.k_norm.weight": "model-00012-of-00014.safetensors",
|
| 608 |
"model.layers.58.self_attn.k_proj.weight": "model-00012-of-00014.safetensors",
|
| 609 |
"model.layers.58.self_attn.o_proj.weight": "model-00012-of-00014.safetensors",
|
|
|
|
| 665 |
"model.layers.62.self_attn.q_norm.weight": "model-00013-of-00014.safetensors",
|
| 666 |
"model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors",
|
| 667 |
"model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors",
|
| 668 |
+
"model.layers.63.input_layernorm.weight": "model-00014-of-00014.safetensors",
|
| 669 |
"model.layers.63.mlp.down_proj.weight": "model-00014-of-00014.safetensors",
|
| 670 |
"model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors",
|
| 671 |
"model.layers.63.mlp.up_proj.weight": "model-00014-of-00014.safetensors",
|
| 672 |
+
"model.layers.63.post_attention_layernorm.weight": "model-00014-of-00014.safetensors",
|
| 673 |
"model.layers.63.self_attn.k_norm.weight": "model-00013-of-00014.safetensors",
|
| 674 |
"model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors",
|
| 675 |
"model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors",
|
|
|
|
| 687 |
"model.layers.7.self_attn.q_norm.weight": "model-00002-of-00014.safetensors",
|
| 688 |
"model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors",
|
| 689 |
"model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors",
|
| 690 |
+
"model.layers.8.input_layernorm.weight": "model-00003-of-00014.safetensors",
|
| 691 |
"model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors",
|
| 692 |
"model.layers.8.mlp.gate_proj.weight": "model-00002-of-00014.safetensors",
|
| 693 |
"model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors",
|
| 694 |
+
"model.layers.8.post_attention_layernorm.weight": "model-00003-of-00014.safetensors",
|
| 695 |
"model.layers.8.self_attn.k_norm.weight": "model-00002-of-00014.safetensors",
|
| 696 |
"model.layers.8.self_attn.k_proj.weight": "model-00002-of-00014.safetensors",
|
| 697 |
"model.layers.8.self_attn.o_proj.weight": "model-00002-of-00014.safetensors",
|