weiweiz1 commited on
Commit
81419bf
·
verified ·
1 Parent(s): 71abea5

Add files using upload-large-folder tool

Browse files
config.json CHANGED
@@ -25,10 +25,295 @@
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
27
  "quantization_config": {
28
- "autoround_version": "0.6.0",
29
  "bits": 4,
30
  "data_type": "int",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  "group_size": 128,
 
32
  "packing_format": "auto_round:auto_gptq",
33
  "quant_method": "auto-round",
34
  "sym": true
@@ -40,7 +325,7 @@
40
  "sliding_window": null,
41
  "tie_word_embeddings": false,
42
  "torch_dtype": "bfloat16",
43
- "transformers_version": "4.53.2",
44
  "use_cache": true,
45
  "use_sliding_window": false,
46
  "vocab_size": 151936
 
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
27
  "quantization_config": {
28
+ "autoround_version": "0.6.1.dev",
29
  "bits": 4,
30
  "data_type": "int",
31
+ "extra_config": {
32
+ "model.layers.0.mlp.gate": {
33
+ "bits": 16
34
+ },
35
+ "model.layers.1.mlp.gate": {
36
+ "bits": 16
37
+ },
38
+ "model.layers.10.mlp.gate": {
39
+ "bits": 16
40
+ },
41
+ "model.layers.11.mlp.gate": {
42
+ "bits": 16
43
+ },
44
+ "model.layers.12.mlp.gate": {
45
+ "bits": 16
46
+ },
47
+ "model.layers.13.mlp.gate": {
48
+ "bits": 16
49
+ },
50
+ "model.layers.14.mlp.gate": {
51
+ "bits": 16
52
+ },
53
+ "model.layers.15.mlp.gate": {
54
+ "bits": 16
55
+ },
56
+ "model.layers.16.mlp.gate": {
57
+ "bits": 16
58
+ },
59
+ "model.layers.17.mlp.gate": {
60
+ "bits": 16
61
+ },
62
+ "model.layers.18.mlp.gate": {
63
+ "bits": 16
64
+ },
65
+ "model.layers.19.mlp.gate": {
66
+ "bits": 16
67
+ },
68
+ "model.layers.2.mlp.gate": {
69
+ "bits": 16
70
+ },
71
+ "model.layers.20.mlp.gate": {
72
+ "bits": 16
73
+ },
74
+ "model.layers.21.mlp.gate": {
75
+ "bits": 16
76
+ },
77
+ "model.layers.22.mlp.gate": {
78
+ "bits": 16
79
+ },
80
+ "model.layers.23.mlp.gate": {
81
+ "bits": 16
82
+ },
83
+ "model.layers.24.mlp.gate": {
84
+ "bits": 16
85
+ },
86
+ "model.layers.25.mlp.gate": {
87
+ "bits": 16
88
+ },
89
+ "model.layers.26.mlp.gate": {
90
+ "bits": 16
91
+ },
92
+ "model.layers.27.mlp.gate": {
93
+ "bits": 16
94
+ },
95
+ "model.layers.28.mlp.gate": {
96
+ "bits": 16
97
+ },
98
+ "model.layers.29.mlp.gate": {
99
+ "bits": 16
100
+ },
101
+ "model.layers.3.mlp.gate": {
102
+ "bits": 16
103
+ },
104
+ "model.layers.30.mlp.gate": {
105
+ "bits": 16
106
+ },
107
+ "model.layers.31.mlp.gate": {
108
+ "bits": 16
109
+ },
110
+ "model.layers.32.mlp.gate": {
111
+ "bits": 16
112
+ },
113
+ "model.layers.33.mlp.gate": {
114
+ "bits": 16
115
+ },
116
+ "model.layers.34.mlp.gate": {
117
+ "bits": 16
118
+ },
119
+ "model.layers.35.mlp.gate": {
120
+ "bits": 16
121
+ },
122
+ "model.layers.36.mlp.gate": {
123
+ "bits": 16
124
+ },
125
+ "model.layers.37.mlp.gate": {
126
+ "bits": 16
127
+ },
128
+ "model.layers.38.mlp.gate": {
129
+ "bits": 16
130
+ },
131
+ "model.layers.39.mlp.gate": {
132
+ "bits": 16
133
+ },
134
+ "model.layers.4.mlp.gate": {
135
+ "bits": 16
136
+ },
137
+ "model.layers.40.mlp.gate": {
138
+ "bits": 16
139
+ },
140
+ "model.layers.41.mlp.gate": {
141
+ "bits": 16
142
+ },
143
+ "model.layers.42.mlp.gate": {
144
+ "bits": 16
145
+ },
146
+ "model.layers.43.mlp.gate": {
147
+ "bits": 16
148
+ },
149
+ "model.layers.44.mlp.gate": {
150
+ "bits": 16
151
+ },
152
+ "model.layers.45.mlp.gate": {
153
+ "bits": 16
154
+ },
155
+ "model.layers.46.mlp.gate": {
156
+ "bits": 16
157
+ },
158
+ "model.layers.47.mlp.gate": {
159
+ "bits": 16
160
+ },
161
+ "model.layers.48.mlp.gate": {
162
+ "bits": 16
163
+ },
164
+ "model.layers.49.mlp.gate": {
165
+ "bits": 16
166
+ },
167
+ "model.layers.5.mlp.gate": {
168
+ "bits": 16
169
+ },
170
+ "model.layers.50.mlp.gate": {
171
+ "bits": 16
172
+ },
173
+ "model.layers.51.mlp.gate": {
174
+ "bits": 16
175
+ },
176
+ "model.layers.52.mlp.gate": {
177
+ "bits": 16
178
+ },
179
+ "model.layers.53.mlp.gate": {
180
+ "bits": 16
181
+ },
182
+ "model.layers.54.mlp.gate": {
183
+ "bits": 16
184
+ },
185
+ "model.layers.55.mlp.gate": {
186
+ "bits": 16
187
+ },
188
+ "model.layers.56.mlp.gate": {
189
+ "bits": 16
190
+ },
191
+ "model.layers.57.mlp.gate": {
192
+ "bits": 16
193
+ },
194
+ "model.layers.58.mlp.gate": {
195
+ "bits": 16
196
+ },
197
+ "model.layers.59.mlp.gate": {
198
+ "bits": 16
199
+ },
200
+ "model.layers.6.mlp.gate": {
201
+ "bits": 16
202
+ },
203
+ "model.layers.60.mlp.gate": {
204
+ "bits": 16
205
+ },
206
+ "model.layers.61.mlp.gate": {
207
+ "bits": 16
208
+ },
209
+ "model.layers.62.mlp.gate": {
210
+ "bits": 16
211
+ },
212
+ "model.layers.63.mlp.gate": {
213
+ "bits": 16
214
+ },
215
+ "model.layers.64.mlp.gate": {
216
+ "bits": 16
217
+ },
218
+ "model.layers.65.mlp.gate": {
219
+ "bits": 16
220
+ },
221
+ "model.layers.66.mlp.gate": {
222
+ "bits": 16
223
+ },
224
+ "model.layers.67.mlp.gate": {
225
+ "bits": 16
226
+ },
227
+ "model.layers.68.mlp.gate": {
228
+ "bits": 16
229
+ },
230
+ "model.layers.69.mlp.gate": {
231
+ "bits": 16
232
+ },
233
+ "model.layers.7.mlp.gate": {
234
+ "bits": 16
235
+ },
236
+ "model.layers.70.mlp.gate": {
237
+ "bits": 16
238
+ },
239
+ "model.layers.71.mlp.gate": {
240
+ "bits": 16
241
+ },
242
+ "model.layers.72.mlp.gate": {
243
+ "bits": 16
244
+ },
245
+ "model.layers.73.mlp.gate": {
246
+ "bits": 16
247
+ },
248
+ "model.layers.74.mlp.gate": {
249
+ "bits": 16
250
+ },
251
+ "model.layers.75.mlp.gate": {
252
+ "bits": 16
253
+ },
254
+ "model.layers.76.mlp.gate": {
255
+ "bits": 16
256
+ },
257
+ "model.layers.77.mlp.gate": {
258
+ "bits": 16
259
+ },
260
+ "model.layers.78.mlp.gate": {
261
+ "bits": 16
262
+ },
263
+ "model.layers.79.mlp.gate": {
264
+ "bits": 16
265
+ },
266
+ "model.layers.8.mlp.gate": {
267
+ "bits": 16
268
+ },
269
+ "model.layers.80.mlp.gate": {
270
+ "bits": 16
271
+ },
272
+ "model.layers.81.mlp.gate": {
273
+ "bits": 16
274
+ },
275
+ "model.layers.82.mlp.gate": {
276
+ "bits": 16
277
+ },
278
+ "model.layers.83.mlp.gate": {
279
+ "bits": 16
280
+ },
281
+ "model.layers.84.mlp.gate": {
282
+ "bits": 16
283
+ },
284
+ "model.layers.85.mlp.gate": {
285
+ "bits": 16
286
+ },
287
+ "model.layers.86.mlp.gate": {
288
+ "bits": 16
289
+ },
290
+ "model.layers.87.mlp.gate": {
291
+ "bits": 16
292
+ },
293
+ "model.layers.88.mlp.gate": {
294
+ "bits": 16
295
+ },
296
+ "model.layers.89.mlp.gate": {
297
+ "bits": 16
298
+ },
299
+ "model.layers.9.mlp.gate": {
300
+ "bits": 16
301
+ },
302
+ "model.layers.90.mlp.gate": {
303
+ "bits": 16
304
+ },
305
+ "model.layers.91.mlp.gate": {
306
+ "bits": 16
307
+ },
308
+ "model.layers.92.mlp.gate": {
309
+ "bits": 16
310
+ },
311
+ "model.layers.93.mlp.gate": {
312
+ "bits": 16
313
+ }
314
+ },
315
  "group_size": 128,
316
+ "nsamples": 512,
317
  "packing_format": "auto_round:auto_gptq",
318
  "quant_method": "auto-round",
319
  "sym": true
 
325
  "sliding_window": null,
326
  "tie_word_embeddings": false,
327
  "torch_dtype": "bfloat16",
328
+ "transformers_version": "4.54.0",
329
  "use_cache": true,
330
  "use_sliding_window": false,
331
  "vocab_size": 151936
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "temperature": 0.7,
10
  "top_k": 20,
11
  "top_p": 0.8,
12
- "transformers_version": "4.53.2"
13
  }
 
9
  "temperature": 0.7,
10
  "top_k": 20,
11
  "top_p": 0.8,
12
+ "transformers_version": "4.54.0"
13
  }
model-00001-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b0ec85884991ad8e315ce79e67a290c0998c0f1c9e7a38842e6aa1266ffcf28
3
- size 4998285064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:592a8a958f2adc893a1199bebcc6c70f3f8ed0dc0be430ce88bf14b9cbe6a715
3
+ size 4997344032
model-00002-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:122fce75c964782050a992397151a4b4b9c0e6500ca6118aa49c707efa9c1a01
3
- size 5000497640
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abb0dc689c547a9f5325c3b31c42a541fe9223798d922c3b42a7fc8a2ae3d58a
3
+ size 5000332616
model-00003-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:278c26744904ed5c930a71e2d61d029af3618e21565002f20ebd63a5235029cd
3
- size 5000498544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1f9bbedaffeb716f0a67c8c85084e7feda3a5ef5da99e724cf10f026d852ca8
3
+ size 5000333504
model-00004-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a08ca1e7885d25217a5c74b9598973babd5871e4e307ced1d039c642ed541a09
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:325ce9e112004b953fda5ef05243dd439996b3e0b3caff01638a5147f05de692
3
+ size 5000337264
model-00005-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:548c56573d735fdf20f8697170e4a1b3b0e9fc390271cd2091de399f3597bfaf
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:005aeae6d1db78bb34b0f1167c0755eafe3fcc1eef0cc20f715e74ec3d99e9a1
3
+ size 5000337264
model-00006-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab6a5b23f702297c10e17815b0dc5b1c5044babfa392bbee2ae84607b0c7db4e
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c6ebbdcefeeccf9e0e315cba7a9d66dae617992470e0ec534662d2f3456c70f
3
+ size 5000337256
model-00007-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca5b854912017a74f718ab094b093b1594a5097af919888b99491f285c9031a9
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5feee975f7bd87034570a0d525db963943b334a85dc38c6d2f2401cee46e84f
3
+ size 5000337272
model-00008-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7837a55d4d878884f1a4e4ed7eaf1a4e4064a4bc26253355698810d801737840
3
- size 4999125448
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf6a86a87783dc3fcaa41f96451e502d8876e728d16dc5aafb50f17f47a9707
3
+ size 4998184392
model-00009-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:068cb1c9d38e2741d99fdac0df2577469de823673e78fcafd64e06a649b4bf01
3
- size 5000502160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac85f57a31b0c2ae36ec5d4361330f1a55be17d33cd57ecb80137d6678ab231
3
+ size 5000337104
model-00010-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eadfaa351a789cf49a5a3185c0a653c98b612733b4407ff6fbcc053f18889a9b
3
- size 5000502232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b5392bceb89dd3e826ac69c1f0d012ee2b8262aedac446b2c4bd115fe0b10b8
3
+ size 5000337208
model-00011-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:163dddcf2f55272f8bc01321e034ce7e45cdfe7d082069c8eb67021a266d907c
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f569d80f1b9947f952f2e9c1937c5d1339e0b4028ee18b84a41863a2d423735
3
+ size 5000337264
model-00012-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7911b8b2767f6e902bb437c2d39465246c538885f36bed7c663f73ac0b31f51f
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3824ccc3f522667e3a2365e40c461121720e2ad12151255777921cc8cf1871ba
3
+ size 5000337264
model-00013-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fa1b8f3009fba90d18f196c9f2f3fca42016741fd8c7cba722581aca380bcd6
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f16b43f2e05efaa48c8645e13667266f403c01dd626834b2e1ee9bb1d22dc93
3
+ size 5000337256
model-00014-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ae2d3202715979ab97b8f197dbde12b0742b6501455a2597fbe6c26e78d0cee
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8787abfef427d16681903c6a4765eb5a3cf9855c795bb319c15d614fcf1b44dd
3
+ size 5000337264
model-00015-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62f5772f4a5c01740c4f8ab41b9861cf21af21bcd37edaa2479e21cf25fb88df
3
- size 5000502368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:549610ccdea282f69863a43229e127f39f62f27a36eb39753516c524681e940f
3
+ size 4988392960
model-00016-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f55a85d05363614606edfef5203d5d0bf30cf9e86c388825795b1f89026c359
3
- size 4999125336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba44f31f41cc338caa41573c3cb03e9a9593d8d7e4e1ce78b82b629aac755efe
3
+ size 5000321680
model-00017-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ed3d486fb601cb1307be60f648056497652439550cd0e24d83e0c8563bb6fd4
3
- size 5000502160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb21d1ed3d9cbd003a2e0647467f759152461c099f06acce6dcd7895749e88e9
3
+ size 5000337144
model-00018-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3255e3f4f42c620d8a8255f6d07408672e0ced23c4eec2e48014d4a0a04e774
3
- size 5000502296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f31c75adcfc003bbbf81575d8716124d3355839b910507fe038d81f4b97443de
3
+ size 5000337264
model-00019-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45a6e06e92a152e44b8e3c15a8376c04294187f0cc916adfbfa2facbd96c831c
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f63295d198e6c28266bfe99789873441d690a43872a2154a22d0c78f91af5be
3
+ size 5000337264
model-00020-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b8553e66167d7c7a2cf63e5662a1f7183d362213131c0da9eaa2d3657cc4d12
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b832e4e969afc1ed60ef99f92c3427371a22ac94ab86323916f2df7e8243a7b
3
+ size 5000337256
model-00021-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5813f24dca5e3cf6daae2d43ba39ee0be74f55e17abd529885c6da3d5458c9a1
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f72e4e9da544ddb984005ea006204c1ed69610807dbf292ac566e864b9a612f
3
+ size 5000337264
model-00022-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fbb59a93f405a3b0d6492ce03cfc97c3cb8a4c8826881fd47cac5fa5115704e
3
- size 5000502312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ffbd7c3cf5fb8e3505d46a10ce34ebaae91f0f7c4390b7a122e742cc99153ee
3
+ size 5000337296
model-00023-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab3b731f338747a70768dc46a721709148b0dcd97f2c9b3b33c06acf241903f1
3
- size 4989335168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adeb101a806674decaa8491bbd38a9d04be459a93cee4927ba511e2d8de79a8b
3
+ size 4998184344
model-00024-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:46bca499693020a1c68269681eded1a3abb05a0328e9380138b70ca05aeb78bc
3
- size 5000485536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2c1d4a68954a8ceccca0cbabe473424fcd9aea71f4237f533670edb39c6fc1
3
+ size 5000337096
model-00025-of-00025.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:070ae1ffc7fa96e901044324ad6ffb4684f9dd97f81fd651dac5861da5157cd3
3
- size 3999915464
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91585050b17f1a4151b4e29781c0e4e724f9090efb05b8da58a629970867b7d7
3
+ size 4079922928
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
quantization_config.json CHANGED
@@ -3,7 +3,292 @@
3
  "group_size": 128,
4
  "sym": true,
5
  "data_type": "int",
6
- "autoround_version": "0.6.0",
 
7
  "quant_method": "auto-round",
8
- "packing_format": "auto_round:auto_gptq"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  }
 
3
  "group_size": 128,
4
  "sym": true,
5
  "data_type": "int",
6
+ "nsamples": 512,
7
+ "autoround_version": "0.6.1.dev",
8
  "quant_method": "auto-round",
9
+ "packing_format": "auto_round:auto_gptq",
10
+ "extra_config": {
11
+ "model.layers.0.mlp.gate": {
12
+ "bits": 16
13
+ },
14
+ "model.layers.1.mlp.gate": {
15
+ "bits": 16
16
+ },
17
+ "model.layers.2.mlp.gate": {
18
+ "bits": 16
19
+ },
20
+ "model.layers.3.mlp.gate": {
21
+ "bits": 16
22
+ },
23
+ "model.layers.4.mlp.gate": {
24
+ "bits": 16
25
+ },
26
+ "model.layers.5.mlp.gate": {
27
+ "bits": 16
28
+ },
29
+ "model.layers.6.mlp.gate": {
30
+ "bits": 16
31
+ },
32
+ "model.layers.7.mlp.gate": {
33
+ "bits": 16
34
+ },
35
+ "model.layers.8.mlp.gate": {
36
+ "bits": 16
37
+ },
38
+ "model.layers.9.mlp.gate": {
39
+ "bits": 16
40
+ },
41
+ "model.layers.10.mlp.gate": {
42
+ "bits": 16
43
+ },
44
+ "model.layers.11.mlp.gate": {
45
+ "bits": 16
46
+ },
47
+ "model.layers.12.mlp.gate": {
48
+ "bits": 16
49
+ },
50
+ "model.layers.13.mlp.gate": {
51
+ "bits": 16
52
+ },
53
+ "model.layers.14.mlp.gate": {
54
+ "bits": 16
55
+ },
56
+ "model.layers.15.mlp.gate": {
57
+ "bits": 16
58
+ },
59
+ "model.layers.16.mlp.gate": {
60
+ "bits": 16
61
+ },
62
+ "model.layers.17.mlp.gate": {
63
+ "bits": 16
64
+ },
65
+ "model.layers.18.mlp.gate": {
66
+ "bits": 16
67
+ },
68
+ "model.layers.19.mlp.gate": {
69
+ "bits": 16
70
+ },
71
+ "model.layers.20.mlp.gate": {
72
+ "bits": 16
73
+ },
74
+ "model.layers.21.mlp.gate": {
75
+ "bits": 16
76
+ },
77
+ "model.layers.22.mlp.gate": {
78
+ "bits": 16
79
+ },
80
+ "model.layers.23.mlp.gate": {
81
+ "bits": 16
82
+ },
83
+ "model.layers.24.mlp.gate": {
84
+ "bits": 16
85
+ },
86
+ "model.layers.25.mlp.gate": {
87
+ "bits": 16
88
+ },
89
+ "model.layers.26.mlp.gate": {
90
+ "bits": 16
91
+ },
92
+ "model.layers.27.mlp.gate": {
93
+ "bits": 16
94
+ },
95
+ "model.layers.28.mlp.gate": {
96
+ "bits": 16
97
+ },
98
+ "model.layers.29.mlp.gate": {
99
+ "bits": 16
100
+ },
101
+ "model.layers.30.mlp.gate": {
102
+ "bits": 16
103
+ },
104
+ "model.layers.31.mlp.gate": {
105
+ "bits": 16
106
+ },
107
+ "model.layers.32.mlp.gate": {
108
+ "bits": 16
109
+ },
110
+ "model.layers.33.mlp.gate": {
111
+ "bits": 16
112
+ },
113
+ "model.layers.34.mlp.gate": {
114
+ "bits": 16
115
+ },
116
+ "model.layers.35.mlp.gate": {
117
+ "bits": 16
118
+ },
119
+ "model.layers.36.mlp.gate": {
120
+ "bits": 16
121
+ },
122
+ "model.layers.37.mlp.gate": {
123
+ "bits": 16
124
+ },
125
+ "model.layers.38.mlp.gate": {
126
+ "bits": 16
127
+ },
128
+ "model.layers.39.mlp.gate": {
129
+ "bits": 16
130
+ },
131
+ "model.layers.40.mlp.gate": {
132
+ "bits": 16
133
+ },
134
+ "model.layers.41.mlp.gate": {
135
+ "bits": 16
136
+ },
137
+ "model.layers.42.mlp.gate": {
138
+ "bits": 16
139
+ },
140
+ "model.layers.43.mlp.gate": {
141
+ "bits": 16
142
+ },
143
+ "model.layers.44.mlp.gate": {
144
+ "bits": 16
145
+ },
146
+ "model.layers.45.mlp.gate": {
147
+ "bits": 16
148
+ },
149
+ "model.layers.46.mlp.gate": {
150
+ "bits": 16
151
+ },
152
+ "model.layers.47.mlp.gate": {
153
+ "bits": 16
154
+ },
155
+ "model.layers.48.mlp.gate": {
156
+ "bits": 16
157
+ },
158
+ "model.layers.49.mlp.gate": {
159
+ "bits": 16
160
+ },
161
+ "model.layers.50.mlp.gate": {
162
+ "bits": 16
163
+ },
164
+ "model.layers.51.mlp.gate": {
165
+ "bits": 16
166
+ },
167
+ "model.layers.52.mlp.gate": {
168
+ "bits": 16
169
+ },
170
+ "model.layers.53.mlp.gate": {
171
+ "bits": 16
172
+ },
173
+ "model.layers.54.mlp.gate": {
174
+ "bits": 16
175
+ },
176
+ "model.layers.55.mlp.gate": {
177
+ "bits": 16
178
+ },
179
+ "model.layers.56.mlp.gate": {
180
+ "bits": 16
181
+ },
182
+ "model.layers.57.mlp.gate": {
183
+ "bits": 16
184
+ },
185
+ "model.layers.58.mlp.gate": {
186
+ "bits": 16
187
+ },
188
+ "model.layers.59.mlp.gate": {
189
+ "bits": 16
190
+ },
191
+ "model.layers.60.mlp.gate": {
192
+ "bits": 16
193
+ },
194
+ "model.layers.61.mlp.gate": {
195
+ "bits": 16
196
+ },
197
+ "model.layers.62.mlp.gate": {
198
+ "bits": 16
199
+ },
200
+ "model.layers.63.mlp.gate": {
201
+ "bits": 16
202
+ },
203
+ "model.layers.64.mlp.gate": {
204
+ "bits": 16
205
+ },
206
+ "model.layers.65.mlp.gate": {
207
+ "bits": 16
208
+ },
209
+ "model.layers.66.mlp.gate": {
210
+ "bits": 16
211
+ },
212
+ "model.layers.67.mlp.gate": {
213
+ "bits": 16
214
+ },
215
+ "model.layers.68.mlp.gate": {
216
+ "bits": 16
217
+ },
218
+ "model.layers.69.mlp.gate": {
219
+ "bits": 16
220
+ },
221
+ "model.layers.70.mlp.gate": {
222
+ "bits": 16
223
+ },
224
+ "model.layers.71.mlp.gate": {
225
+ "bits": 16
226
+ },
227
+ "model.layers.72.mlp.gate": {
228
+ "bits": 16
229
+ },
230
+ "model.layers.73.mlp.gate": {
231
+ "bits": 16
232
+ },
233
+ "model.layers.74.mlp.gate": {
234
+ "bits": 16
235
+ },
236
+ "model.layers.75.mlp.gate": {
237
+ "bits": 16
238
+ },
239
+ "model.layers.76.mlp.gate": {
240
+ "bits": 16
241
+ },
242
+ "model.layers.77.mlp.gate": {
243
+ "bits": 16
244
+ },
245
+ "model.layers.78.mlp.gate": {
246
+ "bits": 16
247
+ },
248
+ "model.layers.79.mlp.gate": {
249
+ "bits": 16
250
+ },
251
+ "model.layers.80.mlp.gate": {
252
+ "bits": 16
253
+ },
254
+ "model.layers.81.mlp.gate": {
255
+ "bits": 16
256
+ },
257
+ "model.layers.82.mlp.gate": {
258
+ "bits": 16
259
+ },
260
+ "model.layers.83.mlp.gate": {
261
+ "bits": 16
262
+ },
263
+ "model.layers.84.mlp.gate": {
264
+ "bits": 16
265
+ },
266
+ "model.layers.85.mlp.gate": {
267
+ "bits": 16
268
+ },
269
+ "model.layers.86.mlp.gate": {
270
+ "bits": 16
271
+ },
272
+ "model.layers.87.mlp.gate": {
273
+ "bits": 16
274
+ },
275
+ "model.layers.88.mlp.gate": {
276
+ "bits": 16
277
+ },
278
+ "model.layers.89.mlp.gate": {
279
+ "bits": 16
280
+ },
281
+ "model.layers.90.mlp.gate": {
282
+ "bits": 16
283
+ },
284
+ "model.layers.91.mlp.gate": {
285
+ "bits": 16
286
+ },
287
+ "model.layers.92.mlp.gate": {
288
+ "bits": 16
289
+ },
290
+ "model.layers.93.mlp.gate": {
291
+ "bits": 16
292
+ }
293
+ }
294
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
- size 11422654
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0acdaba32b920d640afb36af4396c91974e074735636e4016d17a8ed9c03730
3
+ size 11422753