n1ck-guo commited on
Commit
9142816
·
verified ·
1 Parent(s): 62d5300

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. config.json +522 -0
  2. model-00002-of-00072.safetensors +2 -2
  3. model-00003-of-00072.safetensors +2 -2
  4. model-00004-of-00072.safetensors +2 -2
  5. model-00005-of-00072.safetensors +2 -2
  6. model-00006-of-00072.safetensors +2 -2
  7. model-00007-of-00072.safetensors +2 -2
  8. model-00008-of-00072.safetensors +2 -2
  9. model-00009-of-00072.safetensors +2 -2
  10. model-00010-of-00072.safetensors +2 -2
  11. model-00011-of-00072.safetensors +2 -2
  12. model-00012-of-00072.safetensors +2 -2
  13. model-00013-of-00072.safetensors +2 -2
  14. model-00014-of-00072.safetensors +2 -2
  15. model-00015-of-00072.safetensors +2 -2
  16. model-00016-of-00072.safetensors +2 -2
  17. model-00017-of-00072.safetensors +2 -2
  18. model-00018-of-00072.safetensors +2 -2
  19. model-00019-of-00072.safetensors +2 -2
  20. model-00020-of-00072.safetensors +2 -2
  21. model-00021-of-00072.safetensors +2 -2
  22. model-00022-of-00072.safetensors +2 -2
  23. model-00023-of-00072.safetensors +2 -2
  24. model-00024-of-00072.safetensors +2 -2
  25. model-00025-of-00072.safetensors +2 -2
  26. model-00026-of-00072.safetensors +2 -2
  27. model-00027-of-00072.safetensors +2 -2
  28. model-00028-of-00072.safetensors +2 -2
  29. model-00029-of-00072.safetensors +2 -2
  30. model-00030-of-00072.safetensors +2 -2
  31. model-00031-of-00072.safetensors +2 -2
  32. model-00032-of-00072.safetensors +2 -2
  33. model-00033-of-00072.safetensors +2 -2
  34. model-00034-of-00072.safetensors +2 -2
  35. model-00035-of-00072.safetensors +2 -2
  36. model-00036-of-00072.safetensors +2 -2
  37. model-00037-of-00072.safetensors +2 -2
  38. model-00038-of-00072.safetensors +2 -2
  39. model-00039-of-00072.safetensors +2 -2
  40. model-00040-of-00072.safetensors +2 -2
  41. model-00041-of-00072.safetensors +2 -2
  42. model-00042-of-00072.safetensors +2 -2
  43. model-00043-of-00072.safetensors +2 -2
  44. model-00044-of-00072.safetensors +2 -2
  45. model-00045-of-00072.safetensors +2 -2
  46. model-00046-of-00072.safetensors +2 -2
  47. model-00047-of-00072.safetensors +2 -2
  48. model-00048-of-00072.safetensors +2 -2
  49. model-00049-of-00072.safetensors +2 -2
  50. model-00050-of-00072.safetensors +2 -2
config.json CHANGED
@@ -88,6 +88,15 @@
88
  "model.layers.1.self_attn.q_b_proj": {
89
  "bits": 8
90
  },
 
 
 
 
 
 
 
 
 
91
  "model.layers.10.self_attn.kv_a_proj_with_mqa": {
92
  "bits": 8
93
  },
@@ -103,6 +112,15 @@
103
  "model.layers.10.self_attn.q_b_proj": {
104
  "bits": 8
105
  },
 
 
 
 
 
 
 
 
 
106
  "model.layers.11.self_attn.kv_a_proj_with_mqa": {
107
  "bits": 8
108
  },
@@ -118,6 +136,15 @@
118
  "model.layers.11.self_attn.q_b_proj": {
119
  "bits": 8
120
  },
 
 
 
 
 
 
 
 
 
121
  "model.layers.12.self_attn.kv_a_proj_with_mqa": {
122
  "bits": 8
123
  },
@@ -133,6 +160,15 @@
133
  "model.layers.12.self_attn.q_b_proj": {
134
  "bits": 8
135
  },
 
 
 
 
 
 
 
 
 
136
  "model.layers.13.self_attn.kv_a_proj_with_mqa": {
137
  "bits": 8
138
  },
@@ -148,6 +184,15 @@
148
  "model.layers.13.self_attn.q_b_proj": {
149
  "bits": 8
150
  },
 
 
 
 
 
 
 
 
 
151
  "model.layers.14.self_attn.kv_a_proj_with_mqa": {
152
  "bits": 8
153
  },
@@ -163,6 +208,15 @@
163
  "model.layers.14.self_attn.q_b_proj": {
164
  "bits": 8
165
  },
 
 
 
 
 
 
 
 
 
166
  "model.layers.15.self_attn.kv_a_proj_with_mqa": {
167
  "bits": 8
168
  },
@@ -178,6 +232,15 @@
178
  "model.layers.15.self_attn.q_b_proj": {
179
  "bits": 8
180
  },
 
 
 
 
 
 
 
 
 
181
  "model.layers.16.self_attn.kv_a_proj_with_mqa": {
182
  "bits": 8
183
  },
@@ -193,6 +256,15 @@
193
  "model.layers.16.self_attn.q_b_proj": {
194
  "bits": 8
195
  },
 
 
 
 
 
 
 
 
 
196
  "model.layers.17.self_attn.kv_a_proj_with_mqa": {
197
  "bits": 8
198
  },
@@ -208,6 +280,15 @@
208
  "model.layers.17.self_attn.q_b_proj": {
209
  "bits": 8
210
  },
 
 
 
 
 
 
 
 
 
211
  "model.layers.18.self_attn.kv_a_proj_with_mqa": {
212
  "bits": 8
213
  },
@@ -223,6 +304,15 @@
223
  "model.layers.18.self_attn.q_b_proj": {
224
  "bits": 8
225
  },
 
 
 
 
 
 
 
 
 
226
  "model.layers.19.self_attn.kv_a_proj_with_mqa": {
227
  "bits": 8
228
  },
@@ -262,6 +352,15 @@
262
  "model.layers.2.self_attn.q_b_proj": {
263
  "bits": 8
264
  },
 
 
 
 
 
 
 
 
 
265
  "model.layers.20.self_attn.kv_a_proj_with_mqa": {
266
  "bits": 8
267
  },
@@ -277,6 +376,15 @@
277
  "model.layers.20.self_attn.q_b_proj": {
278
  "bits": 8
279
  },
 
 
 
 
 
 
 
 
 
280
  "model.layers.21.self_attn.kv_a_proj_with_mqa": {
281
  "bits": 8
282
  },
@@ -292,6 +400,15 @@
292
  "model.layers.21.self_attn.q_b_proj": {
293
  "bits": 8
294
  },
 
 
 
 
 
 
 
 
 
295
  "model.layers.22.self_attn.kv_a_proj_with_mqa": {
296
  "bits": 8
297
  },
@@ -307,6 +424,15 @@
307
  "model.layers.22.self_attn.q_b_proj": {
308
  "bits": 8
309
  },
 
 
 
 
 
 
 
 
 
310
  "model.layers.23.self_attn.kv_a_proj_with_mqa": {
311
  "bits": 8
312
  },
@@ -322,6 +448,15 @@
322
  "model.layers.23.self_attn.q_b_proj": {
323
  "bits": 8
324
  },
 
 
 
 
 
 
 
 
 
325
  "model.layers.24.self_attn.kv_a_proj_with_mqa": {
326
  "bits": 8
327
  },
@@ -337,6 +472,15 @@
337
  "model.layers.24.self_attn.q_b_proj": {
338
  "bits": 8
339
  },
 
 
 
 
 
 
 
 
 
340
  "model.layers.25.self_attn.kv_a_proj_with_mqa": {
341
  "bits": 8
342
  },
@@ -352,6 +496,15 @@
352
  "model.layers.25.self_attn.q_b_proj": {
353
  "bits": 8
354
  },
 
 
 
 
 
 
 
 
 
355
  "model.layers.26.self_attn.kv_a_proj_with_mqa": {
356
  "bits": 8
357
  },
@@ -367,6 +520,15 @@
367
  "model.layers.26.self_attn.q_b_proj": {
368
  "bits": 8
369
  },
 
 
 
 
 
 
 
 
 
370
  "model.layers.27.self_attn.kv_a_proj_with_mqa": {
371
  "bits": 8
372
  },
@@ -382,6 +544,15 @@
382
  "model.layers.27.self_attn.q_b_proj": {
383
  "bits": 8
384
  },
 
 
 
 
 
 
 
 
 
385
  "model.layers.28.self_attn.kv_a_proj_with_mqa": {
386
  "bits": 8
387
  },
@@ -397,6 +568,15 @@
397
  "model.layers.28.self_attn.q_b_proj": {
398
  "bits": 8
399
  },
 
 
 
 
 
 
 
 
 
400
  "model.layers.29.self_attn.kv_a_proj_with_mqa": {
401
  "bits": 8
402
  },
@@ -412,6 +592,15 @@
412
  "model.layers.29.self_attn.q_b_proj": {
413
  "bits": 8
414
  },
 
 
 
 
 
 
 
 
 
415
  "model.layers.3.self_attn.kv_a_proj_with_mqa": {
416
  "bits": 8
417
  },
@@ -427,6 +616,15 @@
427
  "model.layers.3.self_attn.q_b_proj": {
428
  "bits": 8
429
  },
 
 
 
 
 
 
 
 
 
430
  "model.layers.30.self_attn.kv_a_proj_with_mqa": {
431
  "bits": 8
432
  },
@@ -442,6 +640,15 @@
442
  "model.layers.30.self_attn.q_b_proj": {
443
  "bits": 8
444
  },
 
 
 
 
 
 
 
 
 
445
  "model.layers.31.self_attn.kv_a_proj_with_mqa": {
446
  "bits": 8
447
  },
@@ -457,6 +664,15 @@
457
  "model.layers.31.self_attn.q_b_proj": {
458
  "bits": 8
459
  },
 
 
 
 
 
 
 
 
 
460
  "model.layers.32.self_attn.kv_a_proj_with_mqa": {
461
  "bits": 8
462
  },
@@ -472,6 +688,15 @@
472
  "model.layers.32.self_attn.q_b_proj": {
473
  "bits": 8
474
  },
 
 
 
 
 
 
 
 
 
475
  "model.layers.33.self_attn.kv_a_proj_with_mqa": {
476
  "bits": 8
477
  },
@@ -487,6 +712,15 @@
487
  "model.layers.33.self_attn.q_b_proj": {
488
  "bits": 8
489
  },
 
 
 
 
 
 
 
 
 
490
  "model.layers.34.self_attn.kv_a_proj_with_mqa": {
491
  "bits": 8
492
  },
@@ -502,6 +736,15 @@
502
  "model.layers.34.self_attn.q_b_proj": {
503
  "bits": 8
504
  },
 
 
 
 
 
 
 
 
 
505
  "model.layers.35.self_attn.kv_a_proj_with_mqa": {
506
  "bits": 8
507
  },
@@ -517,6 +760,15 @@
517
  "model.layers.35.self_attn.q_b_proj": {
518
  "bits": 8
519
  },
 
 
 
 
 
 
 
 
 
520
  "model.layers.36.self_attn.kv_a_proj_with_mqa": {
521
  "bits": 8
522
  },
@@ -532,6 +784,15 @@
532
  "model.layers.36.self_attn.q_b_proj": {
533
  "bits": 8
534
  },
 
 
 
 
 
 
 
 
 
535
  "model.layers.37.self_attn.kv_a_proj_with_mqa": {
536
  "bits": 8
537
  },
@@ -547,6 +808,15 @@
547
  "model.layers.37.self_attn.q_b_proj": {
548
  "bits": 8
549
  },
 
 
 
 
 
 
 
 
 
550
  "model.layers.38.self_attn.kv_a_proj_with_mqa": {
551
  "bits": 8
552
  },
@@ -562,6 +832,15 @@
562
  "model.layers.38.self_attn.q_b_proj": {
563
  "bits": 8
564
  },
 
 
 
 
 
 
 
 
 
565
  "model.layers.39.self_attn.kv_a_proj_with_mqa": {
566
  "bits": 8
567
  },
@@ -577,6 +856,15 @@
577
  "model.layers.39.self_attn.q_b_proj": {
578
  "bits": 8
579
  },
 
 
 
 
 
 
 
 
 
580
  "model.layers.4.self_attn.kv_a_proj_with_mqa": {
581
  "bits": 8
582
  },
@@ -592,6 +880,15 @@
592
  "model.layers.4.self_attn.q_b_proj": {
593
  "bits": 8
594
  },
 
 
 
 
 
 
 
 
 
595
  "model.layers.40.self_attn.kv_a_proj_with_mqa": {
596
  "bits": 8
597
  },
@@ -607,6 +904,15 @@
607
  "model.layers.40.self_attn.q_b_proj": {
608
  "bits": 8
609
  },
 
 
 
 
 
 
 
 
 
610
  "model.layers.41.self_attn.kv_a_proj_with_mqa": {
611
  "bits": 8
612
  },
@@ -622,6 +928,15 @@
622
  "model.layers.41.self_attn.q_b_proj": {
623
  "bits": 8
624
  },
 
 
 
 
 
 
 
 
 
625
  "model.layers.42.self_attn.kv_a_proj_with_mqa": {
626
  "bits": 8
627
  },
@@ -637,6 +952,15 @@
637
  "model.layers.42.self_attn.q_b_proj": {
638
  "bits": 8
639
  },
 
 
 
 
 
 
 
 
 
640
  "model.layers.43.self_attn.kv_a_proj_with_mqa": {
641
  "bits": 8
642
  },
@@ -652,6 +976,15 @@
652
  "model.layers.43.self_attn.q_b_proj": {
653
  "bits": 8
654
  },
 
 
 
 
 
 
 
 
 
655
  "model.layers.44.self_attn.kv_a_proj_with_mqa": {
656
  "bits": 8
657
  },
@@ -667,6 +1000,15 @@
667
  "model.layers.44.self_attn.q_b_proj": {
668
  "bits": 8
669
  },
 
 
 
 
 
 
 
 
 
670
  "model.layers.45.self_attn.kv_a_proj_with_mqa": {
671
  "bits": 8
672
  },
@@ -682,6 +1024,15 @@
682
  "model.layers.45.self_attn.q_b_proj": {
683
  "bits": 8
684
  },
 
 
 
 
 
 
 
 
 
685
  "model.layers.46.self_attn.kv_a_proj_with_mqa": {
686
  "bits": 8
687
  },
@@ -697,6 +1048,15 @@
697
  "model.layers.46.self_attn.q_b_proj": {
698
  "bits": 8
699
  },
 
 
 
 
 
 
 
 
 
700
  "model.layers.47.self_attn.kv_a_proj_with_mqa": {
701
  "bits": 8
702
  },
@@ -712,6 +1072,15 @@
712
  "model.layers.47.self_attn.q_b_proj": {
713
  "bits": 8
714
  },
 
 
 
 
 
 
 
 
 
715
  "model.layers.48.self_attn.kv_a_proj_with_mqa": {
716
  "bits": 8
717
  },
@@ -727,6 +1096,15 @@
727
  "model.layers.48.self_attn.q_b_proj": {
728
  "bits": 8
729
  },
 
 
 
 
 
 
 
 
 
730
  "model.layers.49.self_attn.kv_a_proj_with_mqa": {
731
  "bits": 8
732
  },
@@ -742,6 +1120,15 @@
742
  "model.layers.49.self_attn.q_b_proj": {
743
  "bits": 8
744
  },
 
 
 
 
 
 
 
 
 
745
  "model.layers.5.self_attn.kv_a_proj_with_mqa": {
746
  "bits": 8
747
  },
@@ -757,6 +1144,15 @@
757
  "model.layers.5.self_attn.q_b_proj": {
758
  "bits": 8
759
  },
 
 
 
 
 
 
 
 
 
760
  "model.layers.50.self_attn.kv_a_proj_with_mqa": {
761
  "bits": 8
762
  },
@@ -772,6 +1168,15 @@
772
  "model.layers.50.self_attn.q_b_proj": {
773
  "bits": 8
774
  },
 
 
 
 
 
 
 
 
 
775
  "model.layers.51.self_attn.kv_a_proj_with_mqa": {
776
  "bits": 8
777
  },
@@ -787,6 +1192,15 @@
787
  "model.layers.51.self_attn.q_b_proj": {
788
  "bits": 8
789
  },
 
 
 
 
 
 
 
 
 
790
  "model.layers.52.self_attn.kv_a_proj_with_mqa": {
791
  "bits": 8
792
  },
@@ -802,6 +1216,15 @@
802
  "model.layers.52.self_attn.q_b_proj": {
803
  "bits": 8
804
  },
 
 
 
 
 
 
 
 
 
805
  "model.layers.53.self_attn.kv_a_proj_with_mqa": {
806
  "bits": 8
807
  },
@@ -817,6 +1240,15 @@
817
  "model.layers.53.self_attn.q_b_proj": {
818
  "bits": 8
819
  },
 
 
 
 
 
 
 
 
 
820
  "model.layers.54.self_attn.kv_a_proj_with_mqa": {
821
  "bits": 8
822
  },
@@ -832,6 +1264,15 @@
832
  "model.layers.54.self_attn.q_b_proj": {
833
  "bits": 8
834
  },
 
 
 
 
 
 
 
 
 
835
  "model.layers.55.self_attn.kv_a_proj_with_mqa": {
836
  "bits": 8
837
  },
@@ -847,6 +1288,15 @@
847
  "model.layers.55.self_attn.q_b_proj": {
848
  "bits": 8
849
  },
 
 
 
 
 
 
 
 
 
850
  "model.layers.56.self_attn.kv_a_proj_with_mqa": {
851
  "bits": 8
852
  },
@@ -862,6 +1312,15 @@
862
  "model.layers.56.self_attn.q_b_proj": {
863
  "bits": 8
864
  },
 
 
 
 
 
 
 
 
 
865
  "model.layers.57.self_attn.kv_a_proj_with_mqa": {
866
  "bits": 8
867
  },
@@ -877,6 +1336,15 @@
877
  "model.layers.57.self_attn.q_b_proj": {
878
  "bits": 8
879
  },
 
 
 
 
 
 
 
 
 
880
  "model.layers.58.self_attn.kv_a_proj_with_mqa": {
881
  "bits": 8
882
  },
@@ -892,6 +1360,15 @@
892
  "model.layers.58.self_attn.q_b_proj": {
893
  "bits": 8
894
  },
 
 
 
 
 
 
 
 
 
895
  "model.layers.59.self_attn.kv_a_proj_with_mqa": {
896
  "bits": 8
897
  },
@@ -907,6 +1384,15 @@
907
  "model.layers.59.self_attn.q_b_proj": {
908
  "bits": 8
909
  },
 
 
 
 
 
 
 
 
 
910
  "model.layers.6.self_attn.kv_a_proj_with_mqa": {
911
  "bits": 8
912
  },
@@ -922,6 +1408,15 @@
922
  "model.layers.6.self_attn.q_b_proj": {
923
  "bits": 8
924
  },
 
 
 
 
 
 
 
 
 
925
  "model.layers.60.self_attn.kv_a_proj_with_mqa": {
926
  "bits": 8
927
  },
@@ -937,6 +1432,15 @@
937
  "model.layers.60.self_attn.q_b_proj": {
938
  "bits": 8
939
  },
 
 
 
 
 
 
 
 
 
940
  "model.layers.7.self_attn.kv_a_proj_with_mqa": {
941
  "bits": 8
942
  },
@@ -952,6 +1456,15 @@
952
  "model.layers.7.self_attn.q_b_proj": {
953
  "bits": 8
954
  },
 
 
 
 
 
 
 
 
 
955
  "model.layers.8.self_attn.kv_a_proj_with_mqa": {
956
  "bits": 8
957
  },
@@ -967,6 +1480,15 @@
967
  "model.layers.8.self_attn.q_b_proj": {
968
  "bits": 8
969
  },
 
 
 
 
 
 
 
 
 
970
  "model.layers.9.self_attn.kv_a_proj_with_mqa": {
971
  "bits": 8
972
  },
 
88
  "model.layers.1.self_attn.q_b_proj": {
89
  "bits": 8
90
  },
91
+ "model.layers.10.mlp.shared_experts.down_proj": {
92
+ "bits": 8
93
+ },
94
+ "model.layers.10.mlp.shared_experts.gate_proj": {
95
+ "bits": 8
96
+ },
97
+ "model.layers.10.mlp.shared_experts.up_proj": {
98
+ "bits": 8
99
+ },
100
  "model.layers.10.self_attn.kv_a_proj_with_mqa": {
101
  "bits": 8
102
  },
 
112
  "model.layers.10.self_attn.q_b_proj": {
113
  "bits": 8
114
  },
115
+ "model.layers.11.mlp.shared_experts.down_proj": {
116
+ "bits": 8
117
+ },
118
+ "model.layers.11.mlp.shared_experts.gate_proj": {
119
+ "bits": 8
120
+ },
121
+ "model.layers.11.mlp.shared_experts.up_proj": {
122
+ "bits": 8
123
+ },
124
  "model.layers.11.self_attn.kv_a_proj_with_mqa": {
125
  "bits": 8
126
  },
 
136
  "model.layers.11.self_attn.q_b_proj": {
137
  "bits": 8
138
  },
139
+ "model.layers.12.mlp.shared_experts.down_proj": {
140
+ "bits": 8
141
+ },
142
+ "model.layers.12.mlp.shared_experts.gate_proj": {
143
+ "bits": 8
144
+ },
145
+ "model.layers.12.mlp.shared_experts.up_proj": {
146
+ "bits": 8
147
+ },
148
  "model.layers.12.self_attn.kv_a_proj_with_mqa": {
149
  "bits": 8
150
  },
 
160
  "model.layers.12.self_attn.q_b_proj": {
161
  "bits": 8
162
  },
163
+ "model.layers.13.mlp.shared_experts.down_proj": {
164
+ "bits": 8
165
+ },
166
+ "model.layers.13.mlp.shared_experts.gate_proj": {
167
+ "bits": 8
168
+ },
169
+ "model.layers.13.mlp.shared_experts.up_proj": {
170
+ "bits": 8
171
+ },
172
  "model.layers.13.self_attn.kv_a_proj_with_mqa": {
173
  "bits": 8
174
  },
 
184
  "model.layers.13.self_attn.q_b_proj": {
185
  "bits": 8
186
  },
187
+ "model.layers.14.mlp.shared_experts.down_proj": {
188
+ "bits": 8
189
+ },
190
+ "model.layers.14.mlp.shared_experts.gate_proj": {
191
+ "bits": 8
192
+ },
193
+ "model.layers.14.mlp.shared_experts.up_proj": {
194
+ "bits": 8
195
+ },
196
  "model.layers.14.self_attn.kv_a_proj_with_mqa": {
197
  "bits": 8
198
  },
 
208
  "model.layers.14.self_attn.q_b_proj": {
209
  "bits": 8
210
  },
211
+ "model.layers.15.mlp.shared_experts.down_proj": {
212
+ "bits": 8
213
+ },
214
+ "model.layers.15.mlp.shared_experts.gate_proj": {
215
+ "bits": 8
216
+ },
217
+ "model.layers.15.mlp.shared_experts.up_proj": {
218
+ "bits": 8
219
+ },
220
  "model.layers.15.self_attn.kv_a_proj_with_mqa": {
221
  "bits": 8
222
  },
 
232
  "model.layers.15.self_attn.q_b_proj": {
233
  "bits": 8
234
  },
235
+ "model.layers.16.mlp.shared_experts.down_proj": {
236
+ "bits": 8
237
+ },
238
+ "model.layers.16.mlp.shared_experts.gate_proj": {
239
+ "bits": 8
240
+ },
241
+ "model.layers.16.mlp.shared_experts.up_proj": {
242
+ "bits": 8
243
+ },
244
  "model.layers.16.self_attn.kv_a_proj_with_mqa": {
245
  "bits": 8
246
  },
 
256
  "model.layers.16.self_attn.q_b_proj": {
257
  "bits": 8
258
  },
259
+ "model.layers.17.mlp.shared_experts.down_proj": {
260
+ "bits": 8
261
+ },
262
+ "model.layers.17.mlp.shared_experts.gate_proj": {
263
+ "bits": 8
264
+ },
265
+ "model.layers.17.mlp.shared_experts.up_proj": {
266
+ "bits": 8
267
+ },
268
  "model.layers.17.self_attn.kv_a_proj_with_mqa": {
269
  "bits": 8
270
  },
 
280
  "model.layers.17.self_attn.q_b_proj": {
281
  "bits": 8
282
  },
283
+ "model.layers.18.mlp.shared_experts.down_proj": {
284
+ "bits": 8
285
+ },
286
+ "model.layers.18.mlp.shared_experts.gate_proj": {
287
+ "bits": 8
288
+ },
289
+ "model.layers.18.mlp.shared_experts.up_proj": {
290
+ "bits": 8
291
+ },
292
  "model.layers.18.self_attn.kv_a_proj_with_mqa": {
293
  "bits": 8
294
  },
 
304
  "model.layers.18.self_attn.q_b_proj": {
305
  "bits": 8
306
  },
307
+ "model.layers.19.mlp.shared_experts.down_proj": {
308
+ "bits": 8
309
+ },
310
+ "model.layers.19.mlp.shared_experts.gate_proj": {
311
+ "bits": 8
312
+ },
313
+ "model.layers.19.mlp.shared_experts.up_proj": {
314
+ "bits": 8
315
+ },
316
  "model.layers.19.self_attn.kv_a_proj_with_mqa": {
317
  "bits": 8
318
  },
 
352
  "model.layers.2.self_attn.q_b_proj": {
353
  "bits": 8
354
  },
355
+ "model.layers.20.mlp.shared_experts.down_proj": {
356
+ "bits": 8
357
+ },
358
+ "model.layers.20.mlp.shared_experts.gate_proj": {
359
+ "bits": 8
360
+ },
361
+ "model.layers.20.mlp.shared_experts.up_proj": {
362
+ "bits": 8
363
+ },
364
  "model.layers.20.self_attn.kv_a_proj_with_mqa": {
365
  "bits": 8
366
  },
 
376
  "model.layers.20.self_attn.q_b_proj": {
377
  "bits": 8
378
  },
379
+ "model.layers.21.mlp.shared_experts.down_proj": {
380
+ "bits": 8
381
+ },
382
+ "model.layers.21.mlp.shared_experts.gate_proj": {
383
+ "bits": 8
384
+ },
385
+ "model.layers.21.mlp.shared_experts.up_proj": {
386
+ "bits": 8
387
+ },
388
  "model.layers.21.self_attn.kv_a_proj_with_mqa": {
389
  "bits": 8
390
  },
 
400
  "model.layers.21.self_attn.q_b_proj": {
401
  "bits": 8
402
  },
403
+ "model.layers.22.mlp.shared_experts.down_proj": {
404
+ "bits": 8
405
+ },
406
+ "model.layers.22.mlp.shared_experts.gate_proj": {
407
+ "bits": 8
408
+ },
409
+ "model.layers.22.mlp.shared_experts.up_proj": {
410
+ "bits": 8
411
+ },
412
  "model.layers.22.self_attn.kv_a_proj_with_mqa": {
413
  "bits": 8
414
  },
 
424
  "model.layers.22.self_attn.q_b_proj": {
425
  "bits": 8
426
  },
427
+ "model.layers.23.mlp.shared_experts.down_proj": {
428
+ "bits": 8
429
+ },
430
+ "model.layers.23.mlp.shared_experts.gate_proj": {
431
+ "bits": 8
432
+ },
433
+ "model.layers.23.mlp.shared_experts.up_proj": {
434
+ "bits": 8
435
+ },
436
  "model.layers.23.self_attn.kv_a_proj_with_mqa": {
437
  "bits": 8
438
  },
 
448
  "model.layers.23.self_attn.q_b_proj": {
449
  "bits": 8
450
  },
451
+ "model.layers.24.mlp.shared_experts.down_proj": {
452
+ "bits": 8
453
+ },
454
+ "model.layers.24.mlp.shared_experts.gate_proj": {
455
+ "bits": 8
456
+ },
457
+ "model.layers.24.mlp.shared_experts.up_proj": {
458
+ "bits": 8
459
+ },
460
  "model.layers.24.self_attn.kv_a_proj_with_mqa": {
461
  "bits": 8
462
  },
 
472
  "model.layers.24.self_attn.q_b_proj": {
473
  "bits": 8
474
  },
475
+ "model.layers.25.mlp.shared_experts.down_proj": {
476
+ "bits": 8
477
+ },
478
+ "model.layers.25.mlp.shared_experts.gate_proj": {
479
+ "bits": 8
480
+ },
481
+ "model.layers.25.mlp.shared_experts.up_proj": {
482
+ "bits": 8
483
+ },
484
  "model.layers.25.self_attn.kv_a_proj_with_mqa": {
485
  "bits": 8
486
  },
 
496
  "model.layers.25.self_attn.q_b_proj": {
497
  "bits": 8
498
  },
499
+ "model.layers.26.mlp.shared_experts.down_proj": {
500
+ "bits": 8
501
+ },
502
+ "model.layers.26.mlp.shared_experts.gate_proj": {
503
+ "bits": 8
504
+ },
505
+ "model.layers.26.mlp.shared_experts.up_proj": {
506
+ "bits": 8
507
+ },
508
  "model.layers.26.self_attn.kv_a_proj_with_mqa": {
509
  "bits": 8
510
  },
 
520
  "model.layers.26.self_attn.q_b_proj": {
521
  "bits": 8
522
  },
523
+ "model.layers.27.mlp.shared_experts.down_proj": {
524
+ "bits": 8
525
+ },
526
+ "model.layers.27.mlp.shared_experts.gate_proj": {
527
+ "bits": 8
528
+ },
529
+ "model.layers.27.mlp.shared_experts.up_proj": {
530
+ "bits": 8
531
+ },
532
  "model.layers.27.self_attn.kv_a_proj_with_mqa": {
533
  "bits": 8
534
  },
 
544
  "model.layers.27.self_attn.q_b_proj": {
545
  "bits": 8
546
  },
547
+ "model.layers.28.mlp.shared_experts.down_proj": {
548
+ "bits": 8
549
+ },
550
+ "model.layers.28.mlp.shared_experts.gate_proj": {
551
+ "bits": 8
552
+ },
553
+ "model.layers.28.mlp.shared_experts.up_proj": {
554
+ "bits": 8
555
+ },
556
  "model.layers.28.self_attn.kv_a_proj_with_mqa": {
557
  "bits": 8
558
  },
 
568
  "model.layers.28.self_attn.q_b_proj": {
569
  "bits": 8
570
  },
571
+ "model.layers.29.mlp.shared_experts.down_proj": {
572
+ "bits": 8
573
+ },
574
+ "model.layers.29.mlp.shared_experts.gate_proj": {
575
+ "bits": 8
576
+ },
577
+ "model.layers.29.mlp.shared_experts.up_proj": {
578
+ "bits": 8
579
+ },
580
  "model.layers.29.self_attn.kv_a_proj_with_mqa": {
581
  "bits": 8
582
  },
 
592
  "model.layers.29.self_attn.q_b_proj": {
593
  "bits": 8
594
  },
595
+ "model.layers.3.mlp.shared_experts.down_proj": {
596
+ "bits": 8
597
+ },
598
+ "model.layers.3.mlp.shared_experts.gate_proj": {
599
+ "bits": 8
600
+ },
601
+ "model.layers.3.mlp.shared_experts.up_proj": {
602
+ "bits": 8
603
+ },
604
  "model.layers.3.self_attn.kv_a_proj_with_mqa": {
605
  "bits": 8
606
  },
 
616
  "model.layers.3.self_attn.q_b_proj": {
617
  "bits": 8
618
  },
619
+ "model.layers.30.mlp.shared_experts.down_proj": {
620
+ "bits": 8
621
+ },
622
+ "model.layers.30.mlp.shared_experts.gate_proj": {
623
+ "bits": 8
624
+ },
625
+ "model.layers.30.mlp.shared_experts.up_proj": {
626
+ "bits": 8
627
+ },
628
  "model.layers.30.self_attn.kv_a_proj_with_mqa": {
629
  "bits": 8
630
  },
 
640
  "model.layers.30.self_attn.q_b_proj": {
641
  "bits": 8
642
  },
643
+ "model.layers.31.mlp.shared_experts.down_proj": {
644
+ "bits": 8
645
+ },
646
+ "model.layers.31.mlp.shared_experts.gate_proj": {
647
+ "bits": 8
648
+ },
649
+ "model.layers.31.mlp.shared_experts.up_proj": {
650
+ "bits": 8
651
+ },
652
  "model.layers.31.self_attn.kv_a_proj_with_mqa": {
653
  "bits": 8
654
  },
 
664
  "model.layers.31.self_attn.q_b_proj": {
665
  "bits": 8
666
  },
667
+ "model.layers.32.mlp.shared_experts.down_proj": {
668
+ "bits": 8
669
+ },
670
+ "model.layers.32.mlp.shared_experts.gate_proj": {
671
+ "bits": 8
672
+ },
673
+ "model.layers.32.mlp.shared_experts.up_proj": {
674
+ "bits": 8
675
+ },
676
  "model.layers.32.self_attn.kv_a_proj_with_mqa": {
677
  "bits": 8
678
  },
 
688
  "model.layers.32.self_attn.q_b_proj": {
689
  "bits": 8
690
  },
691
+ "model.layers.33.mlp.shared_experts.down_proj": {
692
+ "bits": 8
693
+ },
694
+ "model.layers.33.mlp.shared_experts.gate_proj": {
695
+ "bits": 8
696
+ },
697
+ "model.layers.33.mlp.shared_experts.up_proj": {
698
+ "bits": 8
699
+ },
700
  "model.layers.33.self_attn.kv_a_proj_with_mqa": {
701
  "bits": 8
702
  },
 
712
  "model.layers.33.self_attn.q_b_proj": {
713
  "bits": 8
714
  },
715
+ "model.layers.34.mlp.shared_experts.down_proj": {
716
+ "bits": 8
717
+ },
718
+ "model.layers.34.mlp.shared_experts.gate_proj": {
719
+ "bits": 8
720
+ },
721
+ "model.layers.34.mlp.shared_experts.up_proj": {
722
+ "bits": 8
723
+ },
724
  "model.layers.34.self_attn.kv_a_proj_with_mqa": {
725
  "bits": 8
726
  },
 
736
  "model.layers.34.self_attn.q_b_proj": {
737
  "bits": 8
738
  },
739
+ "model.layers.35.mlp.shared_experts.down_proj": {
740
+ "bits": 8
741
+ },
742
+ "model.layers.35.mlp.shared_experts.gate_proj": {
743
+ "bits": 8
744
+ },
745
+ "model.layers.35.mlp.shared_experts.up_proj": {
746
+ "bits": 8
747
+ },
748
  "model.layers.35.self_attn.kv_a_proj_with_mqa": {
749
  "bits": 8
750
  },
 
760
  "model.layers.35.self_attn.q_b_proj": {
761
  "bits": 8
762
  },
763
+ "model.layers.36.mlp.shared_experts.down_proj": {
764
+ "bits": 8
765
+ },
766
+ "model.layers.36.mlp.shared_experts.gate_proj": {
767
+ "bits": 8
768
+ },
769
+ "model.layers.36.mlp.shared_experts.up_proj": {
770
+ "bits": 8
771
+ },
772
  "model.layers.36.self_attn.kv_a_proj_with_mqa": {
773
  "bits": 8
774
  },
 
784
  "model.layers.36.self_attn.q_b_proj": {
785
  "bits": 8
786
  },
787
+ "model.layers.37.mlp.shared_experts.down_proj": {
788
+ "bits": 8
789
+ },
790
+ "model.layers.37.mlp.shared_experts.gate_proj": {
791
+ "bits": 8
792
+ },
793
+ "model.layers.37.mlp.shared_experts.up_proj": {
794
+ "bits": 8
795
+ },
796
  "model.layers.37.self_attn.kv_a_proj_with_mqa": {
797
  "bits": 8
798
  },
 
808
  "model.layers.37.self_attn.q_b_proj": {
809
  "bits": 8
810
  },
811
+ "model.layers.38.mlp.shared_experts.down_proj": {
812
+ "bits": 8
813
+ },
814
+ "model.layers.38.mlp.shared_experts.gate_proj": {
815
+ "bits": 8
816
+ },
817
+ "model.layers.38.mlp.shared_experts.up_proj": {
818
+ "bits": 8
819
+ },
820
  "model.layers.38.self_attn.kv_a_proj_with_mqa": {
821
  "bits": 8
822
  },
 
832
  "model.layers.38.self_attn.q_b_proj": {
833
  "bits": 8
834
  },
835
+ "model.layers.39.mlp.shared_experts.down_proj": {
836
+ "bits": 8
837
+ },
838
+ "model.layers.39.mlp.shared_experts.gate_proj": {
839
+ "bits": 8
840
+ },
841
+ "model.layers.39.mlp.shared_experts.up_proj": {
842
+ "bits": 8
843
+ },
844
  "model.layers.39.self_attn.kv_a_proj_with_mqa": {
845
  "bits": 8
846
  },
 
856
  "model.layers.39.self_attn.q_b_proj": {
857
  "bits": 8
858
  },
859
+ "model.layers.4.mlp.shared_experts.down_proj": {
860
+ "bits": 8
861
+ },
862
+ "model.layers.4.mlp.shared_experts.gate_proj": {
863
+ "bits": 8
864
+ },
865
+ "model.layers.4.mlp.shared_experts.up_proj": {
866
+ "bits": 8
867
+ },
868
  "model.layers.4.self_attn.kv_a_proj_with_mqa": {
869
  "bits": 8
870
  },
 
880
  "model.layers.4.self_attn.q_b_proj": {
881
  "bits": 8
882
  },
883
+ "model.layers.40.mlp.shared_experts.down_proj": {
884
+ "bits": 8
885
+ },
886
+ "model.layers.40.mlp.shared_experts.gate_proj": {
887
+ "bits": 8
888
+ },
889
+ "model.layers.40.mlp.shared_experts.up_proj": {
890
+ "bits": 8
891
+ },
892
  "model.layers.40.self_attn.kv_a_proj_with_mqa": {
893
  "bits": 8
894
  },
 
904
  "model.layers.40.self_attn.q_b_proj": {
905
  "bits": 8
906
  },
907
+ "model.layers.41.mlp.shared_experts.down_proj": {
908
+ "bits": 8
909
+ },
910
+ "model.layers.41.mlp.shared_experts.gate_proj": {
911
+ "bits": 8
912
+ },
913
+ "model.layers.41.mlp.shared_experts.up_proj": {
914
+ "bits": 8
915
+ },
916
  "model.layers.41.self_attn.kv_a_proj_with_mqa": {
917
  "bits": 8
918
  },
 
928
  "model.layers.41.self_attn.q_b_proj": {
929
  "bits": 8
930
  },
931
+ "model.layers.42.mlp.shared_experts.down_proj": {
932
+ "bits": 8
933
+ },
934
+ "model.layers.42.mlp.shared_experts.gate_proj": {
935
+ "bits": 8
936
+ },
937
+ "model.layers.42.mlp.shared_experts.up_proj": {
938
+ "bits": 8
939
+ },
940
  "model.layers.42.self_attn.kv_a_proj_with_mqa": {
941
  "bits": 8
942
  },
 
952
  "model.layers.42.self_attn.q_b_proj": {
953
  "bits": 8
954
  },
955
+ "model.layers.43.mlp.shared_experts.down_proj": {
956
+ "bits": 8
957
+ },
958
+ "model.layers.43.mlp.shared_experts.gate_proj": {
959
+ "bits": 8
960
+ },
961
+ "model.layers.43.mlp.shared_experts.up_proj": {
962
+ "bits": 8
963
+ },
964
  "model.layers.43.self_attn.kv_a_proj_with_mqa": {
965
  "bits": 8
966
  },
 
976
  "model.layers.43.self_attn.q_b_proj": {
977
  "bits": 8
978
  },
979
+ "model.layers.44.mlp.shared_experts.down_proj": {
980
+ "bits": 8
981
+ },
982
+ "model.layers.44.mlp.shared_experts.gate_proj": {
983
+ "bits": 8
984
+ },
985
+ "model.layers.44.mlp.shared_experts.up_proj": {
986
+ "bits": 8
987
+ },
988
  "model.layers.44.self_attn.kv_a_proj_with_mqa": {
989
  "bits": 8
990
  },
 
1000
  "model.layers.44.self_attn.q_b_proj": {
1001
  "bits": 8
1002
  },
1003
+ "model.layers.45.mlp.shared_experts.down_proj": {
1004
+ "bits": 8
1005
+ },
1006
+ "model.layers.45.mlp.shared_experts.gate_proj": {
1007
+ "bits": 8
1008
+ },
1009
+ "model.layers.45.mlp.shared_experts.up_proj": {
1010
+ "bits": 8
1011
+ },
1012
  "model.layers.45.self_attn.kv_a_proj_with_mqa": {
1013
  "bits": 8
1014
  },
 
1024
  "model.layers.45.self_attn.q_b_proj": {
1025
  "bits": 8
1026
  },
1027
+ "model.layers.46.mlp.shared_experts.down_proj": {
1028
+ "bits": 8
1029
+ },
1030
+ "model.layers.46.mlp.shared_experts.gate_proj": {
1031
+ "bits": 8
1032
+ },
1033
+ "model.layers.46.mlp.shared_experts.up_proj": {
1034
+ "bits": 8
1035
+ },
1036
  "model.layers.46.self_attn.kv_a_proj_with_mqa": {
1037
  "bits": 8
1038
  },
 
1048
  "model.layers.46.self_attn.q_b_proj": {
1049
  "bits": 8
1050
  },
1051
+ "model.layers.47.mlp.shared_experts.down_proj": {
1052
+ "bits": 8
1053
+ },
1054
+ "model.layers.47.mlp.shared_experts.gate_proj": {
1055
+ "bits": 8
1056
+ },
1057
+ "model.layers.47.mlp.shared_experts.up_proj": {
1058
+ "bits": 8
1059
+ },
1060
  "model.layers.47.self_attn.kv_a_proj_with_mqa": {
1061
  "bits": 8
1062
  },
 
1072
  "model.layers.47.self_attn.q_b_proj": {
1073
  "bits": 8
1074
  },
1075
+ "model.layers.48.mlp.shared_experts.down_proj": {
1076
+ "bits": 8
1077
+ },
1078
+ "model.layers.48.mlp.shared_experts.gate_proj": {
1079
+ "bits": 8
1080
+ },
1081
+ "model.layers.48.mlp.shared_experts.up_proj": {
1082
+ "bits": 8
1083
+ },
1084
  "model.layers.48.self_attn.kv_a_proj_with_mqa": {
1085
  "bits": 8
1086
  },
 
1096
  "model.layers.48.self_attn.q_b_proj": {
1097
  "bits": 8
1098
  },
1099
+ "model.layers.49.mlp.shared_experts.down_proj": {
1100
+ "bits": 8
1101
+ },
1102
+ "model.layers.49.mlp.shared_experts.gate_proj": {
1103
+ "bits": 8
1104
+ },
1105
+ "model.layers.49.mlp.shared_experts.up_proj": {
1106
+ "bits": 8
1107
+ },
1108
  "model.layers.49.self_attn.kv_a_proj_with_mqa": {
1109
  "bits": 8
1110
  },
 
1120
  "model.layers.49.self_attn.q_b_proj": {
1121
  "bits": 8
1122
  },
1123
+ "model.layers.5.mlp.shared_experts.down_proj": {
1124
+ "bits": 8
1125
+ },
1126
+ "model.layers.5.mlp.shared_experts.gate_proj": {
1127
+ "bits": 8
1128
+ },
1129
+ "model.layers.5.mlp.shared_experts.up_proj": {
1130
+ "bits": 8
1131
+ },
1132
  "model.layers.5.self_attn.kv_a_proj_with_mqa": {
1133
  "bits": 8
1134
  },
 
1144
  "model.layers.5.self_attn.q_b_proj": {
1145
  "bits": 8
1146
  },
1147
+ "model.layers.50.mlp.shared_experts.down_proj": {
1148
+ "bits": 8
1149
+ },
1150
+ "model.layers.50.mlp.shared_experts.gate_proj": {
1151
+ "bits": 8
1152
+ },
1153
+ "model.layers.50.mlp.shared_experts.up_proj": {
1154
+ "bits": 8
1155
+ },
1156
  "model.layers.50.self_attn.kv_a_proj_with_mqa": {
1157
  "bits": 8
1158
  },
 
1168
  "model.layers.50.self_attn.q_b_proj": {
1169
  "bits": 8
1170
  },
1171
+ "model.layers.51.mlp.shared_experts.down_proj": {
1172
+ "bits": 8
1173
+ },
1174
+ "model.layers.51.mlp.shared_experts.gate_proj": {
1175
+ "bits": 8
1176
+ },
1177
+ "model.layers.51.mlp.shared_experts.up_proj": {
1178
+ "bits": 8
1179
+ },
1180
  "model.layers.51.self_attn.kv_a_proj_with_mqa": {
1181
  "bits": 8
1182
  },
 
1192
  "model.layers.51.self_attn.q_b_proj": {
1193
  "bits": 8
1194
  },
1195
+ "model.layers.52.mlp.shared_experts.down_proj": {
1196
+ "bits": 8
1197
+ },
1198
+ "model.layers.52.mlp.shared_experts.gate_proj": {
1199
+ "bits": 8
1200
+ },
1201
+ "model.layers.52.mlp.shared_experts.up_proj": {
1202
+ "bits": 8
1203
+ },
1204
  "model.layers.52.self_attn.kv_a_proj_with_mqa": {
1205
  "bits": 8
1206
  },
 
1216
  "model.layers.52.self_attn.q_b_proj": {
1217
  "bits": 8
1218
  },
1219
+ "model.layers.53.mlp.shared_experts.down_proj": {
1220
+ "bits": 8
1221
+ },
1222
+ "model.layers.53.mlp.shared_experts.gate_proj": {
1223
+ "bits": 8
1224
+ },
1225
+ "model.layers.53.mlp.shared_experts.up_proj": {
1226
+ "bits": 8
1227
+ },
1228
  "model.layers.53.self_attn.kv_a_proj_with_mqa": {
1229
  "bits": 8
1230
  },
 
1240
  "model.layers.53.self_attn.q_b_proj": {
1241
  "bits": 8
1242
  },
1243
+ "model.layers.54.mlp.shared_experts.down_proj": {
1244
+ "bits": 8
1245
+ },
1246
+ "model.layers.54.mlp.shared_experts.gate_proj": {
1247
+ "bits": 8
1248
+ },
1249
+ "model.layers.54.mlp.shared_experts.up_proj": {
1250
+ "bits": 8
1251
+ },
1252
  "model.layers.54.self_attn.kv_a_proj_with_mqa": {
1253
  "bits": 8
1254
  },
 
1264
  "model.layers.54.self_attn.q_b_proj": {
1265
  "bits": 8
1266
  },
1267
+ "model.layers.55.mlp.shared_experts.down_proj": {
1268
+ "bits": 8
1269
+ },
1270
+ "model.layers.55.mlp.shared_experts.gate_proj": {
1271
+ "bits": 8
1272
+ },
1273
+ "model.layers.55.mlp.shared_experts.up_proj": {
1274
+ "bits": 8
1275
+ },
1276
  "model.layers.55.self_attn.kv_a_proj_with_mqa": {
1277
  "bits": 8
1278
  },
 
1288
  "model.layers.55.self_attn.q_b_proj": {
1289
  "bits": 8
1290
  },
1291
+ "model.layers.56.mlp.shared_experts.down_proj": {
1292
+ "bits": 8
1293
+ },
1294
+ "model.layers.56.mlp.shared_experts.gate_proj": {
1295
+ "bits": 8
1296
+ },
1297
+ "model.layers.56.mlp.shared_experts.up_proj": {
1298
+ "bits": 8
1299
+ },
1300
  "model.layers.56.self_attn.kv_a_proj_with_mqa": {
1301
  "bits": 8
1302
  },
 
1312
  "model.layers.56.self_attn.q_b_proj": {
1313
  "bits": 8
1314
  },
1315
+ "model.layers.57.mlp.shared_experts.down_proj": {
1316
+ "bits": 8
1317
+ },
1318
+ "model.layers.57.mlp.shared_experts.gate_proj": {
1319
+ "bits": 8
1320
+ },
1321
+ "model.layers.57.mlp.shared_experts.up_proj": {
1322
+ "bits": 8
1323
+ },
1324
  "model.layers.57.self_attn.kv_a_proj_with_mqa": {
1325
  "bits": 8
1326
  },
 
1336
  "model.layers.57.self_attn.q_b_proj": {
1337
  "bits": 8
1338
  },
1339
+ "model.layers.58.mlp.shared_experts.down_proj": {
1340
+ "bits": 8
1341
+ },
1342
+ "model.layers.58.mlp.shared_experts.gate_proj": {
1343
+ "bits": 8
1344
+ },
1345
+ "model.layers.58.mlp.shared_experts.up_proj": {
1346
+ "bits": 8
1347
+ },
1348
  "model.layers.58.self_attn.kv_a_proj_with_mqa": {
1349
  "bits": 8
1350
  },
 
1360
  "model.layers.58.self_attn.q_b_proj": {
1361
  "bits": 8
1362
  },
1363
+ "model.layers.59.mlp.shared_experts.down_proj": {
1364
+ "bits": 8
1365
+ },
1366
+ "model.layers.59.mlp.shared_experts.gate_proj": {
1367
+ "bits": 8
1368
+ },
1369
+ "model.layers.59.mlp.shared_experts.up_proj": {
1370
+ "bits": 8
1371
+ },
1372
  "model.layers.59.self_attn.kv_a_proj_with_mqa": {
1373
  "bits": 8
1374
  },
 
1384
  "model.layers.59.self_attn.q_b_proj": {
1385
  "bits": 8
1386
  },
1387
+ "model.layers.6.mlp.shared_experts.down_proj": {
1388
+ "bits": 8
1389
+ },
1390
+ "model.layers.6.mlp.shared_experts.gate_proj": {
1391
+ "bits": 8
1392
+ },
1393
+ "model.layers.6.mlp.shared_experts.up_proj": {
1394
+ "bits": 8
1395
+ },
1396
  "model.layers.6.self_attn.kv_a_proj_with_mqa": {
1397
  "bits": 8
1398
  },
 
1408
  "model.layers.6.self_attn.q_b_proj": {
1409
  "bits": 8
1410
  },
1411
+ "model.layers.60.mlp.shared_experts.down_proj": {
1412
+ "bits": 8
1413
+ },
1414
+ "model.layers.60.mlp.shared_experts.gate_proj": {
1415
+ "bits": 8
1416
+ },
1417
+ "model.layers.60.mlp.shared_experts.up_proj": {
1418
+ "bits": 8
1419
+ },
1420
  "model.layers.60.self_attn.kv_a_proj_with_mqa": {
1421
  "bits": 8
1422
  },
 
1432
  "model.layers.60.self_attn.q_b_proj": {
1433
  "bits": 8
1434
  },
1435
+ "model.layers.7.mlp.shared_experts.down_proj": {
1436
+ "bits": 8
1437
+ },
1438
+ "model.layers.7.mlp.shared_experts.gate_proj": {
1439
+ "bits": 8
1440
+ },
1441
+ "model.layers.7.mlp.shared_experts.up_proj": {
1442
+ "bits": 8
1443
+ },
1444
  "model.layers.7.self_attn.kv_a_proj_with_mqa": {
1445
  "bits": 8
1446
  },
 
1456
  "model.layers.7.self_attn.q_b_proj": {
1457
  "bits": 8
1458
  },
1459
+ "model.layers.8.mlp.shared_experts.down_proj": {
1460
+ "bits": 8
1461
+ },
1462
+ "model.layers.8.mlp.shared_experts.gate_proj": {
1463
+ "bits": 8
1464
+ },
1465
+ "model.layers.8.mlp.shared_experts.up_proj": {
1466
+ "bits": 8
1467
+ },
1468
  "model.layers.8.self_attn.kv_a_proj_with_mqa": {
1469
  "bits": 8
1470
  },
 
1480
  "model.layers.8.self_attn.q_b_proj": {
1481
  "bits": 8
1482
  },
1483
+ "model.layers.9.mlp.shared_experts.down_proj": {
1484
+ "bits": 8
1485
+ },
1486
+ "model.layers.9.mlp.shared_experts.gate_proj": {
1487
+ "bits": 8
1488
+ },
1489
+ "model.layers.9.mlp.shared_experts.up_proj": {
1490
+ "bits": 8
1491
+ },
1492
  "model.layers.9.self_attn.kv_a_proj_with_mqa": {
1493
  "bits": 8
1494
  },
model-00002-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5a5ba61ae9bf2b3907e48105177902846e063f40a76301c19907c68f47c95b1
3
- size 5000051392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0462b3a214fa5ea5bad4b053f2a70b6d65919d962a0e1ac9f894161efbc496a7
3
+ size 4999591680
model-00003-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b08a3f557e84a92e41f1e92f99f550cc0d8a8442f21b95eb3fa480703eecc1f8
3
- size 4995991608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88b741b175aa1149079d35cc5bfb86bc16aa1e0e88f1e2ebf8831fafd427c63a
3
+ size 4995762096
model-00004-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7020b0c09d1aca0b1a508d4d3039b3904dbd74953334cd8a2769ec3608b9108b
3
- size 5000050912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f7d2a3285f7458d3e92649afdac0b9f973ab5a08167511e6f65b6e42253d2de
3
+ size 4999591192
model-00005-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72f624ca05d3b54dfc34267ee659ba12d15d5b372db2feec13fe2b496a2d8211
3
- size 5000222944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd8e279298f6750bf7a9a6a9011e78a2d2558cc3478fe9aa0468ed5579dd1738
3
+ size 4999591176
model-00006-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3a0f1fb2164bab6a13fa7a54501d09996c34d7e41a6cb2d39d9d621e0410ad4
3
- size 4992940416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:525493633455c0f46e00b6bb06417ecc803a746b8848132d9c2a30f88b6f1087
3
+ size 4999591384
model-00007-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a4be835f87031118ce99f89b55553273a1956f44fb987f632e5bdc7374688407
3
- size 5000051336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3aae5be40df7755d18385d1292b3ab4ca53f2897238760bdeca657d75181a48f
3
+ size 4999591608
model-00008-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f68acaf80fe8ed102b0a1f946c9fc49548de37d408c89460d8fd3a49d98f49dd
3
- size 4995991936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfa4a06b23c694d9afe0b166d7bb9a2baf60eb2ff20eb34177e935e9a24411c3
3
+ size 4995762400
model-00009-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:347da340973bc3c9e78814c9b0b37007d2342092961b2260ecb8360db190d912
3
- size 5000050912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c838136cccadf85d169666aca32792bd1cb29c2dc973c81946887c41e4d87fb
3
+ size 4999591192
model-00010-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f808ce67c8deee35f928cd4f507d0b7d91b9782752adce381940569cb390982f
3
- size 5000224376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:370201acc095b9bed2173c26048e6e003c19bc62ed36958c35bf33687c3687b1
3
+ size 4999592560
model-00011-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9c32b02f7fc2fbaa588c4a1a561c3c845b5cdcc97e4f9d4d990558d3db30414
3
- size 4992942152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a40bc6754a40b1b52f337f904f10b0b7466e89e7d44a9a5267781b2ed6ff36e7
3
+ size 4999593080
model-00012-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9112c487f3ecf185605d4cf4dac92138519dbe5ace64f3a9be1c9a04630eea4b
3
- size 5000053152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d06ca481c6a2acac2274992d5b0969d64b2e13de8bf2ae5455dfaea0fa0ee548
3
+ size 4999593480
model-00013-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c184fc0ee3bf5e33047f40f4c2855844c909040a4b95b2cff34bc99b018ff625
3
- size 5000225280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b486cbea28c5e55dd0c9c17257eac85a5f0f7744440e0acc6fe646e764aad94
3
+ size 4999593536
model-00014-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9909ab2a8a7ce7bc5a7c5b19feeb2d9f3e060c4d0ca8179c58db9a738be83436
3
- size 4996051216
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce5a242eb6094293854ffa84c07120c31a1be234b9778d6c3d3c8328fa35c106
3
+ size 4995764144
model-00015-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da31552606f00a3a21049637416f845c0f55cedc40e060f120114447503cf0d3
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af629bee2dc2c2b286e5249a4df175ded456fb124f4cd5ba956c24cc409b4644
3
+ size 4999593096
model-00016-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0bc3bad1cac7808d24cd57b4cb3409052a9b06de766d9413fea2f8c868b6600
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:204e129139fd8ea6951f2d6f171d7380b6a0edf4dddbb0a6bbd5a9f5567e01c7
3
+ size 4999593080
model-00017-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17b5c9c247567c7624beea886458abeae76c9958aacf888925f01879dd5ac209
3
- size 4992942208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a05e3ae2f3998deb65c2a457e23ce91dabe6d91da413b29990dd44c02673804
3
+ size 4999593248
model-00018-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c04f2c0095391732d42259480fda5d2219e457d01e0179e36e450ca078f4db3
3
- size 5000053248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09edbe9a68debfa5c46ab73d29e69e98de50676e96c8be8f3f004e1abfc4e142
3
+ size 4999593512
model-00019-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7aba70afcdf3ce089821a36d63c5dbfec323c4bac17c8354057770ecb9a11e3
3
- size 4887946080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:baa360db2d5af6ceca018b5c9dbf11341f4671f2eb6469d450cdbd95e69a2017
3
+ size 4995764408
model-00020-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:10a936066825e6e83310f5beda543af76caa7454c228261ccb6ebcecb388b5d9
3
- size 4993923416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9426b4e92393fde4ae09268bd50dacaf35c2712534250a9f957198d4ff6ccaea
3
+ size 4999593096
model-00021-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbf954fdfee1482d82ae4b7dd3b81f3c96e94503c3cd0b879cd84d8601f39e6b
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92a858c6b64bb5062e27e82fb71e29bc56acf0be7bc2ec690941de70f31ca4c8
3
+ size 4999593080
model-00022-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1e391cdccc84a5b3ec1723c5f871e2070a0d521ff77c356bf36ae76f6878b03
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8964b092f1345171f6c942c2dc3c4643e80e26120160b562b22a62f907b62174
3
+ size 4999593080
model-00023-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:702015a7919d71819afb95f0c9e3cf74216fcf4596f9062189e73fcf569f9d39
3
- size 4992942408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06df694bb89764e1845a26c86b71b9102a95c8deee7852cafcf605d2f9a9a64d
3
+ size 4999593440
model-00024-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:422dd7e4209744716eeaa76779f30457b7d1b2e6bf105d4947e671ce6921e523
3
- size 5000053248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94ecaee51f1a488ab3626a2f9119b563cfeaa6433f89eb59cc416eadec984022
3
+ size 4999593512
model-00025-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:476c13bc42d1fffcabcf1f83229aa435201c5a2f92e008305e3278d674fa792e
3
- size 4995993824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14c7cb322477eb98733db90c8b08b3e753e25fb61026c00684cf6da1804607ef
3
+ size 4995764216
model-00026-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:441cf0656c2358e3e15a326b8f280de0ae09a4b7481e0752d85571a6e7ad76c1
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ee4323ce69e1141320852699c39ccb8a7c29e9541af4a8d5ba8a384d7450430
3
+ size 4999593096
model-00027-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c0eaca5ef29ed2e0b985af88687d13699f075332e45f006c66530e0c740985c
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3586365b51e3ef370eb3d62a637013040b5fc558dd94ae9563698d9ea04d45bd
3
+ size 4999593080
model-00028-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4db427a1585b57fdd8fc3b84138c00548c9b310e0d5fb2fe3f69944eae6dc5d0
3
- size 4992942152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eea128b89640d9864d22ff5543adecc4200dba21cf272c3c2ade8341541a837
3
+ size 4999593208
model-00029-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e276ebe719c214518a53b38396ad526159d32071d8df3fbd99316e9add87be57
3
- size 5000053232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61efba4225c03154ffd9ac51d9c05d195058986934fbdb175bfca9e9cf89df93
3
+ size 4999593512
model-00030-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9ae27cc538f8ab8da4ce267c76d7b0f4dc33bccaf5067a0b12ba5cf143a2e31
3
- size 5000225304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c0d0d0c31bee444b012680d07a2a3432474cfda0a71574dfcbceb3cfbd2f8fd
3
+ size 4999205472
model-00031-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:983b30e8172ebda2b844d5287959e8c50d731fa3371a207257d680aa5b9121cb
3
- size 4996051112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dc916619de62cd657edc348e89feae03adfefea7efb8feef09ec2927fd37bc3
3
+ size 4996152080
model-00032-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ee2587fcdbcb7966e2e46e557b9a86df222f86fc9ff3c78a37b08b45c1134b5
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0f92fe783a8acc864054da76d0ca2d75d30c951b0bbffae1bac2340ceffd120
3
+ size 4999593080
model-00033-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3d88968cec7322b8b0559f518f0d2b3fab1adbbe1940a3e0078fdb810f4a040
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d398780f22d6402471e7220b0e29074f8d89675698197ac5d9ee441c06dcbac3
3
+ size 4999593080
model-00034-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0f2715617fe58702606a3891d3eff57ce5f6b402cbeeefacea858a64b3f0b01
3
- size 4992942288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7d991a04fd2d0341fcc2a25b8a177f945e2520a68622199f27a5c9dfafbbf51
3
+ size 4999593400
model-00035-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa403dc7a766d421d56db4fab0e8144c33307cbface657e58ce08dd0723a21d7
3
- size 5000053248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba2bbff3643ddb289667d98511aa0df31893007fda01d413cea8bfad500ae95c
3
+ size 4999593512
model-00036-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3af20d5f58e03c46ebcb02a9127eb3a80cd96257e778c7000b2a28a609c1864f
3
- size 4999664696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c076d23a0c6bee26147aa7d18e4c54fe39a1e0bf15f389351a0ead3ccb13095c
3
+ size 4995764256
model-00037-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74459e2953f3b36afa8426eab858336930b597ac3c652bb16e30ba4fb2f301db
3
- size 4996611560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf065e01d8d893952a188ffe99a3a9b6303bebe80a310632ec5ad8f1a84e59f2
3
+ size 4999593096
model-00038-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3d06df18ed937935a1fbafb1d212ac8c0c5225f2a32285f8f364092586c45f5
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faebe749b3e0a34c2dd74c18c6194cf45c8c306310c8388b3e3f7a2cffaf4285
3
+ size 4999593080
model-00039-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8de186711e1ef31eea8eb178c4cb45f39130d77623617982726c2f73ec384667
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2e7e686f371d25e83370114112e11a0b48bbf7d578741bbef4ed240dd28a4ee
3
+ size 4999593160
model-00040-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5ab209db638b7e349a3f8266abed57b7bcb84320b4e33aff59e8e2aff184f5b
3
- size 4992942440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea1f8e8784ad460c66fb17164e34674b002e4bbb161f1c616a0897007183f4b2
3
+ size 4999593512
model-00041-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b52b41a366e94ca5c038940517cd4f594850c5bcc2cc65f27686e267e99b7675
3
- size 5000053248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88c5b246725c00848b61517495f2006aa421c0dea9191ae62f65b8d92c64cc40
3
+ size 4993807032
model-00042-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4f148561dd6769a5ff5c8242aaea132cbe7c13031dc6af0f5b7a72bb5800d2a
3
- size 4995993792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5894290536addb4b108a7bd5597e32d036ccb0eee11834fd98ee188983724852
3
+ size 4993923416
model-00043-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b06564b3422d93e8d8a171ab270de1baff4e4aca1d997b06a496079b1078f78c
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4659c84b744787225d253bc2774d9f2ef0af1117d379b7f726a67be479c825d8
3
+ size 4999593096
model-00044-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff9c0c8c6a4aef00b34a1b2766ae347ab853629c7da6f9c48e942c65a55c381c
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3154593a83985f8cc0bbc80e79fa387ca5a077af6690ac9cc5201ae471399a6
3
+ size 4999593080
model-00045-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c4b7abd0c88a285639dcefe4ed905d0d0b31958c94ea3353df623737a0e3f27
3
- size 4992942176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdd23d19b26a6c8d381b3acc906c190f6ddb27724803a81d2ffd42cb9f785aee
3
+ size 4999593360
model-00046-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2139f7b5a065a037e76b1d3a88e60f86157040fc93608e629c7fbc3d2390c289
3
- size 5000053248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bea09251968dc9ca17adc10092e10161e79e9858f4e7509d94cbb12486da76d
3
+ size 4999593512
model-00047-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:07bef778b2417fa450ee481f892353398c18abc875b238478112bc3677691e89
3
- size 5000225336
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:207264998aad27b1f372140968aa40d061b6128f6db5d4cf57086a4240f99bf6
3
+ size 4995764296
model-00048-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c50b59bc6c2522e7bd5fc0ff827fd239a6cc8cbc3b0f6f9ee608a5eba8f294fe
3
- size 4996051040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:799f9792c3a5025b0849b50976fa3250302b9e910504230317ffaa2576ae86c6
3
+ size 4999593096
model-00049-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8834e9c291ed773cbae76ff5f7b29df7641639d6dd3abcc03f0c9de82126a510
3
- size 5000052816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbe6bbd60af7680e0642aeba1bd9b1041ec9171cde7613530f9f03d5c83144d1
3
+ size 4999593080
model-00050-of-00072.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51501512b569be6b1756375d1c980e6ac1884e479ef303345619cd78a7960105
3
- size 5000224856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73cbdc09aefed8d32ddbe0f214e07f7c891c5a15c906c57d7253f32c54cec09a
3
+ size 4999593120