cpatonn commited on
Commit
34a2b1b
·
verified ·
1 Parent(s): 8f04ff4

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -106,7 +106,7 @@
106
  "quantization_status": "compressed",
107
  "sparsity_config": {},
108
  "transform_config": {},
109
- "version": "0.11.1.a20250912"
110
  },
111
  "rms_norm_eps": 1e-06,
112
  "rope_scaling": null,
@@ -114,8 +114,8 @@
114
  "router_aux_loss_coef": 0.001,
115
  "sliding_window": null,
116
  "tie_word_embeddings": false,
117
- "transformers_version": "4.57.0.dev0",
118
  "use_cache": true,
119
  "use_sliding_window": false,
120
  "vocab_size": 151936
121
- }
 
106
  "quantization_status": "compressed",
107
  "sparsity_config": {},
108
  "transform_config": {},
109
+ "version": "0.12.3.a20251114"
110
  },
111
  "rms_norm_eps": 1e-06,
112
  "rope_scaling": null,
 
114
  "router_aux_loss_coef": 0.001,
115
  "sliding_window": null,
116
  "tie_word_embeddings": false,
117
+ "transformers_version": "4.57.1",
118
  "use_cache": true,
119
  "use_sliding_window": false,
120
  "vocab_size": 151936
121
+ }
generation_config.json CHANGED
@@ -9,5 +9,5 @@
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
- "transformers_version": "4.57.0.dev0"
13
  }
 
9
  "temperature": 0.6,
10
  "top_k": 20,
11
  "top_p": 0.95,
12
+ "transformers_version": "4.57.1"
13
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:078d12c1ff7aaf6ff9a2943d06e8907de460581376ecdca38e9ca4013201c4a4
3
  size 5001707008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae59a04c559fd19fffb4ea029b584f4c2121fe82dfe50908ef9a5109ba1674ac
3
  size 5001707008
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:545186f5ff36c0aa2ebc4e7b8154058d6ed3b92ec2796f9c11367e75ea357239
3
  size 5001283696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82134082d19a2399f9e518777bd9e6d8ae5ef3bbb73a2ff4e3d4727049af7e33
3
  size 5001283696
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:250b38969c285a6213621aef4e4cdab99a85aacda3ebaa0a18d20d788e121386
3
  size 5001283912
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d377ef3c4f84df9711ccb5c64e2d59521839d68404544d5779340c7f2b0818d
3
  size 5001283912
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f157386d73df8142fafc87e23a76961d397edd37e7fa8884e588b931f8ca62c
3
  size 3090232736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39bb63e3056ca762f9f2dbeeab33f53c74f0b8f85c584b5a2bd9ae7ac9471c54
3
  size 3090232736
recipe.yaml CHANGED
@@ -1,5 +1,5 @@
1
- quant_stage:
2
- quant_modifiers:
3
  AWQModifier:
4
  config_groups:
5
  group_0:
@@ -13,6 +13,8 @@ quant_stage:
13
  block_structure: null
14
  dynamic: false
15
  actorder: null
 
 
16
  observer: mse
17
  observer_kwargs: {}
18
  input_activations: null
@@ -30,4 +32,6 @@ quant_stage:
30
  balance_layers: ['re:.*mlp.experts.*.gate_proj$', 're:.*mlp.experts.*.up_proj$']
31
  - smooth_layer: re:.*up_proj$
32
  balance_layers: ['re:.*down_proj$']
 
33
  duo_scaling: true
 
 
1
+ default_stage:
2
+ default_modifiers:
3
  AWQModifier:
4
  config_groups:
5
  group_0:
 
13
  block_structure: null
14
  dynamic: false
15
  actorder: null
16
+ scale_dtype: null
17
+ zp_dtype: null
18
  observer: mse
19
  observer_kwargs: {}
20
  input_activations: null
 
32
  balance_layers: ['re:.*mlp.experts.*.gate_proj$', 're:.*mlp.experts.*.up_proj$']
33
  - smooth_layer: re:.*up_proj$
34
  balance_layers: ['re:.*down_proj$']
35
+ offload_device: !!python/object/apply:torch.device [cpu]
36
  duo_scaling: true
37
+ n_grid: 20
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ae1a036a9837df9caeebb840d09d80e8feef0f6d2bae982970d1ad34f5946aff
3
- size 11422753
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654