woodchen7 commited on
Commit
67580f8
·
verified ·
1 Parent(s): 07ee42b

Upload angelslim_config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. angelslim_config.json +76 -0
angelslim_config.json ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_config": {
3
+ "name": "Qwen",
4
+ "model_path": "Base Model Path",
5
+ "trust_remote_code": true,
6
+ "torch_dtype": "auto",
7
+ "device_map": "auto",
8
+ "low_cpu_mem_usage": true,
9
+ "use_cache": false,
10
+ "cache_dir": null
11
+ },
12
+ "compression_config": {
13
+ "name": [
14
+ "PTQ"
15
+ ],
16
+ "quantization": {
17
+ "name": "nvfp4",
18
+ "bits": 4,
19
+ "quant_method": {
20
+ "weight": "per-block",
21
+ "activation": "per-block",
22
+ "group_size": 16
23
+ },
24
+ "quant_helpers": [],
25
+ "smooth_alpha": 0.5,
26
+ "low_memory": false,
27
+ "modules_to_quantize": [],
28
+ "zero_point": true,
29
+ "mse_range": false,
30
+ "ignore_layers": [
31
+ "lm_head",
32
+ "model.embed_tokens",
33
+ "lm_head"
34
+ ],
35
+ "quant_analyse": false,
36
+ "quant_vit": false
37
+ },
38
+ "cache": null
39
+ },
40
+ "dataset_config": {
41
+ "name": "TextDataset",
42
+ "data_path": "/cfs_cloud_code/wdarrenhu/open_slim/openslim_v5/OpenSlim/dataset/hunyuan/7b_dense_qwen3_235b_a22b_output_256.jsonl",
43
+ "max_seq_length": 4096,
44
+ "num_samples": 256,
45
+ "batch_size": 1,
46
+ "shuffle": false,
47
+ "inference_settings": null
48
+ },
49
+ "global_config": {
50
+ "save_path": "Save Model Path",
51
+ "max_seq_length": 4096,
52
+ "hidden_size": 5120,
53
+ "model_arch_type": "qwen3",
54
+ "deploy_backend": "vllm"
55
+ },
56
+ "infer_config": null,
57
+ "debug_info": {
58
+ "python": "3.10.12 (main, Aug 29 2024, 16:22:46) [GCC 9.3.0]",
59
+ "angelslim": {
60
+ "name": "angelslim",
61
+ "version": "0.0.0.dev0",
62
+ "source": "pip"
63
+ },
64
+ "torch": {
65
+ "name": "torch",
66
+ "version": "2.3.1",
67
+ "source": "pip"
68
+ },
69
+ "transformers": {
70
+ "name": "transformers",
71
+ "version": "4.56.1",
72
+ "source": "pip"
73
+ },
74
+ "torch_cuda_version": "12.3"
75
+ }
76
+ }