nabi-chan commited on
Commit
545c023
·
verified ·
1 Parent(s): ff3b8cc

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
LICENSE ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ Apache License
3
+ Version 2.0, January 2004
4
+ http://www.apache.org/licenses/
5
+
6
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
7
+
8
+ 1. Definitions.
9
+
10
+ "License" shall mean the terms and conditions for use, reproduction,
11
+ and distribution as defined by Sections 1 through 9 of this document.
12
+
13
+ "Licensor" shall mean the copyright owner or entity authorized by
14
+ the copyright owner that is granting the License.
15
+
16
+ "Legal Entity" shall mean the union of the acting entity and all
17
+ other entities that control, are controlled by, or are under common
18
+ control with that entity. For the purposes of this definition,
19
+ "control" means (i) the power, direct or indirect, to cause the
20
+ direction or management of such entity, whether by contract or
21
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
22
+ outstanding shares, or (iii) beneficial ownership of such entity.
23
+
24
+ "You" (or "Your") shall mean an individual or Legal Entity
25
+ exercising permissions granted by this License.
26
+
27
+ "Source" form shall mean the preferred form for making modifications,
28
+ including but not limited to software source code, documentation
29
+ source, and configuration files.
30
+
31
+ "Object" form shall mean any form resulting from mechanical
32
+ transformation or translation of a Source form, including but
33
+ not limited to compiled object code, generated documentation,
34
+ and conversions to other media types.
35
+
36
+ "Work" shall mean the work of authorship, whether in Source or
37
+ Object form, made available under the License, as indicated by a
38
+ copyright notice that is included in or attached to the work
39
+ (an example is provided in the Appendix below).
40
+
41
+ "Derivative Works" shall mean any work, whether in Source or Object
42
+ form, that is based on (or derived from) the Work and for which the
43
+ editorial revisions, annotations, elaborations, or other modifications
44
+ represent, as a whole, an original work of authorship. For the purposes
45
+ of this License, Derivative Works shall not include works that remain
46
+ separable from, or merely link (or bind by name) to the interfaces of,
47
+ the Work and Derivative Works thereof.
48
+
49
+ "Contribution" shall mean any work of authorship, including
50
+ the original version of the Work and any modifications or additions
51
+ to that Work or Derivative Works thereof, that is intentionally
52
+ submitted to Licensor for inclusion in the Work by the copyright owner
53
+ or by an individual or Legal Entity authorized to submit on behalf of
54
+ the copyright owner. For the purposes of this definition, "submitted"
55
+ means any form of electronic, verbal, or written communication sent
56
+ to the Licensor or its representatives, including but not limited to
57
+ communication on electronic mailing lists, source code control systems,
58
+ and issue tracking systems that are managed by, or on behalf of, the
59
+ Licensor for the purpose of discussing and improving the Work, but
60
+ excluding communication that is conspicuously marked or otherwise
61
+ designated in writing by the copyright owner as "Not a Contribution."
62
+
63
+ "Contributor" shall mean Licensor and any individual or Legal Entity
64
+ on behalf of whom a Contribution has been received by Licensor and
65
+ subsequently incorporated within the Work.
66
+
67
+ 2. Grant of Copyright License. Subject to the terms and conditions of
68
+ this License, each Contributor hereby grants to You a perpetual,
69
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
70
+ copyright license to reproduce, prepare Derivative Works of,
71
+ publicly display, publicly perform, sublicense, and distribute the
72
+ Work and such Derivative Works in Source or Object form.
73
+
74
+ 3. Grant of Patent License. Subject to the terms and conditions of
75
+ this License, each Contributor hereby grants to You a perpetual,
76
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
77
+ (except as stated in this section) patent license to make, have made,
78
+ use, offer to sell, sell, import, and otherwise transfer the Work,
79
+ where such license applies only to those patent claims licensable
80
+ by such Contributor that are necessarily infringed by their
81
+ Contribution(s) alone or by combination of their Contribution(s)
82
+ with the Work to which such Contribution(s) was submitted. If You
83
+ institute patent litigation against any entity (including a
84
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
85
+ or a Contribution incorporated within the Work constitutes direct
86
+ or contributory patent infringement, then any patent licenses
87
+ granted to You under this License for that Work shall terminate
88
+ as of the date such litigation is filed.
89
+
90
+ 4. Redistribution. You may reproduce and distribute copies of the
91
+ Work or Derivative Works thereof in any medium, with or without
92
+ modifications, and in Source or Object form, provided that You
93
+ meet the following conditions:
94
+
95
+ (a) You must give any other recipients of the Work or
96
+ Derivative Works a copy of this License; and
97
+
98
+ (b) You must cause any modified files to carry prominent notices
99
+ stating that You changed the files; and
100
+
101
+ (c) You must retain, in the Source form of any Derivative Works
102
+ that You distribute, all copyright, patent, trademark, and
103
+ attribution notices from the Source form of the Work,
104
+ excluding those notices that do not pertain to any part of
105
+ the Derivative Works; and
106
+
107
+ (d) If the Work includes a "NOTICE" text file as part of its
108
+ distribution, then any Derivative Works that You distribute must
109
+ include a readable copy of the attribution notices contained
110
+ within such NOTICE file, excluding those notices that do not
111
+ pertain to any part of the Derivative Works, in at least one
112
+ of the following places: within a NOTICE text file distributed
113
+ as part of the Derivative Works; within the Source form or
114
+ documentation, if provided along with the Derivative Works; or,
115
+ within a display generated by the Derivative Works, if and
116
+ wherever such third-party notices normally appear. The contents
117
+ of the NOTICE file are for informational purposes only and
118
+ do not modify the License. You may add Your own attribution
119
+ notices within Derivative Works that You distribute, alongside
120
+ or as an addendum to the NOTICE text from the Work, provided
121
+ that such additional attribution notices cannot be construed
122
+ as modifying the License.
123
+
124
+ You may add Your own copyright statement to Your modifications and
125
+ may provide additional or different license terms and conditions
126
+ for use, reproduction, or distribution of Your modifications, or
127
+ for any such Derivative Works as a whole, provided Your use,
128
+ reproduction, and distribution of the Work otherwise complies with
129
+ the conditions stated in this License.
130
+
131
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
132
+ any Contribution intentionally submitted for inclusion in the Work
133
+ by You to the Licensor shall be under the terms and conditions of
134
+ this License, without any additional terms or conditions.
135
+ Notwithstanding the above, nothing herein shall supersede or modify
136
+ the terms of any separate license agreement you may have executed
137
+ with Licensor regarding such Contributions.
138
+
139
+ 6. Trademarks. This License does not grant permission to use the trade
140
+ names, trademarks, service marks, or product names of the Licensor,
141
+ except as required for reasonable and customary use in describing the
142
+ origin of the Work and reproducing the content of the NOTICE file.
143
+
144
+ 7. Disclaimer of Warranty. Unless required by applicable law or
145
+ agreed to in writing, Licensor provides the Work (and each
146
+ Contributor provides its Contributions) on an "AS IS" BASIS,
147
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
148
+ implied, including, without limitation, any warranties or conditions
149
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
150
+ PARTICULAR PURPOSE. You are solely responsible for determining the
151
+ appropriateness of using or redistributing the Work and assume any
152
+ risks associated with Your exercise of permissions under this License.
153
+
154
+ 8. Limitation of Liability. In no event and under no legal theory,
155
+ whether in tort (including negligence), contract, or otherwise,
156
+ unless required by applicable law (such as deliberate and grossly
157
+ negligent acts) or agreed to in writing, shall any Contributor be
158
+ liable to You for damages, including any direct, indirect, special,
159
+ incidental, or consequential damages of any character arising as a
160
+ result of this License or out of the use or inability to use the
161
+ Work (including but not limited to damages for loss of goodwill,
162
+ work stoppage, computer failure or malfunction, or any and all
163
+ other commercial damages or losses), even if such Contributor
164
+ has been advised of the possibility of such damages.
165
+
166
+ 9. Accepting Warranty or Additional Liability. While redistributing
167
+ the Work or Derivative Works thereof, You may choose to offer,
168
+ and charge a fee for, acceptance of support, warranty, indemnity,
169
+ or other liability obligations and/or rights consistent with this
170
+ License. However, in accepting such obligations, You may act only
171
+ on Your own behalf and on Your sole responsibility, not on behalf
172
+ of any other Contributor, and only if You agree to indemnify,
173
+ defend, and hold each Contributor harmless for any liability
174
+ incurred by, or claims asserted against, such Contributor by reason
175
+ of your accepting any such warranty or additional liability.
176
+
177
+ END OF TERMS AND CONDITIONS
178
+
179
+ APPENDIX: How to apply the Apache License to your work.
180
+
181
+ To apply the Apache License to your work, attach the following
182
+ boilerplate notice, with the fields enclosed by brackets "[]"
183
+ replaced with your own identifying information. (Don't include
184
+ the brackets!) The text should be enclosed in the appropriate
185
+ comment syntax for the file format. We also recommend that a
186
+ file or class name and description of purpose be included on the
187
+ same "printed page" as the copyright notice for easier
188
+ identification within third-party archives.
189
+
190
+ Copyright 2026 Alibaba Cloud
191
+
192
+ Licensed under the Apache License, Version 2.0 (the "License");
193
+ you may not use this file except in compliance with the License.
194
+ You may obtain a copy of the License at
195
+
196
+ http://www.apache.org/licenses/LICENSE-2.0
197
+
198
+ Unless required by applicable law or agreed to in writing, software
199
+ distributed under the License is distributed on an "AS IS" BASIS,
200
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
201
+ See the License for the specific language governing permissions and
202
+ limitations under the License.
README.md ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: mlx
3
+ license: "apache-2.0"
4
+ license_link: https://huggingface.co/Qwen/Qwen3.6-27B/blob/main/LICENSE
5
+ pipeline_tag: text-generation
6
+ language:
7
+ - en
8
+ - ko
9
+ - zh
10
+ - ja
11
+ tags:
12
+ - mlx
13
+ - "mlx-5bit"
14
+ - quantized
15
+ - safetensors
16
+ - apple-silicon
17
+ - Qwen
18
+ - Qwen3.6
19
+ - Qwen3_5
20
+ - abliterated
21
+ - uncensored
22
+ base_model:
23
+ - huihui-ai/Huihui-Qwen3.6-27B-abliterated
24
+ ---
25
+
26
+ # 🌌 `huihui-ai/Huihui-Qwen3.6-27B-abliterated` converted to MLX 5-bit
27
+
28
+ ## About This Quantization
29
+
30
+ **Apple Sllicon / MLX 5-bit**
31
+
32
+ - **Source Model (BF16)** : [huihui-ai/Huihui-Qwen3.6-27B-abliterated](https://huggingface.co/huihui-ai/Huihui-Qwen3.6-27B-abliterated)
33
+ - **Quantized By:** [@nabi-chan](https://huggingface.co/nabi-chan)
34
+
35
+ ### Quickstart
36
+
37
+ #### Install
38
+
39
+ ```bash
40
+ pip install -U "mlx-lm>=0.31.2"
41
+ ```
42
+
43
+ #### Python
44
+
45
+ ```python
46
+ from mlx_lm import load, generate
47
+
48
+ model, tokenizer = load("nabi-chan/.Huihui-Qwen3.6-27B-abliterated-MLX-5bit.tmp-1478")
49
+ print(generate(model, tokenizer, prompt="Explain quantum entanglement simply.", max_tokens=128))
50
+ ```
51
+
52
+ #### CLI
53
+
54
+ ```bash
55
+ python3 -m mlx_lm generate \
56
+ --model nabi-chan/.Huihui-Qwen3.6-27B-abliterated-MLX-5bit.tmp-1478 \
57
+ --prompt "Write a haiku about Apple Silicon." \
58
+ --max-tokens 128
59
+ ```
60
+
61
+ ### Quantization Details
62
+
63
+ | Property | Value |
64
+ | --------------------- | ---------------------------------------------------------------------------------------------- |
65
+ | **Method** | MLX affine quantization |
66
+ | **Bits / weight** | 5 |
67
+ | **Group size** | 64 |
68
+ | **Non-quant dtype** | bfloat16 |
69
+ | **Quantizer version** | `mlx` : 0.31.2 / `mlx-lm` : 0.31.3 / `mlx-vlm`: 0.4.4 |
70
+
71
+ > [!WARNING]
72
+ > Protected tensors keep their original dtype. In VLM models, vision tensors and some guarded layers may remain unquantized.
73
+
74
+ ---
75
+
76
+ Everything below is huihui-ai's original model card, preserved verbatim.
77
+
78
+ ---
79
+
80
+ # huihui-ai/Huihui-Qwen3.6-27B-abliterated
81
+
82
+
83
+ This is an uncensored version of [Qwen/Qwen3.6-27B](https://huggingface.co/Qwen/Qwen3.6-27B) created with abliteration (see [remove-refusals-with-transformers](https://github.com/Sumandora/remove-refusals-with-transformers) to know more about it).
84
+ This is a crude, proof-of-concept implementation to remove refusals from an LLM model without using TransformerLens.
85
+
86
+
87
+ ## ollama
88
+
89
+ Please use the latest version of [ollama](https://github.com/ollama/ollama/releases/tag)
90
+
91
+ You can use [huihui_ai/qwen3.6-abliterated:27b](https://ollama.com/huihui_ai/qwen3.6-abliterated:27b) directly,
92
+ ```
93
+ ollama run huihui_ai/qwen3.6-abliterated:27b
94
+ ```
95
+
96
+ ### Usage Warnings
97
+
98
+
99
+ - **Risk of Sensitive or Controversial Outputs**: This model’s safety filtering has been significantly reduced, potentially generating sensitive, controversial, or inappropriate content. Users should exercise caution and rigorously review generated outputs.
100
+
101
+ - **Not Suitable for All Audiences**: Due to limited content filtering, the model’s outputs may be inappropriate for public settings, underage users, or applications requiring high security.
102
+
103
+ - **Legal and Ethical Responsibilities**: Users must ensure their usage complies with local laws and ethical standards. Generated content may carry legal or ethical risks, and users are solely responsible for any consequences.
104
+
105
+ - **Research and Experimental Use**: It is recommended to use this model for research, testing, or controlled environments, avoiding direct use in production or public-facing commercial applications.
106
+
107
+ - **Monitoring and Review Recommendations**: Users are strongly advised to monitor model outputs in real-time and conduct manual reviews when necessary to prevent the dissemination of inappropriate content.
108
+
109
+ - **No Default Safety Guarantees**: Unlike standard models, this model has not undergone rigorous safety optimization. huihui.ai bears no responsibility for any consequences arising from its use.
110
+
111
+
112
+ ### Donation
113
+ ##### Your donation helps us continue our further development and improvement, a cup of coffee can do it.
114
+ - bitcoin:
115
+ ```
116
+ bc1qqnkhuchxw0zqjh2ku3lu4hq45hc6gy84uk70ge
117
+ ```
118
+ - Support our work on [Ko-fi](https://ko-fi.com/huihuiai)!
chat_template.jinja ADDED
@@ -0,0 +1,154 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- set image_count = namespace(value=0) %}
2
+ {%- set video_count = namespace(value=0) %}
3
+ {%- macro render_content(content, do_vision_count, is_system_content=false) %}
4
+ {%- if content is string %}
5
+ {{- content }}
6
+ {%- elif content is iterable and content is not mapping %}
7
+ {%- for item in content %}
8
+ {%- if 'image' in item or 'image_url' in item or item.type == 'image' %}
9
+ {%- if is_system_content %}
10
+ {{- raise_exception('System message cannot contain images.') }}
11
+ {%- endif %}
12
+ {%- if do_vision_count %}
13
+ {%- set image_count.value = image_count.value + 1 %}
14
+ {%- endif %}
15
+ {%- if add_vision_id %}
16
+ {{- 'Picture ' ~ image_count.value ~ ': ' }}
17
+ {%- endif %}
18
+ {{- '<|vision_start|><|image_pad|><|vision_end|>' }}
19
+ {%- elif 'video' in item or item.type == 'video' %}
20
+ {%- if is_system_content %}
21
+ {{- raise_exception('System message cannot contain videos.') }}
22
+ {%- endif %}
23
+ {%- if do_vision_count %}
24
+ {%- set video_count.value = video_count.value + 1 %}
25
+ {%- endif %}
26
+ {%- if add_vision_id %}
27
+ {{- 'Video ' ~ video_count.value ~ ': ' }}
28
+ {%- endif %}
29
+ {{- '<|vision_start|><|video_pad|><|vision_end|>' }}
30
+ {%- elif 'text' in item %}
31
+ {{- item.text }}
32
+ {%- else %}
33
+ {{- raise_exception('Unexpected item type in content.') }}
34
+ {%- endif %}
35
+ {%- endfor %}
36
+ {%- elif content is none or content is undefined %}
37
+ {{- '' }}
38
+ {%- else %}
39
+ {{- raise_exception('Unexpected content type.') }}
40
+ {%- endif %}
41
+ {%- endmacro %}
42
+ {%- if not messages %}
43
+ {{- raise_exception('No messages provided.') }}
44
+ {%- endif %}
45
+ {%- if tools and tools is iterable and tools is not mapping %}
46
+ {{- '<|im_start|>system\n' }}
47
+ {{- "# Tools\n\nYou have access to the following functions:\n\n<tools>" }}
48
+ {%- for tool in tools %}
49
+ {{- "\n" }}
50
+ {{- tool | tojson }}
51
+ {%- endfor %}
52
+ {{- "\n</tools>" }}
53
+ {{- '\n\nIf you choose to call a function ONLY reply in the following format with NO suffix:\n\n<tool_call>\n<function=example_function_name>\n<parameter=example_parameter_1>\nvalue_1\n</parameter>\n<parameter=example_parameter_2>\nThis is the value for the second parameter\nthat can span\nmultiple lines\n</parameter>\n</function>\n</tool_call>\n\n<IMPORTANT>\nReminder:\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\n- Required parameters MUST be specified\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\n</IMPORTANT>' }}
54
+ {%- if messages[0].role == 'system' %}
55
+ {%- set content = render_content(messages[0].content, false, true)|trim %}
56
+ {%- if content %}
57
+ {{- '\n\n' + content }}
58
+ {%- endif %}
59
+ {%- endif %}
60
+ {{- '<|im_end|>\n' }}
61
+ {%- else %}
62
+ {%- if messages[0].role == 'system' %}
63
+ {%- set content = render_content(messages[0].content, false, true)|trim %}
64
+ {{- '<|im_start|>system\n' + content + '<|im_end|>\n' }}
65
+ {%- endif %}
66
+ {%- endif %}
67
+ {%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}
68
+ {%- for message in messages[::-1] %}
69
+ {%- set index = (messages|length - 1) - loop.index0 %}
70
+ {%- if ns.multi_step_tool and message.role == "user" %}
71
+ {%- set content = render_content(message.content, false)|trim %}
72
+ {%- if not(content.startswith('<tool_response>') and content.endswith('</tool_response>')) %}
73
+ {%- set ns.multi_step_tool = false %}
74
+ {%- set ns.last_query_index = index %}
75
+ {%- endif %}
76
+ {%- endif %}
77
+ {%- endfor %}
78
+ {%- if ns.multi_step_tool %}
79
+ {{- raise_exception('No user query found in messages.') }}
80
+ {%- endif %}
81
+ {%- for message in messages %}
82
+ {%- set content = render_content(message.content, true)|trim %}
83
+ {%- if message.role == "system" %}
84
+ {%- if not loop.first %}
85
+ {{- raise_exception('System message must be at the beginning.') }}
86
+ {%- endif %}
87
+ {%- elif message.role == "user" %}
88
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
89
+ {%- elif message.role == "assistant" %}
90
+ {%- set reasoning_content = '' %}
91
+ {%- if message.reasoning_content is string %}
92
+ {%- set reasoning_content = message.reasoning_content %}
93
+ {%- else %}
94
+ {%- if '</think>' in content %}
95
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
96
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
97
+ {%- endif %}
98
+ {%- endif %}
99
+ {%- set reasoning_content = reasoning_content|trim %}
100
+ {%- if (preserve_thinking is defined and preserve_thinking is true) or (loop.index0 > ns.last_query_index) %}
101
+ {{- '<|im_start|>' + message.role + '\n<think>\n' + reasoning_content + '\n</think>\n\n' + content }}
102
+ {%- else %}
103
+ {{- '<|im_start|>' + message.role + '\n' + content }}
104
+ {%- endif %}
105
+ {%- if message.tool_calls and message.tool_calls is iterable and message.tool_calls is not mapping %}
106
+ {%- for tool_call in message.tool_calls %}
107
+ {%- if tool_call.function is defined %}
108
+ {%- set tool_call = tool_call.function %}
109
+ {%- endif %}
110
+ {%- if loop.first %}
111
+ {%- if content|trim %}
112
+ {{- '\n\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
113
+ {%- else %}
114
+ {{- '<tool_call>\n<function=' + tool_call.name + '>\n' }}
115
+ {%- endif %}
116
+ {%- else %}
117
+ {{- '\n<tool_call>\n<function=' + tool_call.name + '>\n' }}
118
+ {%- endif %}
119
+ {%- if tool_call.arguments is defined %}
120
+ {%- for args_name, args_value in tool_call.arguments|items %}
121
+ {{- '<parameter=' + args_name + '>\n' }}
122
+ {%- set args_value = args_value | string if args_value is string else args_value | tojson | safe %}
123
+ {{- args_value }}
124
+ {{- '\n</parameter>\n' }}
125
+ {%- endfor %}
126
+ {%- endif %}
127
+ {{- '</function>\n</tool_call>' }}
128
+ {%- endfor %}
129
+ {%- endif %}
130
+ {{- '<|im_end|>\n' }}
131
+ {%- elif message.role == "tool" %}
132
+ {%- if loop.previtem and loop.previtem.role != "tool" %}
133
+ {{- '<|im_start|>user' }}
134
+ {%- endif %}
135
+ {{- '\n<tool_response>\n' }}
136
+ {{- content }}
137
+ {{- '\n</tool_response>' }}
138
+ {%- if not loop.last and loop.nextitem.role != "tool" %}
139
+ {{- '<|im_end|>\n' }}
140
+ {%- elif loop.last %}
141
+ {{- '<|im_end|>\n' }}
142
+ {%- endif %}
143
+ {%- else %}
144
+ {{- raise_exception('Unexpected message role.') }}
145
+ {%- endif %}
146
+ {%- endfor %}
147
+ {%- if add_generation_prompt %}
148
+ {{- '<|im_start|>assistant\n' }}
149
+ {%- if enable_thinking is defined and enable_thinking is false %}
150
+ {{- '<think>\n\n</think>\n\n' }}
151
+ {%- else %}
152
+ {{- '<think>\n' }}
153
+ {%- endif %}
154
+ {%- endif %}
config.json ADDED
@@ -0,0 +1,654 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3_5ForConditionalGeneration"
4
+ ],
5
+ "eos_token_id": [
6
+ 248046,
7
+ 248044
8
+ ],
9
+ "image_token_id": 248056,
10
+ "language_model_only": false,
11
+ "model_type": "qwen3_5",
12
+ "quantization": {
13
+ "group_size": 64,
14
+ "bits": 5,
15
+ "mode": "affine",
16
+ "language_model.model.embed_tokens": {
17
+ "bits": 8,
18
+ "group_size": 64,
19
+ "mode": "affine"
20
+ },
21
+ "language_model.model.layers.0.linear_attn.out_proj": {
22
+ "bits": 5,
23
+ "group_size": 64,
24
+ "mode": "affine"
25
+ },
26
+ "language_model.model.layers.1.linear_attn.out_proj": {
27
+ "bits": 5,
28
+ "group_size": 64,
29
+ "mode": "affine"
30
+ },
31
+ "language_model.model.layers.2.linear_attn.out_proj": {
32
+ "bits": 5,
33
+ "group_size": 64,
34
+ "mode": "affine"
35
+ },
36
+ "language_model.model.layers.4.linear_attn.out_proj": {
37
+ "bits": 5,
38
+ "group_size": 64,
39
+ "mode": "affine"
40
+ },
41
+ "language_model.model.layers.5.linear_attn.out_proj": {
42
+ "bits": 5,
43
+ "group_size": 64,
44
+ "mode": "affine"
45
+ },
46
+ "language_model.model.layers.6.linear_attn.out_proj": {
47
+ "bits": 5,
48
+ "group_size": 64,
49
+ "mode": "affine"
50
+ },
51
+ "language_model.model.layers.8.linear_attn.out_proj": {
52
+ "bits": 5,
53
+ "group_size": 64,
54
+ "mode": "affine"
55
+ },
56
+ "language_model.model.layers.9.linear_attn.out_proj": {
57
+ "bits": 5,
58
+ "group_size": 64,
59
+ "mode": "affine"
60
+ },
61
+ "language_model.model.layers.10.linear_attn.out_proj": {
62
+ "bits": 5,
63
+ "group_size": 64,
64
+ "mode": "affine"
65
+ },
66
+ "language_model.model.layers.12.linear_attn.out_proj": {
67
+ "bits": 5,
68
+ "group_size": 64,
69
+ "mode": "affine"
70
+ },
71
+ "language_model.model.layers.13.linear_attn.out_proj": {
72
+ "bits": 5,
73
+ "group_size": 64,
74
+ "mode": "affine"
75
+ },
76
+ "language_model.model.layers.14.linear_attn.out_proj": {
77
+ "bits": 5,
78
+ "group_size": 64,
79
+ "mode": "affine"
80
+ },
81
+ "language_model.model.layers.16.linear_attn.out_proj": {
82
+ "bits": 5,
83
+ "group_size": 64,
84
+ "mode": "affine"
85
+ },
86
+ "language_model.model.layers.17.linear_attn.out_proj": {
87
+ "bits": 5,
88
+ "group_size": 64,
89
+ "mode": "affine"
90
+ },
91
+ "language_model.model.layers.18.linear_attn.out_proj": {
92
+ "bits": 5,
93
+ "group_size": 64,
94
+ "mode": "affine"
95
+ },
96
+ "language_model.model.layers.20.linear_attn.out_proj": {
97
+ "bits": 5,
98
+ "group_size": 64,
99
+ "mode": "affine"
100
+ },
101
+ "language_model.model.layers.21.linear_attn.out_proj": {
102
+ "bits": 5,
103
+ "group_size": 64,
104
+ "mode": "affine"
105
+ },
106
+ "language_model.model.layers.22.linear_attn.out_proj": {
107
+ "bits": 5,
108
+ "group_size": 64,
109
+ "mode": "affine"
110
+ },
111
+ "language_model.model.layers.24.linear_attn.out_proj": {
112
+ "bits": 5,
113
+ "group_size": 64,
114
+ "mode": "affine"
115
+ },
116
+ "language_model.model.layers.25.linear_attn.out_proj": {
117
+ "bits": 5,
118
+ "group_size": 64,
119
+ "mode": "affine"
120
+ },
121
+ "language_model.model.layers.26.linear_attn.out_proj": {
122
+ "bits": 5,
123
+ "group_size": 64,
124
+ "mode": "affine"
125
+ },
126
+ "language_model.model.layers.28.linear_attn.out_proj": {
127
+ "bits": 5,
128
+ "group_size": 64,
129
+ "mode": "affine"
130
+ },
131
+ "language_model.model.layers.29.linear_attn.out_proj": {
132
+ "bits": 5,
133
+ "group_size": 64,
134
+ "mode": "affine"
135
+ },
136
+ "language_model.model.layers.30.linear_attn.out_proj": {
137
+ "bits": 5,
138
+ "group_size": 64,
139
+ "mode": "affine"
140
+ },
141
+ "language_model.model.layers.32.linear_attn.out_proj": {
142
+ "bits": 5,
143
+ "group_size": 64,
144
+ "mode": "affine"
145
+ },
146
+ "language_model.model.layers.33.linear_attn.out_proj": {
147
+ "bits": 5,
148
+ "group_size": 64,
149
+ "mode": "affine"
150
+ },
151
+ "language_model.model.layers.34.linear_attn.out_proj": {
152
+ "bits": 5,
153
+ "group_size": 64,
154
+ "mode": "affine"
155
+ },
156
+ "language_model.model.layers.36.linear_attn.out_proj": {
157
+ "bits": 5,
158
+ "group_size": 64,
159
+ "mode": "affine"
160
+ },
161
+ "language_model.model.layers.37.linear_attn.out_proj": {
162
+ "bits": 5,
163
+ "group_size": 64,
164
+ "mode": "affine"
165
+ },
166
+ "language_model.model.layers.38.linear_attn.out_proj": {
167
+ "bits": 5,
168
+ "group_size": 64,
169
+ "mode": "affine"
170
+ },
171
+ "language_model.model.layers.40.linear_attn.out_proj": {
172
+ "bits": 5,
173
+ "group_size": 64,
174
+ "mode": "affine"
175
+ },
176
+ "language_model.model.layers.41.linear_attn.out_proj": {
177
+ "bits": 5,
178
+ "group_size": 64,
179
+ "mode": "affine"
180
+ },
181
+ "language_model.model.layers.42.linear_attn.out_proj": {
182
+ "bits": 5,
183
+ "group_size": 64,
184
+ "mode": "affine"
185
+ },
186
+ "language_model.model.layers.44.linear_attn.out_proj": {
187
+ "bits": 5,
188
+ "group_size": 64,
189
+ "mode": "affine"
190
+ },
191
+ "language_model.model.layers.45.linear_attn.out_proj": {
192
+ "bits": 5,
193
+ "group_size": 64,
194
+ "mode": "affine"
195
+ },
196
+ "language_model.model.layers.46.linear_attn.out_proj": {
197
+ "bits": 5,
198
+ "group_size": 64,
199
+ "mode": "affine"
200
+ },
201
+ "language_model.model.layers.48.linear_attn.out_proj": {
202
+ "bits": 5,
203
+ "group_size": 64,
204
+ "mode": "affine"
205
+ },
206
+ "language_model.model.layers.49.linear_attn.out_proj": {
207
+ "bits": 5,
208
+ "group_size": 64,
209
+ "mode": "affine"
210
+ },
211
+ "language_model.model.layers.50.linear_attn.out_proj": {
212
+ "bits": 5,
213
+ "group_size": 64,
214
+ "mode": "affine"
215
+ },
216
+ "language_model.model.layers.52.linear_attn.out_proj": {
217
+ "bits": 5,
218
+ "group_size": 64,
219
+ "mode": "affine"
220
+ },
221
+ "language_model.model.layers.53.linear_attn.out_proj": {
222
+ "bits": 5,
223
+ "group_size": 64,
224
+ "mode": "affine"
225
+ },
226
+ "language_model.model.layers.54.linear_attn.out_proj": {
227
+ "bits": 5,
228
+ "group_size": 64,
229
+ "mode": "affine"
230
+ },
231
+ "language_model.model.layers.56.linear_attn.out_proj": {
232
+ "bits": 5,
233
+ "group_size": 64,
234
+ "mode": "affine"
235
+ },
236
+ "language_model.model.layers.57.linear_attn.out_proj": {
237
+ "bits": 5,
238
+ "group_size": 64,
239
+ "mode": "affine"
240
+ },
241
+ "language_model.model.layers.58.linear_attn.out_proj": {
242
+ "bits": 5,
243
+ "group_size": 64,
244
+ "mode": "affine"
245
+ },
246
+ "language_model.model.layers.60.linear_attn.out_proj": {
247
+ "bits": 5,
248
+ "group_size": 64,
249
+ "mode": "affine"
250
+ },
251
+ "language_model.model.layers.61.linear_attn.out_proj": {
252
+ "bits": 5,
253
+ "group_size": 64,
254
+ "mode": "affine"
255
+ },
256
+ "language_model.model.layers.62.linear_attn.out_proj": {
257
+ "bits": 5,
258
+ "group_size": 64,
259
+ "mode": "affine"
260
+ },
261
+ "language_model.lm_head": {
262
+ "bits": 8,
263
+ "group_size": 64,
264
+ "mode": "affine"
265
+ }
266
+ },
267
+ "quantization_config": {
268
+ "group_size": 64,
269
+ "bits": 5,
270
+ "mode": "affine",
271
+ "language_model.model.embed_tokens": {
272
+ "bits": 8,
273
+ "group_size": 64,
274
+ "mode": "affine"
275
+ },
276
+ "language_model.model.layers.0.linear_attn.out_proj": {
277
+ "bits": 5,
278
+ "group_size": 64,
279
+ "mode": "affine"
280
+ },
281
+ "language_model.model.layers.1.linear_attn.out_proj": {
282
+ "bits": 5,
283
+ "group_size": 64,
284
+ "mode": "affine"
285
+ },
286
+ "language_model.model.layers.2.linear_attn.out_proj": {
287
+ "bits": 5,
288
+ "group_size": 64,
289
+ "mode": "affine"
290
+ },
291
+ "language_model.model.layers.4.linear_attn.out_proj": {
292
+ "bits": 5,
293
+ "group_size": 64,
294
+ "mode": "affine"
295
+ },
296
+ "language_model.model.layers.5.linear_attn.out_proj": {
297
+ "bits": 5,
298
+ "group_size": 64,
299
+ "mode": "affine"
300
+ },
301
+ "language_model.model.layers.6.linear_attn.out_proj": {
302
+ "bits": 5,
303
+ "group_size": 64,
304
+ "mode": "affine"
305
+ },
306
+ "language_model.model.layers.8.linear_attn.out_proj": {
307
+ "bits": 5,
308
+ "group_size": 64,
309
+ "mode": "affine"
310
+ },
311
+ "language_model.model.layers.9.linear_attn.out_proj": {
312
+ "bits": 5,
313
+ "group_size": 64,
314
+ "mode": "affine"
315
+ },
316
+ "language_model.model.layers.10.linear_attn.out_proj": {
317
+ "bits": 5,
318
+ "group_size": 64,
319
+ "mode": "affine"
320
+ },
321
+ "language_model.model.layers.12.linear_attn.out_proj": {
322
+ "bits": 5,
323
+ "group_size": 64,
324
+ "mode": "affine"
325
+ },
326
+ "language_model.model.layers.13.linear_attn.out_proj": {
327
+ "bits": 5,
328
+ "group_size": 64,
329
+ "mode": "affine"
330
+ },
331
+ "language_model.model.layers.14.linear_attn.out_proj": {
332
+ "bits": 5,
333
+ "group_size": 64,
334
+ "mode": "affine"
335
+ },
336
+ "language_model.model.layers.16.linear_attn.out_proj": {
337
+ "bits": 5,
338
+ "group_size": 64,
339
+ "mode": "affine"
340
+ },
341
+ "language_model.model.layers.17.linear_attn.out_proj": {
342
+ "bits": 5,
343
+ "group_size": 64,
344
+ "mode": "affine"
345
+ },
346
+ "language_model.model.layers.18.linear_attn.out_proj": {
347
+ "bits": 5,
348
+ "group_size": 64,
349
+ "mode": "affine"
350
+ },
351
+ "language_model.model.layers.20.linear_attn.out_proj": {
352
+ "bits": 5,
353
+ "group_size": 64,
354
+ "mode": "affine"
355
+ },
356
+ "language_model.model.layers.21.linear_attn.out_proj": {
357
+ "bits": 5,
358
+ "group_size": 64,
359
+ "mode": "affine"
360
+ },
361
+ "language_model.model.layers.22.linear_attn.out_proj": {
362
+ "bits": 5,
363
+ "group_size": 64,
364
+ "mode": "affine"
365
+ },
366
+ "language_model.model.layers.24.linear_attn.out_proj": {
367
+ "bits": 5,
368
+ "group_size": 64,
369
+ "mode": "affine"
370
+ },
371
+ "language_model.model.layers.25.linear_attn.out_proj": {
372
+ "bits": 5,
373
+ "group_size": 64,
374
+ "mode": "affine"
375
+ },
376
+ "language_model.model.layers.26.linear_attn.out_proj": {
377
+ "bits": 5,
378
+ "group_size": 64,
379
+ "mode": "affine"
380
+ },
381
+ "language_model.model.layers.28.linear_attn.out_proj": {
382
+ "bits": 5,
383
+ "group_size": 64,
384
+ "mode": "affine"
385
+ },
386
+ "language_model.model.layers.29.linear_attn.out_proj": {
387
+ "bits": 5,
388
+ "group_size": 64,
389
+ "mode": "affine"
390
+ },
391
+ "language_model.model.layers.30.linear_attn.out_proj": {
392
+ "bits": 5,
393
+ "group_size": 64,
394
+ "mode": "affine"
395
+ },
396
+ "language_model.model.layers.32.linear_attn.out_proj": {
397
+ "bits": 5,
398
+ "group_size": 64,
399
+ "mode": "affine"
400
+ },
401
+ "language_model.model.layers.33.linear_attn.out_proj": {
402
+ "bits": 5,
403
+ "group_size": 64,
404
+ "mode": "affine"
405
+ },
406
+ "language_model.model.layers.34.linear_attn.out_proj": {
407
+ "bits": 5,
408
+ "group_size": 64,
409
+ "mode": "affine"
410
+ },
411
+ "language_model.model.layers.36.linear_attn.out_proj": {
412
+ "bits": 5,
413
+ "group_size": 64,
414
+ "mode": "affine"
415
+ },
416
+ "language_model.model.layers.37.linear_attn.out_proj": {
417
+ "bits": 5,
418
+ "group_size": 64,
419
+ "mode": "affine"
420
+ },
421
+ "language_model.model.layers.38.linear_attn.out_proj": {
422
+ "bits": 5,
423
+ "group_size": 64,
424
+ "mode": "affine"
425
+ },
426
+ "language_model.model.layers.40.linear_attn.out_proj": {
427
+ "bits": 5,
428
+ "group_size": 64,
429
+ "mode": "affine"
430
+ },
431
+ "language_model.model.layers.41.linear_attn.out_proj": {
432
+ "bits": 5,
433
+ "group_size": 64,
434
+ "mode": "affine"
435
+ },
436
+ "language_model.model.layers.42.linear_attn.out_proj": {
437
+ "bits": 5,
438
+ "group_size": 64,
439
+ "mode": "affine"
440
+ },
441
+ "language_model.model.layers.44.linear_attn.out_proj": {
442
+ "bits": 5,
443
+ "group_size": 64,
444
+ "mode": "affine"
445
+ },
446
+ "language_model.model.layers.45.linear_attn.out_proj": {
447
+ "bits": 5,
448
+ "group_size": 64,
449
+ "mode": "affine"
450
+ },
451
+ "language_model.model.layers.46.linear_attn.out_proj": {
452
+ "bits": 5,
453
+ "group_size": 64,
454
+ "mode": "affine"
455
+ },
456
+ "language_model.model.layers.48.linear_attn.out_proj": {
457
+ "bits": 5,
458
+ "group_size": 64,
459
+ "mode": "affine"
460
+ },
461
+ "language_model.model.layers.49.linear_attn.out_proj": {
462
+ "bits": 5,
463
+ "group_size": 64,
464
+ "mode": "affine"
465
+ },
466
+ "language_model.model.layers.50.linear_attn.out_proj": {
467
+ "bits": 5,
468
+ "group_size": 64,
469
+ "mode": "affine"
470
+ },
471
+ "language_model.model.layers.52.linear_attn.out_proj": {
472
+ "bits": 5,
473
+ "group_size": 64,
474
+ "mode": "affine"
475
+ },
476
+ "language_model.model.layers.53.linear_attn.out_proj": {
477
+ "bits": 5,
478
+ "group_size": 64,
479
+ "mode": "affine"
480
+ },
481
+ "language_model.model.layers.54.linear_attn.out_proj": {
482
+ "bits": 5,
483
+ "group_size": 64,
484
+ "mode": "affine"
485
+ },
486
+ "language_model.model.layers.56.linear_attn.out_proj": {
487
+ "bits": 5,
488
+ "group_size": 64,
489
+ "mode": "affine"
490
+ },
491
+ "language_model.model.layers.57.linear_attn.out_proj": {
492
+ "bits": 5,
493
+ "group_size": 64,
494
+ "mode": "affine"
495
+ },
496
+ "language_model.model.layers.58.linear_attn.out_proj": {
497
+ "bits": 5,
498
+ "group_size": 64,
499
+ "mode": "affine"
500
+ },
501
+ "language_model.model.layers.60.linear_attn.out_proj": {
502
+ "bits": 5,
503
+ "group_size": 64,
504
+ "mode": "affine"
505
+ },
506
+ "language_model.model.layers.61.linear_attn.out_proj": {
507
+ "bits": 5,
508
+ "group_size": 64,
509
+ "mode": "affine"
510
+ },
511
+ "language_model.model.layers.62.linear_attn.out_proj": {
512
+ "bits": 5,
513
+ "group_size": 64,
514
+ "mode": "affine"
515
+ },
516
+ "language_model.lm_head": {
517
+ "bits": 8,
518
+ "group_size": 64,
519
+ "mode": "affine"
520
+ }
521
+ },
522
+ "text_config": {
523
+ "attention_bias": false,
524
+ "attention_dropout": 0.0,
525
+ "attn_output_gate": true,
526
+ "bos_token_id": 248044,
527
+ "dtype": "bfloat16",
528
+ "eos_token_id": 248044,
529
+ "full_attention_interval": 4,
530
+ "head_dim": 256,
531
+ "hidden_act": "silu",
532
+ "hidden_size": 5120,
533
+ "initializer_range": 0.02,
534
+ "intermediate_size": 17408,
535
+ "layer_types": [
536
+ "linear_attention",
537
+ "linear_attention",
538
+ "linear_attention",
539
+ "full_attention",
540
+ "linear_attention",
541
+ "linear_attention",
542
+ "linear_attention",
543
+ "full_attention",
544
+ "linear_attention",
545
+ "linear_attention",
546
+ "linear_attention",
547
+ "full_attention",
548
+ "linear_attention",
549
+ "linear_attention",
550
+ "linear_attention",
551
+ "full_attention",
552
+ "linear_attention",
553
+ "linear_attention",
554
+ "linear_attention",
555
+ "full_attention",
556
+ "linear_attention",
557
+ "linear_attention",
558
+ "linear_attention",
559
+ "full_attention",
560
+ "linear_attention",
561
+ "linear_attention",
562
+ "linear_attention",
563
+ "full_attention",
564
+ "linear_attention",
565
+ "linear_attention",
566
+ "linear_attention",
567
+ "full_attention",
568
+ "linear_attention",
569
+ "linear_attention",
570
+ "linear_attention",
571
+ "full_attention",
572
+ "linear_attention",
573
+ "linear_attention",
574
+ "linear_attention",
575
+ "full_attention",
576
+ "linear_attention",
577
+ "linear_attention",
578
+ "linear_attention",
579
+ "full_attention",
580
+ "linear_attention",
581
+ "linear_attention",
582
+ "linear_attention",
583
+ "full_attention",
584
+ "linear_attention",
585
+ "linear_attention",
586
+ "linear_attention",
587
+ "full_attention",
588
+ "linear_attention",
589
+ "linear_attention",
590
+ "linear_attention",
591
+ "full_attention",
592
+ "linear_attention",
593
+ "linear_attention",
594
+ "linear_attention",
595
+ "full_attention",
596
+ "linear_attention",
597
+ "linear_attention",
598
+ "linear_attention",
599
+ "full_attention"
600
+ ],
601
+ "linear_conv_kernel_dim": 4,
602
+ "linear_key_head_dim": 128,
603
+ "linear_num_key_heads": 16,
604
+ "linear_num_value_heads": 48,
605
+ "linear_value_head_dim": 128,
606
+ "mamba_ssm_dtype": "float32",
607
+ "max_position_embeddings": 262144,
608
+ "model_type": "qwen3_5_text",
609
+ "mtp_num_hidden_layers": 1,
610
+ "mtp_use_dedicated_embeddings": false,
611
+ "num_attention_heads": 24,
612
+ "num_hidden_layers": 64,
613
+ "num_key_value_heads": 4,
614
+ "output_gate_type": "swish",
615
+ "pad_token_id": null,
616
+ "partial_rotary_factor": 0.25,
617
+ "rms_norm_eps": 1e-06,
618
+ "rope_parameters": {
619
+ "mrope_interleaved": true,
620
+ "mrope_section": [
621
+ 11,
622
+ 11,
623
+ 10
624
+ ],
625
+ "partial_rotary_factor": 0.25,
626
+ "rope_theta": 10000000,
627
+ "rope_type": "default"
628
+ },
629
+ "tie_word_embeddings": false,
630
+ "use_cache": true,
631
+ "vocab_size": 248320
632
+ },
633
+ "tie_word_embeddings": false,
634
+ "transformers_version": "4.57.1",
635
+ "video_token_id": 248057,
636
+ "vision_config": {
637
+ "deepstack_visual_indexes": [],
638
+ "depth": 27,
639
+ "hidden_act": "gelu_pytorch_tanh",
640
+ "hidden_size": 1152,
641
+ "in_channels": 3,
642
+ "initializer_range": 0.02,
643
+ "intermediate_size": 4304,
644
+ "model_type": "qwen3_5",
645
+ "num_heads": 16,
646
+ "num_position_embeddings": 2304,
647
+ "out_hidden_size": 5120,
648
+ "patch_size": 16,
649
+ "spatial_merge_size": 2,
650
+ "temporal_patch_size": 2
651
+ },
652
+ "vision_end_token_id": 248054,
653
+ "vision_start_token_id": 248053
654
+ }
configuration.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"framework":"Pytorch","task":"image-text-to-text"}
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 248044,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 248046,
6
+ 248044
7
+ ],
8
+ "pad_token_id": 248044,
9
+ "temperature": 1.0,
10
+ "top_k": 20,
11
+ "top_p": 0.95
12
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8607e425d696342167d0826294cb2268da557794507dab25222b0b4f7f665fc
3
+ size 5351056366
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ce1d8d8ca3467593171094f0c79d50cece73b722342a266a8a3e4654036a4c2
3
+ size 5352661019
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d81f377f79e3a23560953d8442f19650639b003453394038b5e545a6e07fca6c
3
+ size 5316214974
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ad672bd3b9cf837f3d679efff597b0f9aa9ba7f8517df0d459ab270701c0f1
3
+ size 4349827000
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "size": {
3
+ "longest_edge": 16777216,
4
+ "shortest_edge": 65536
5
+ },
6
+ "patch_size": 16,
7
+ "temporal_patch_size": 2,
8
+ "merge_size": 2,
9
+ "image_mean": [
10
+ 0.5,
11
+ 0.5,
12
+ 0.5
13
+ ],
14
+ "image_std": [
15
+ 0.5,
16
+ 0.5,
17
+ 0.5
18
+ ],
19
+ "processor_class": "Qwen3VLProcessor",
20
+ "image_processor_type": "Qwen2VLImageProcessorFast"
21
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f9e4d4901a92b997e463c1f46055088b6cca5ca61a6522d1b9f64c4bb81cb42
3
+ size 12807982
tokenizer_config.json ADDED
@@ -0,0 +1,305 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "248044": {
5
+ "content": "<|endoftext|>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "248045": {
13
+ "content": "<|im_start|>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "248046": {
21
+ "content": "<|im_end|>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "248047": {
29
+ "content": "<|object_ref_start|>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "248048": {
37
+ "content": "<|object_ref_end|>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "248049": {
45
+ "content": "<|box_start|>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "248050": {
53
+ "content": "<|box_end|>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "248051": {
61
+ "content": "<|quad_start|>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "248052": {
69
+ "content": "<|quad_end|>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "248053": {
77
+ "content": "<|vision_start|>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "248054": {
85
+ "content": "<|vision_end|>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "248055": {
93
+ "content": "<|vision_pad|>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "248056": {
101
+ "content": "<|image_pad|>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "248057": {
109
+ "content": "<|video_pad|>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "248058": {
117
+ "content": "<tool_call>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": false
123
+ },
124
+ "248059": {
125
+ "content": "</tool_call>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": false
131
+ },
132
+ "248060": {
133
+ "content": "<|fim_prefix|>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": false
139
+ },
140
+ "248061": {
141
+ "content": "<|fim_middle|>",
142
+ "lstrip": false,
143
+ "normalized": false,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": false
147
+ },
148
+ "248062": {
149
+ "content": "<|fim_suffix|>",
150
+ "lstrip": false,
151
+ "normalized": false,
152
+ "rstrip": false,
153
+ "single_word": false,
154
+ "special": false
155
+ },
156
+ "248063": {
157
+ "content": "<|fim_pad|>",
158
+ "lstrip": false,
159
+ "normalized": false,
160
+ "rstrip": false,
161
+ "single_word": false,
162
+ "special": false
163
+ },
164
+ "248064": {
165
+ "content": "<|repo_name|>",
166
+ "lstrip": false,
167
+ "normalized": false,
168
+ "rstrip": false,
169
+ "single_word": false,
170
+ "special": false
171
+ },
172
+ "248065": {
173
+ "content": "<|file_sep|>",
174
+ "lstrip": false,
175
+ "normalized": false,
176
+ "rstrip": false,
177
+ "single_word": false,
178
+ "special": false
179
+ },
180
+ "248066": {
181
+ "content": "<tool_response>",
182
+ "lstrip": false,
183
+ "normalized": false,
184
+ "rstrip": false,
185
+ "single_word": false,
186
+ "special": false
187
+ },
188
+ "248067": {
189
+ "content": "</tool_response>",
190
+ "lstrip": false,
191
+ "normalized": false,
192
+ "rstrip": false,
193
+ "single_word": false,
194
+ "special": false
195
+ },
196
+ "248068": {
197
+ "content": "<think>",
198
+ "lstrip": false,
199
+ "normalized": false,
200
+ "rstrip": false,
201
+ "single_word": false,
202
+ "special": false
203
+ },
204
+ "248069": {
205
+ "content": "</think>",
206
+ "lstrip": false,
207
+ "normalized": false,
208
+ "rstrip": false,
209
+ "single_word": false,
210
+ "special": false
211
+ },
212
+ "248070": {
213
+ "content": "<|audio_start|>",
214
+ "lstrip": false,
215
+ "normalized": false,
216
+ "rstrip": false,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "248071": {
221
+ "content": "<|audio_end|>",
222
+ "lstrip": false,
223
+ "normalized": false,
224
+ "rstrip": false,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "248072": {
229
+ "content": "<tts_pad>",
230
+ "lstrip": false,
231
+ "normalized": false,
232
+ "rstrip": false,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "248073": {
237
+ "content": "<tts_text_bos>",
238
+ "lstrip": false,
239
+ "normalized": false,
240
+ "rstrip": false,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "248074": {
245
+ "content": "<tts_text_eod>",
246
+ "lstrip": false,
247
+ "normalized": false,
248
+ "rstrip": false,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "248075": {
253
+ "content": "<tts_text_bos_single>",
254
+ "lstrip": false,
255
+ "normalized": false,
256
+ "rstrip": false,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "248076": {
261
+ "content": "<|audio_pad|>",
262
+ "lstrip": false,
263
+ "normalized": false,
264
+ "rstrip": false,
265
+ "single_word": false,
266
+ "special": true
267
+ }
268
+ },
269
+ "additional_special_tokens": [
270
+ "<|im_start|>",
271
+ "<|im_end|>",
272
+ "<|object_ref_start|>",
273
+ "<|object_ref_end|>",
274
+ "<|box_start|>",
275
+ "<|box_end|>",
276
+ "<|quad_start|>",
277
+ "<|quad_end|>",
278
+ "<|vision_start|>",
279
+ "<|vision_end|>",
280
+ "<|vision_pad|>",
281
+ "<|image_pad|>",
282
+ "<|video_pad|>"
283
+ ],
284
+ "bos_token": null,
285
+ "chat_template": "{%- set image_count = namespace(value=0) %}\n{%- set video_count = namespace(value=0) %}\n{%- macro render_content(content, do_vision_count, is_system_content=false) %}\n {%- if content is string %}\n {{- content }}\n {%- elif content is iterable and content is not mapping %}\n {%- for item in content %}\n {%- if 'image' in item or 'image_url' in item or item.type == 'image' %}\n {%- if is_system_content %}\n {{- raise_exception('System message cannot contain images.') }}\n {%- endif %}\n {%- if do_vision_count %}\n {%- set image_count.value = image_count.value + 1 %}\n {%- endif %}\n {%- if add_vision_id %}\n {{- 'Picture ' ~ image_count.value ~ ': ' }}\n {%- endif %}\n {{- '<|vision_start|><|image_pad|><|vision_end|>' }}\n {%- elif 'video' in item or item.type == 'video' %}\n {%- if is_system_content %}\n {{- raise_exception('System message cannot contain videos.') }}\n {%- endif %}\n {%- if do_vision_count %}\n {%- set video_count.value = video_count.value + 1 %}\n {%- endif %}\n {%- if add_vision_id %}\n {{- 'Video ' ~ video_count.value ~ ': ' }}\n {%- endif %}\n {{- '<|vision_start|><|video_pad|><|vision_end|>' }}\n {%- elif 'text' in item %}\n {{- item.text }}\n {%- else %}\n {{- raise_exception('Unexpected item type in content.') }}\n {%- endif %}\n {%- endfor %}\n {%- elif content is none or content is undefined %}\n {{- '' }}\n {%- else %}\n {{- raise_exception('Unexpected content type.') }}\n {%- endif %}\n{%- endmacro %}\n{%- if not messages %}\n {{- raise_exception('No messages provided.') }}\n{%- endif %}\n{%- if tools and tools is iterable and tools is not mapping %}\n {{- '<|im_start|>system\\n' }}\n {{- \"# Tools\\n\\nYou have access to the following functions:\\n\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\" }}\n {{- '\\n\\nIf you choose to call a function ONLY reply in the following format with NO suffix:\\n\\n<tool_call>\\n<function=example_function_name>\\n<parameter=example_parameter_1>\\nvalue_1\\n</parameter>\\n<parameter=example_parameter_2>\\nThis is the value for the second parameter\\nthat can span\\nmultiple lines\\n</parameter>\\n</function>\\n</tool_call>\\n\\n<IMPORTANT>\\nReminder:\\n- Function calls MUST follow the specified format: an inner <function=...></function> block must be nested within <tool_call></tool_call> XML tags\\n- Required parameters MUST be specified\\n- You may provide optional reasoning for your function call in natural language BEFORE the function call, but NOT after\\n- If there is no function call available, answer the question like normal with your current knowledge and do not tell the user about function calls\\n</IMPORTANT>' }}\n {%- if messages[0].role == 'system' %}\n {%- set content = render_content(messages[0].content, false, true)|trim %}\n {%- if content %}\n {{- '\\n\\n' + content }}\n {%- endif %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {%- set content = render_content(messages[0].content, false, true)|trim %}\n {{- '<|im_start|>system\\n' + content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" %}\n {%- set content = render_content(message.content, false)|trim %}\n {%- if not(content.startswith('<tool_response>') and content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if ns.multi_step_tool %}\n {{- raise_exception('No user query found in messages.') }}\n{%- endif %}\n{%- for message in messages %}\n {%- set content = render_content(message.content, true)|trim %}\n {%- if message.role == \"system\" %}\n {%- if not loop.first %}\n {{- raise_exception('System message must be at the beginning.') }}\n {%- endif %}\n {%- elif message.role == \"user\" %}\n {{- '<|im_start|>' + message.role + '\\n' + content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is string %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in content %}\n {%- set reasoning_content = content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- set content = content.split('</think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- set reasoning_content = reasoning_content|trim %}\n {%- if (preserve_thinking is defined and preserve_thinking is true) or (loop.index0 > ns.last_query_index) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content + '\\n</think>\\n\\n' + content }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls and message.tool_calls is iterable and message.tool_calls is not mapping %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {%- if loop.first %}\n {%- if content|trim %}\n {{- '\\n\\n<tool_call>\\n<function=' + tool_call.name + '>\\n' }}\n {%- else %}\n {{- '<tool_call>\\n<function=' + tool_call.name + '>\\n' }}\n {%- endif %}\n {%- else %}\n {{- '\\n<tool_call>\\n<function=' + tool_call.name + '>\\n' }}\n {%- endif %}\n {%- if tool_call.arguments is defined %}\n {%- for args_name, args_value in tool_call.arguments|items %}\n {{- '<parameter=' + args_name + '>\\n' }}\n {%- set args_value = args_value | string if args_value is string else args_value | tojson | safe %}\n {{- args_value }}\n {{- '\\n</parameter>\\n' }}\n {%- endfor %}\n {%- endif %}\n {{- '</function>\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.previtem and loop.previtem.role != \"tool\" %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- content }}\n {{- '\\n</tool_response>' }}\n {%- if not loop.last and loop.nextitem.role != \"tool\" %}\n {{- '<|im_end|>\\n' }}\n {%- elif loop.last %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- else %}\n {{- raise_exception('Unexpected message role.') }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- else %}\n {{- '<think>\\n' }}\n {%- endif %}\n{%- endif %}",
286
+ "clean_up_tokenization_spaces": false,
287
+ "eos_token": "<|im_end|>",
288
+ "errors": "replace",
289
+ "model_max_length": 262144,
290
+ "pad_token": "<|endoftext|>",
291
+ "split_special_tokens": false,
292
+ "tokenizer_class": "Qwen2Tokenizer",
293
+ "unk_token": null,
294
+ "add_bos_token": false,
295
+ "pretokenize_regex": "(?i:'s|'t|'re|'ve|'m|'ll|'d)|[^\\r\\n\\p{L}\\p{N}]?[\\p{L}\\p{M}]+|\\p{N}| ?[^\\s\\p{L}\\p{M}\\p{N}]+[\\r\\n]*|\\s*[\\r\\n]+|\\s+(?!\\S)|\\s+",
296
+ "extra_special_tokens": {
297
+ "audio_bos_token": "<|audio_start|>",
298
+ "audio_eos_token": "<|audio_end|>",
299
+ "audio_token": "<|audio_pad|>",
300
+ "image_token": "<|image_pad|>",
301
+ "video_token": "<|video_pad|>",
302
+ "vision_bos_token": "<|vision_start|>",
303
+ "vision_eos_token": "<|vision_end|>"
304
+ }
305
+ }
video_preprocessor_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "size": {
3
+ "longest_edge": 25165824,
4
+ "shortest_edge": 4096
5
+ },
6
+ "patch_size": 16,
7
+ "temporal_patch_size": 2,
8
+ "merge_size": 2,
9
+ "image_mean": [
10
+ 0.5,
11
+ 0.5,
12
+ 0.5
13
+ ],
14
+ "image_std": [
15
+ 0.5,
16
+ 0.5,
17
+ 0.5
18
+ ],
19
+ "processor_class": "Qwen3VLProcessor",
20
+ "video_processor_type": "Qwen3VLVideoProcessor"
21
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff