0.8B with mtp is slower than without mtp
#15 opened 7 days ago
by
WEISHU
fix chat template to avoid empty historical `<think>` blocks
1
#14 opened 8 days ago
by
latent-variable
When vllm deployment is started using mtp config, an error will be reported. Please help!
2
#13 opened 13 days ago
by
WEISHU
anis
#12 opened 28 days ago
by
Aniskhan29
Add ScreenSpot-Pro evaluation result
#11 opened 29 days ago
by
merve
How to change the model to ONNX format
#9 opened about 1 month ago
by
WangJH97
Can I run this on the macbook apple Silicon?
3
#8 opened about 1 month ago
by
anuedaraanu
Request format for transformers serve for Qwen3.5
➕ 6
1
#7 opened about 1 month ago
by
Crockrocks12
Is there a comparison between Qwen3.5-0.8B and Qwen3-0.6B?
1
#6 opened about 1 month ago
by
sssssp
Can I run these models on google collab ? Please Help!!
1
#5 opened about 1 month ago
by
saikumar305
Could someone make Qwen/Qwen3.5-0.4B?
3
#4 opened about 1 month ago
by
MihaiPopa-1
Will there be embedding only versions of these models?
👀 2
#3 opened about 1 month ago
by
vincevannnoort
Installation Video and Testing - Step by Step
👍 1
1
#2 opened about 1 month ago
by
fahdmirzac
Add MMLU-Pro evaluation result
1
#1 opened about 1 month ago
by
SaylorTwift