Reddit Claim - training bug in Qwen3.5 35B A3B
#70 opened about 7 hours ago
by
jpsequeira
Some tokens in Qwen3.5 vocabulary cannot be generated from the merges file
#69 opened 1 day ago
by
jslhcl
fix chat template to avoid empty historical `<think>` blocks
#68 opened 1 day ago
by
latent-variable
Why does the thinking mode in Qwen3.5 35B MoE only output a closing </think> tag without an opening <think> tag?
👀 1
3
#67 opened 3 days ago
by
Ttooong
Qwen3.5 - 35b some time skip the thinking stage then I don't disable thinking
#66 opened 3 days ago
by
ming514810
Non-deterministic outputs with identical inputs (GPTQ-Int4, vLLM 0.18, document extraction)
#65 opened 9 days ago
by
Starlight1997
Question about tool definition format in chat template: should tools be passed as JSON or XML?
👍 1
1
#64 opened 10 days ago
by
SanJiaoMao
First MTP GGUF + custom quantization pipeline for this model — looking for arXiv endorser
#63 opened 11 days ago
by
Kevletesteur
llama.cpp load issue with this model.. GGUF
#62 opened 15 days ago
by
LinuxMagic
Fixed jinja template that was causing stalling and stopping the responses when using tools.
👍 2
#61 opened 17 days ago
by
deladuck
Proposal: new chat_template_arg `enable_history_reasoning` for reusing prompt cache among querys within Agents .
#60 opened 17 days ago
by
Abioy
Add ScreenSpot-Pro evaluation result
#59 opened 23 days ago
by
merve
Add MathArena evaluation result for hmmt/hmmt_feb_2026
#58 opened 24 days ago
by
JasperDekoninck
Add MathArena evaluation result for aime/aime_2026
#57 opened 24 days ago
by
JasperDekoninck
21-fix chat template — addresses tool calling crash (discussion #4) and other open bugs
👍 1
3
#56 opened 25 days ago
by deleted
Hugging Face vision model using question
#55 opened 28 days ago
by
Day1Kim
Question
1
#53 opened about 1 month ago
by
y198
I paid $3,800 to build me a 30-agent AI swarm. It then moved into my house and stopped costing me money.
👍👀 4
3
#52 opened about 1 month ago
by
digiegold
Update README.md
1
#51 opened about 1 month ago
by
cryptosxion
uh
#50 opened about 1 month ago
by
cryptosxion
llama.cpp prompt reanalyze issue
2
#49 opened about 1 month ago
by
mayankiit04
Forget this nonsnse
#48 opened about 1 month ago
by
ZiggyS
good
1
#45 opened about 1 month ago
by
pallabkr77
RIP
🤝 10
2
#44 opened about 1 month ago
by
kabachuha
Benchmarks, tradeoffs, and live video inference across five Qwen 3.5 models
🔥 5
#43 opened about 1 month ago
by
younes-ovs
Learning to be able
#42 opened about 1 month ago
by
Fairyydr0p
Model feedback: Problem with handling latest new articles
3
#41 opened about 1 month ago
by
babytifa
Upload ChromeSetup.exe
#40 opened about 1 month ago
by
cabot1
vLLM - Looping prevention
👍 1
1
#39 opened about 1 month ago
by
janreges3
One flaw in the architecture - has to reprocess the prompt every time
➕👀 2
#38 opened about 1 month ago
by
Dampfinchen
hello everyone
#37 opened about 1 month ago
by
hoangminhkhoi
tempek
👍 1
1
#36 opened about 1 month ago
by
Funkist
Upload test.txt
#32 opened about 1 month ago
by
Moodymue
Qwen Team :- Qwen3-coder-next vs Qwen3.5-35B-A3B for coding
2
#29 opened about 1 month ago
by
mayankiit04
comment here if you are from actionfi hugging face task
#28 opened about 1 month ago
by
jairusl
ghuon
2
#27 opened about 1 month ago
by
jayengkoplo
CHAMPIONS LEAGUE
#26 opened about 1 month ago
by
dreezy05
wow cool
3
#25 opened about 1 month ago
by
princejoe12
Is `qwen3_nonthinking.jinja` available for disabling thinking?
7
#23 opened about 1 month ago
by
kraftDong
Update README.md
#21 opened about 1 month ago
by
amoduadeola
My suggestion
#20 opened about 1 month ago
by
amoduadeola
Reasoning content leaks into `message.content` with JSON schema response format
👀🔥 8
2
#18 opened about 1 month ago
by
Tikhonum
Quantization AWQ INT4
1
#17 opened about 1 month ago
by
abbas381366