
QwQ-32B, from Alibaba Qwen team, is a new open-source 32B LLM achieving DeepSeek-R1 level reasoning via scaled Reinforcement Learning. Features a "thinking mode" for complex tasks.
QwQ-32B, from Alibaba Qwen team, is a new open-source 32B LLM achieving DeepSeek-R1 level reasoning via scaled Reinforcement Learning. Features a "thinking mode" for complex tasks.
Hi everyone!
Check out QwQ-32B, a new open-source language model from the Qwen team. It's achieving something remarkable: reasoning performance comparable to DeepSeek-R1, but with a model that's 20 times smaller (32B parameters vs. 671B)!
This is a big deal because:
🤯 Size/Performance Ratio: It punches way above its weight class in reasoning, math, and coding tasks.
🧠 Scaled Reinforcement Learning: They achieved this by scaling up Reinforcement Learning (RL) on a strong foundation model (Qwen2.5-32B).
🤔 "Thinking Mode": Like some other recent models, it has a special "thinking mode" (activated with tags) that allows for longer chains of thought.
✅ Open Source: Model weights are available under Apache 2.0.
Available now in Qwen Chat and HF Spaces.
The implications of getting this level of reasoning performance from a 32B model are huge. It opens up possibilities for deploying powerful AI on less powerful hardware, reducing costs, and making advanced reasoning capabilities more accessible.
@sonu_goswami2 It reveals all the details of its reasoning, clear and transparent – just like R1!
Grimo
@zaczuo cmon this is so fckin good! we've been playing with it on qwen chat as well as groq cloud, performance was astonishing! great job qwen team!
@sonu_goswami2 @zaczuo Nice! Will check it out!
I use hugging face every day for my job. I didn't know that models there can be launched on PH. Very cool!
MGX (MetaGPT X)
This score and size are amazing, I have a few questions
What is the difference between it and DeepSeek R1?
Do you have plans to make a larger size reasoning model?
Thank you for your contribution to the open source community!