Latest Alibaba AI model demos AI improvements | Computer Weekly
Briefly

Alibaba Cloud has launched the QwQ-32B, a new open-source large language model with 32 billion parameters. Despite its smaller size, it shows performance similar to larger AI models such as DeepSeek-R1, which has 671 billion parameters. The performance benchmarks include various evaluation criteria for reasoning, coding, and tool utilization. Alibaba highlights the effectiveness of reinforcement learning in enhancing the model’s capabilities, offering advancements in mathematical reasoning and coding proficiency. The model's design enables critical thinking and tool use based on environmental feedback, indicating a significant leap in AI model training efficiency.
The QwQ-32B model by Alibaba Cloud showcases that a compact language model can rival larger models through innovative reinforcement learning techniques.
Alibaba stated that QwQ-32B achieves performance similar to DeepSeek-R1 despite utilizing significantly fewer parameters, illustrating the strengths of RL in model training.
Read at ComputerWeekly.com
[
|
]