fromInfoQ
3 days agoDeepSeek-V3.2 Outperforms GPT-5 on Reasoning Tasks
DeepSeek applied three new techniques in the development of DeepSeek-V3.2. First, they used a more efficient attention mechanism called DeepSeek Sparse Attention (DSA) that reduces the computational complexity of the model. They also scaled the reinforcement learning phase, which consumed more compute budget than did pre-training. Finally, they developed an agentic task synthesis pipeline to improve the models' tool use.
Artificial intelligence