In new benchmark tests, Alibaba’s Qwen3-Coder outperformed rival models from Moonshot AI and DeepSeek
Touted as the company’s “most advanced agentic AI coding model to date”, the Qwen3-Coder-480B-A35B-Instruct – built on a so-called Mixture of Experts (MoE) architecture – features a total of 480 billion parameters, 35 billion of which are active, and supports a 256,000-token context window, expandable to 1 million tokens through extrapolation methods.
Your personal data will be processed and information from your device (cookies, unique identifiers, and other device data) may be stored by, accessed by and shared with 88 TCF vendor(s) and 20 ad partner(s), or used specifically by this site or app.
Some vendors may process your personal data on the basis of legitimate interest, which you can object to by managing your options below. Look for a link at the bottom of this page to manage or withdraw consent in privacy and cookie settings.
Consent







