PulseAugur
LIVE 12:25:41
research · [1 source] ·
0
research

Qwen3-Next-80B-A3B-Base model targets ultimate training and inference efficiency

The Qwen3-Next-80B-A3B-Base model has been released, focusing on enhanced training and inference efficiency. This release aims to provide a more optimized solution for users working with large language models. Further details on its specific performance improvements and architectural innovations are expected. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

RANK_REASON Release of a new model from a non-frontier lab.

Read on Smol AINews →

COVERAGE [1]

  1. Smol AINews TIER_1 ·

    Qwen3-Next-80B-A3B-Base: Towards Ultimate Training & Inference Efficiency

    **MoE (Mixture of Experts) models** have become essential in frontier AI models, with **Qwen3-Next** pushing sparsity further by activating only **3.7% of parameters** (3B out of 80B) using a hybrid architecture combining **Gated DeltaNet** and **Gated Attention**. This new desig…