OptionProbability
uses some non-AdamW optimizer
Some image input (multimodality)
>= 50% on TerminalBench (https://www.tbench.ai/leaderboard)
>= 60% on BrowseComp (https://www.kaggle.com/benchmarks/openai/browsecomp)
1M+ Context
>=25T pretraining tokens
>= 73% on SWE-Bench Verified (according to epoch.ai)
>= 2 shared experts
Some variation of NSA (Native Sparse Attention)
>=16 active experts
Gemini 2.5 Pro tier or higher on FictionBench (90.6%+ at 192k)
>=512 experts
intra-expert communication
DS-MoE with adaptative expert count
>= 44% on Humanity's Last Exam (text only) at scale.com leaderboard
DeepSeek reports some results with a full-blown deep research agent, and emphasizes that this is the intended use-mode
>=52B active parameters
>=1.5T parameters
Releases before November
91
87
84
82
81
73
59
58
52
48
42
37
35
34
30
24
0
OptionVotes
NO
YES
1032
975
1281
950