You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
## Description
The script that generates this table has logic to sort newer/larger
models first. While it was working in some cases, it didn't correctly
sort Qwen3.5 before Qwen3 because the "." was sorting after " ". The
script was fixed.
## Testing
- [x] Local build succeeds without errors (`mint dev`)
- [x] Local link check succeeds without errors (`mint broken-links`)
- [x] PR tests succeed
Copy file name to clipboardExpand all lines: inference/models.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,11 +23,11 @@ W&B Inference provides access to several open-source foundation models. Each mod
23
23
| OpenAI GPT OSS 120B |`openai/gpt-oss-120b`| Text | 131k | 5.1B-117B (Active-Total) | Efficient Mixture-of-Experts model designed for high-reasoning, agentic and general-purpose use cases. |
24
24
| OpenAI GPT OSS 20B |`openai/gpt-oss-20b`| Text | 131k | 3.6B-20B (Active-Total) | Lower latency Mixture-of-Experts model trained on OpenAI's Harmony response format with reasoning capabilities. |
25
25
| OpenPipe Qwen3 14B Instruct |`OpenPipe/Qwen3-14B-Instruct`| Text | 32.8k | 14.8B (Total) | An efficient multilingual, dense, instruction-tuned model, optimized by OpenPipe for building agents with finetuning. |
26
+
| Qwen3.5 35B A3B |`Qwen/Qwen3.5-35B-A3B`| Text, Vision | 262k | 3B-35B (Active-Total) | Qwen3.5-35B-A3B is an open-weights multimodal MoE model built for efficient, high-throughput inference across chat, reasoning, and agentic tasks. |
26
27
| Qwen3 235B A22B Thinking-2507 |`Qwen/Qwen3-235B-A22B-Thinking-2507`| Text | 262k | 22B-235B (Active-Total) | High-performance Mixture-of-Experts model optimized for structured reasoning, math, and long-form generation. |
| Qwen3 30B A3B |`Qwen/Qwen3-30B-A3B-Instruct-2507`| Text | 262k | 3.3B-30.5B (Active-Total) | Qwen3-30B-A3B-Instruct-2507 is a 30.5B MoE instruction-tuned model with enhanced reasoning, coding, and long-context understanding. |
29
30
| Qwen3 Coder 480B A35B |`Qwen/Qwen3-Coder-480B-A35B-Instruct`| Text | 262k | 35B-480B (Active-Total) | Mixture-of-Experts model optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning. |
30
-
| Qwen3.5 35B A3B |`Qwen/Qwen3.5-35B-A3B`| Text, Vision | 262k | 3B-35B (Active-Total) | Qwen3.5-35B-A3B is an open-weights multimodal MoE model built for efficient, high-throughput inference across chat, reasoning, and agentic tasks. |
31
31
| Z.AI GLM 5 |`zai-org/GLM-5-FP8`| Text | 200k | 40B-744B (Active-Total) | Mixture-of-Experts model for long-horizon agentic tasks with strong performance on reasoning and coding. |
32
32
| Meta Llama 4 Scout (deprecated) |`meta-llama/Llama-4-Scout-17B-16E-Instruct`| Text, Vision | 64k | 17B-109B (Active-Total) | Multimodal model integrating text and image understanding, ideal for visual tasks and combined analysis. |
0 commit comments