Qwen/Qwen3-32B¶
Model Information¶
Qwen/Qwen3-32B
is a multilingual, instruction-tuned large language model developed by Alibaba Cloud. It is part of the Qwen3 series, designed to rival top-tier open and commercial LLMs in reasoning, multilingual QA, coding, and alignment. With 32 billion parameters, it balances strong performance with manageable deployment costs.
- Model Developer: Alibaba Cloud (Qwen Team)
- Model Release Date: May 2024
- Supported Languages: English, Chinese, French, Spanish, German, Japanese, Korean, Portuguese, and other major languages.
Model Architecture¶
Qwen/Qwen3-32B
is a dense, decoder-only transformer model designed for instruction-following and multilingual tasks.
Key Architecture Details:
- Model Type: Decoder-only transformer
- Parameters: 32B
- Context Length: Up to 128K tokens
-
Training Strategy:
- Pretraining on a diverse multilingual and code corpus
- Instruction fine-tuning using SFT and preference-aligned data
- Safety alignment and steering tuning for compliant outputs
- Tokenizer: Optimized multilingual tokenizer from Qwen3 series
- Capabilities:
- Strong few-shot and zero-shot generalization
- Robust code generation and math reasoning
- Steerable behavior through system prompts
Benchmark Scores¶
Category | Benchmark | Shots | Metric | Qwen3-32B |
---|---|---|---|---|
General | MMLU (CoT) | 0 | Acc. (avg) | 88.0 |
MMLU Pro (CoT) | 5 | Acc. (avg) | 60.1 | |
Steerability | IFEval | – | – | 92.1 |
Reasoning | GPQA Diamond (CoT) | 0 | Accuracy | 47.3 |
Code | HumanEval | 0 | Pass@1 | 85.4 |
MBPP EvalPlus (base) | 0 | Pass@1 | 85.9 | |
Math | MATH (CoT) | 0 | Sympy Score | 59.4 |
Tool Use | BFCL v2 | 0 | AST Macro Avg. | 80.3 |
Multilingual | MGSM | 0 | EM (exact match) | 77.6 |
Qwen3-32B competes closely with leading models like GPT-4o and Gemini Flash, offering robust multilingual reasoning, math, and tool use performance.