Skip to content

Qwen/Qwen3-32B

Model Information

Qwen/Qwen3-32B is a multilingual, instruction-tuned large language model developed by Alibaba Cloud. It is part of the Qwen3 series, designed to rival top-tier open and commercial LLMs in reasoning, multilingual QA, coding, and alignment. With 32 billion parameters, it balances strong performance with manageable deployment costs.

  • Model Developer: Alibaba Cloud (Qwen Team)
  • Model Release Date: May 2024
  • Supported Languages: English, Chinese, French, Spanish, German, Japanese, Korean, Portuguese, and other major languages.

Model Architecture

Qwen/Qwen3-32B is a dense, decoder-only transformer model designed for instruction-following and multilingual tasks.

Key Architecture Details:

  • Model Type: Decoder-only transformer
  • Parameters: 32B
  • Context Length: Up to 128K tokens
  • Training Strategy:

    • Pretraining on a diverse multilingual and code corpus
    • Instruction fine-tuning using SFT and preference-aligned data
    • Safety alignment and steering tuning for compliant outputs
    • Tokenizer: Optimized multilingual tokenizer from Qwen3 series
    • Capabilities:
    • Strong few-shot and zero-shot generalization
    • Robust code generation and math reasoning
    • Steerable behavior through system prompts

Benchmark Scores

Category Benchmark Shots Metric Qwen3-32B
General MMLU (CoT) 0 Acc. (avg) 88.0
MMLU Pro (CoT) 5 Acc. (avg) 60.1
Steerability IFEval 92.1
Reasoning GPQA Diamond (CoT) 0 Accuracy 47.3
Code HumanEval 0 Pass@1 85.4
MBPP EvalPlus (base) 0 Pass@1 85.9
Math MATH (CoT) 0 Sympy Score 59.4
Tool Use BFCL v2 0 AST Macro Avg. 80.3
Multilingual MGSM 0 EM (exact match) 77.6

Qwen3-32B competes closely with leading models like GPT-4o and Gemini Flash, offering robust multilingual reasoning, math, and tool use performance.


References