Alibaba logo

Qwen2.5 32B

AlibabaOpen WeightsPending Human Review

A highly efficient dense model designed to balance performance and memory requirements. It is open-sourced under Apache 2.0 license and supports long-context understanding. It outperforms many larger models (e.g., Llama-3.1-70B) in specific benchmarks while fitting on consumer-grade hardware.

2024-09-19
32B
Decoder-only Transformer
Apache 2.0

Specifications

Parameters
32B
Architecture
Decoder-only Transformer
License
Apache 2.0
Context Window
128,000 tokens
Max Output
8,192 tokens
Training Data Cutoff
Sep 2024
Type
text
Modalities
text

Benchmark Scores

Advanced Specifications

Model Family
Qwen
API Access
Available
Chat Interface
Available
Multilingual Support
Yes
Variants
Qwen2.5-32B-InstructQwen2.5-32B (Base)
Hardware Support
CUDAMetal

Capabilities & Limitations

Capabilities
reasoningcodingmathematicsmultilinguallong-context processing
Function Calling Support
Yes
Tool Use Support
Yes

Related Models