Qwen2.5 32B
A highly efficient dense model designed to balance performance and memory requirements. It is open-sourced under Apache 2.0 license and supports long-context understanding. It outperforms many larger models (e.g., Llama-3.1-70B) in specific benchmarks while fitting on consumer-grade hardware.
2024-09-19
32B
Decoder-only Transformer
Apache 2.0
Specifications
- Parameters
- 32B
- Architecture
- Decoder-only Transformer
- License
- Apache 2.0
- Context Window
- 128,000 tokens
- Max Output
- 8,192 tokens
- Training Data Cutoff
- Sep 2024
- Type
- text
- Modalities
- text
Benchmark Scores
Advanced Specifications
- Model Family
- Qwen
- API Access
- Available
- Chat Interface
- Available
- Multilingual Support
- Yes
- Variants
- Qwen2.5-32B-InstructQwen2.5-32B (Base)
- Hardware Support
- CUDAMetal
Capabilities & Limitations
- Capabilities
- reasoningcodingmathematicsmultilinguallong-context processing
- Function Calling Support
- Yes
- Tool Use Support
- Yes