o1
The o1 model series is trained with large-scale reinforcement learning to reason using chain-of-thought. OpenAI o1 is the next model in this series (previously OpenAI o1-preview), with advanced reasoning capabilities that provide new avenues for improving safety and robustness. The model can reason about safety policies in context when responding to potentially unsafe prompts, through deliberative alignment.
Specifications
- Architecture
- Decoder-only Transformer
- License
- Proprietary
- Context Window
- 200,000 tokens
- Max Output
- 100,000 tokens
- Training Data Cutoff
- Sep 30, 2023
- Type
- multimodal
- Modalities
- textimage
Benchmark Scores
Evaluates models on competitive programming problems from the Codeforces platform....
American Invitational Mathematics Examination (AIME) 2024 problems....
Massive Multitask Language Understanding (MMLU) tests knowledge across 57 subjects including mathema...
Massive Multi-discipline Multimodal Understanding (MMMU) evaluates multimodal understanding across 3...
Graduate-level Problems in Quantitative Analysis (GPQA) evaluates advanced reasoning on graduate-lev...
A dataset of 12,500 challenging competition mathematics problems requiring multi-step reasoning....
Evaluates mathematical reasoning in visual contexts, combining vision and mathematical problem-solvi...
Advanced Specifications
- Model Family
- o-series
- API Access
- Available
- Chat Interface
- Available
- Multilingual Support
- Yes
Capabilities & Limitations
- Capabilities
- reasoningcodemathsciencechain-of-thoughtdeliberative alignment
- Known Limitations
- potential risks from heightened intelligence
- Notable Use Cases
- complex reasoningscientific problem solvingadvanced mathematicscomplex coding taskssafety-critical applications
- Function Calling Support
- No
- Tool Use Support
- No