o1
The o1 model series is trained with large-scale reinforcement learning to reason using chain-of-thought. OpenAI o1 is the next model in this series (previously OpenAI o1-preview), with advanced reasoning capabilities that provide new avenues for improving safety and robustness. The model can reason about safety policies in context when responding to potentially unsafe prompts, through deliberative alignment.
Specifications
- Architecture
- Decoder-only Transformer
- License
- Proprietary
- Context Window
- 200,000 tokens
- Max Output
- 100,000 tokens
- Training Data Cutoff
- Sep 30, 2023
- Type
- multimodal
- Modalities
- textimage
Benchmark Scores
A benchmark for measuring browsing agents' ability to navigate the web and find hard-to-find, entang...
Advanced competitive programming benchmark for evaluating large language models on algorithmic probl...
Massive Multitask Language Understanding (MMLU) tests knowledge across 57 subjects including mathema...
A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Expert AGI with 11.5...
Graduate-level Problems in Quantitative Analysis (GPQA) evaluates advanced reasoning on graduate-lev...
A dataset of 12,500 challenging competition mathematics problems requiring multi-step reasoning....
Evaluates mathematical reasoning in visual contexts, combining vision and mathematical problem-solvi...
The FACTS Grounding Leaderboard evaluates LLMs' ability to generate factually accurate long-form res...
Advanced Specifications
- Model Family
- o-series
- API Access
- Available
- Chat Interface
- Available
- Multilingual Support
- Yes
Capabilities & Limitations
- Capabilities
- reasoningcodemathsciencechain-of-thoughtdeliberative alignment
- Known Limitations
- potential risks from heightened intelligence
- Notable Use Cases
- complex reasoningscientific problem solvingadvanced mathematicscomplex coding taskssafety-critical applications
- Function Calling Support
- No
- Tool Use Support
- No