o1

o1

o1 achieves 86% accuracy on Mathematics Olympiad benchmarks (vs. GPT-4o's 13%), offers PhD-level STEM proficiency, and maintains a 0.17% deceptive response rate in synthetic testing.

ConversationWeb DesignCode GenerationReasoning
Provider
Openai
Release Date
December 5, 2024
Size
MEDIUM
Parameters
175B

Benchmark Performance

Performance metrics on industry standard AI benchmarks that measure capabilities across reasoning, knowledge, and specialized tasks.

MMLU

91.8%
Source:github.com

GPQA

75.7%
Source:github.com

MATH

96.4%
Source:github.com

MGSM

89.3%
Source:github.com

Model Insights

All Model Responses

Related Models