Simon Willison’s Weblog

Subscribe
Atom feed for inference-scaling

32 posts tagged “inference-scaling”

Improving performance of LLMs through inference - spending more tokens "thinking" about a problem, as seen in OpenAI's o1, DeepSeek's R1 and Qwen's QwQ.

2024

o1-mini is the most surprising research result I've seen in the past year

Obviously I cannot spill the secret, but a small model getting >60% on AIME math competition is so good that it's hard to believe

Jason Wei, OpenAI

# 12th September 2024, 11:45 pm / o1, generative-ai, openai, ai, llms, inference-scaling

Notes on OpenAI’s new o1 chain-of-thought models

OpenAI released two major new preview models today: o1-preview and o1-mini (that mini one is not a preview)—previously rumored as having the codename “strawberry”. There’s a lot to understand about these models—they’re not as simple as the next step up from GPT-4o, instead introducing some major trade-offs in terms of cost and performance in exchange for improved “reasoning” capabilities.

[... 1,568 words]