How has no one posted this yet? Anyways, pretty cool stuff, scaling inference time through RL assisted reasoning. Big improvement on math, coding, reasoning type benchmarks, no improvement on writing quality etc. It's accessible to pro users now with 30msg cap on o1-preview, and 50msg cap on o1-mini. Note that o1-preview is not as good as the actual o1 model in their benchmarks, like significantly worse. Also, OpenAI seems to be encouraging people to NOT hype it on twitter, saying it's only better at some tasks, not overall, but is a promising path for the future.
Here are some funny things I noticed from various demos and the website:
Holy shit it has perfectly emulated the mind of the median voter ( the joke is that it thought for 7 seconds)
Liberalism wins once again - can't censor CoT reasoning steps or performance gets fricked & its a safety issue. But they're gonna hide it from us plebs for competitive advantage & wrongthink
In the one example they gave where they showed what it's actually doing during the reasoning stage, the model literally says "hmm." lmao. We're literally creating our children.
?? it's just like me, frfr
Jump in the discussion.
No email address required.
hmm, today I will read the mind of the ai while it is thinking
"HMM..."
Jump in the discussion.
No email address required.
More options
Context