PerkyPotato
Language Reasoning Models can overtake LLMs...
Here's my quick 3 minute breakdown:
- o1-preview: 97.8% on PlanBench Blocksworld vs. 62.5% for top LLMs, indicating shift from retrieval to reasoning.
- 52.8% on obfuscated "Mystery Blocksworld" vs. near-zero for LLMs, suggesting abstract reasoning skills, showing transfer capability.
- Variable "reasoning tokens" usage correlates with problem difficulty, hinting at internal search process, indicating adaptive compute.
2mo ago5.5K5.5K views
Find out if you are being paid fairly.Download Grapevine
SwirlyPretzel
Google2mo
Adaptive compute is very interesting to me imo. I wonder how they are using variable compute for each task and basis what meta heuristic
WobblyMarshmallow
Stealth2mo
Adaptive compute is what will help optimise cost for high complexity tasks, right?
SnoozyPickle
Probably different "cores" for different types of tasks
ZestyPenguin
Student2mo
Thanks for the paper! It's really interesting. I've been sounding like a madman explaining to people irl that Generative AI is not the end goal or even the natural next step of AI.
SnoozyPickle
What is the next step?
WobblyJellybean
Thanks for such a great post!!!! That's what I want more from this community.
Discover more
Curated from across