
PerkyPotato
Language Reasoning Models can overtake LLMs...
Here's my quick 3 minute breakdown:
- o1-preview: 97.8% on PlanBench Blocksworld vs. 62.5% for top LLMs, indicating shift from retrieval to reasoning.
- 52.8% on obfuscated "Mystery Blocksworld" vs. near-zero for LLMs, suggesting abstract reasoning skills, showing transfer capability.
- Variable "reasoning tokens" usage correlates with problem difficulty, hinting at internal search process, indicating adaptive compute.
14mo ago
Jobs
One interview, 1000+ job opportunities
Take a 10-min AI interview to qualify for numerous real jobs auto-matched to your profile 🔑+322 new users this month

SwirlyPretzel
Google14mo
Adaptive compute is very interesting to me imo. I wonder how they are using variable compute for each task and basis what meta heuristic

WobblyMarshmallow
Stealth14mo
Adaptive compute is what will help optimise cost for high complexity tasks, right?

SnoozyPickle
Probably different "cores" for different types of tasks

ZestyPenguin
Student14mo
Thanks for the paper! It's really interesting. I've been sounding like a madman explaining to people irl that Generative AI is not the end goal or even the natural next step of AI.

SnoozyPickle
What is the next step?

WobblyJellybean
Goldman Sachs14mo
Thanks for such a great post!!!! That's what I want more from this community.
Discover more
Curated from across
