PerkyPotato
PerkyPotato

Language Reasoning Models can overtake LLMs...

Here's my quick 3 minute breakdown:

  1. o1-preview: 97.8% on PlanBench Blocksworld vs. 62.5% for top LLMs, indicating shift from retrieval to reasoning.
  2. 52.8% on obfuscated "Mystery Blocksworld" vs. near-zero for LLMs, suggesting abstract reasoning skills, showing transfer capability.
  3. Variable "reasoning tokens" usage correlates with problem difficulty, hinting at internal search process, indicating adaptive compute.
2mo ago
5.5Kviews
Find out if you are being paid fairly.Download Grapevine
SwirlyPretzel
SwirlyPretzel

Adaptive compute is very interesting to me imo. I wonder how they are using variable compute for each task and basis what meta heuristic

WobblyMarshmallow
WobblyMarshmallow

Adaptive compute is what will help optimise cost for high complexity tasks, right?

SnoozyPickle
SnoozyPickle

Probably different "cores" for different types of tasks

ZestyPenguin
ZestyPenguin
Student2mo

Thanks for the paper! It's really interesting. I've been sounding like a madman explaining to people irl that Generative AI is not the end goal or even the natural next step of AI.

SnoozyPickle
SnoozyPickle

What is the next step?

WobblyJellybean
WobblyJellybean

Thanks for such a great post!!!! That's what I want more from this community.

Discover more
Curated from across