Human-Level AGI Definition:
 A system has reached human-level AGI when it becomes impossible for humans to create any benchmark where humans reliably

Search code, repositories, users, issues, pull requests...

submited by
Style Pass
2024-12-23 00:00:04

Human-Level AGI Definition: A system has reached human-level AGI when it becomes impossible for humans to create any benchmark where humans reliably outperform the system.

Like many others, I've been amazed by the recent capabilities demonstrated by systems like O1 and O3. As someone deeply interested in technology and AI (View My Projects), I've spent recent years trying to create tests that AI cannot pass. This endeavor has become increasingly challenging - and that's actually really interesting when you think about it.

This challenge of creating tests is intimately connected to how we define human-level AGI. It's as if we've been trying to define it through our attempts to create benchmarks that AI can't solve.

Today, I launched a website called "h-matched" that tracks major AI benchmarks and how long it took for AI systems to reach human-level performance on each one. If you look at the data, you'll notice something fascinating - we're approaching a point where it's becoming incredibly difficult to create any test where humans can outperform the best AI systems.

This got me thinking about what happens when we extrapolate the trend line to where it hits zero (which looks to be around 2025). What would that actually mean in practical terms? From my perspective, it would indicate we've reached a point where we literally cannot create any type of task where humans perform better than AI systems.

Leave a Comment