Hacker news

  • Top
  • New
  • Past
  • Ask
  • Show
  • Jobs

Measuring progress toward AGI: A cognitive framework (https://blog.google)

127 points by surprisetalk 1 day ago | 201 comments | View on ycombinator

pocketarc about 24 hours ago |

When people imagined AI/AGI, they imagined something that can reason like we can, except at the speed of a computer, which we always envisioned would lead to the singularity. In a short period of time, AI would be so far ahead of us and our existing ideas, that the world would become unrecognizable.

That's not what's happening here, and it's worth remembering: A caveman from 200K years ago would have been just as intelligent as any of us here today, despite not having language or technology, or any knowledge.

In Carolyn Porco's words: "These beings, with soaring imagination, eventually flung themselves and their machines into interplanetary space."

When you think of it that way, it should be obvious that LLMs are not AGI. And that's OK! They're a remarkable piece of technology anyway! It turns out that LLMs are actually good enough for a lot of use cases that would otherwise have required human intelligence.

And I echo ArekDymalski's sentiment that it's good to have benchmarks to structure the discussions around the "intelligence level" of LLMs. That _is_ useful, and the more progress we make, the better. But we're not on the way to AGI.

tyleo 1 day ago |

It still seems like something is missing from all these frameworks.

I feel like an average human wouldn't pass some of these metrics yet they are "generally intelligent". On the other hand they also wouldn't pass a lot of the expert questions that AI is good at.

We're measuring something, and I think optimizing it is useful, I'd even say it is "intelligent" in some ways, but it doesn't seem "intelligent" in the same way that humans are.

orangebread about 23 hours ago |

As an engineer who is also spiritual at the core, it seems obvious to me the missing piece: consciousness.

Hear me out.

I love AI and have been using it since ChatGPT 3.5. The obvious question when I first used it was "does this qualify as sentience?" The answer is less obvious. Over the next 3 years we saw EXPONENTIAL intelligence gains where intelligence has now become a commodity, yet we are still unable to determine what qualifies as "AGI".

My thoughts: As humans, we possess our own internal drive and our own perspective. Think of humans as distilled intelligence, we each have our own specialty and motivations. Einstein was a genius physicist but you wouldn't ask him for his expertise on medicine.

What people are describing as AGI is essentially a godlike human. What would make more sense is if the AGI spawned a "distilled" version with a focused agenda/motivation to behave autonomously. But even then, there are limitations. What is the solution? A trillion tokens of system prompt to act as the "soul"/consciousness of this AI agent?

This goes back to my original statement, what is missing is a level of consciousness. Unless this AGI can power itself and somehow the universe recognizes its complexity and existence and bestows it with consciousness I don't think this is phsyically attainable.

yellow_lead about 24 hours ago |

It's kind of funny that Google's idea of evaluating AGI is outsourcing the work to a Kaggle competition.

ArekDymalski 1 day ago |

It's good to have some kind of benchmark at least to structure the ongoing, fruitless discussion around "are we there already?".

However I must admit that including the last point that is partially hinting at the emotional or rather social intelligence surprised me. It makes this list go beyond usual understanding of AGI and moves it toward something like AGI-we-actually-want. But for that purpose this last point isn't ok narrow, too specific. And so is the whole list.

To be actually useful the AGI-we-actually-want benchmark should not only include positive indicators but also a list of unwanted behaviors to ensure this thing that used to be called alignment I guess.

lccerina about 20 hours ago |

Every week we are 50% closer to shifting the goalpost...

from the paper "AI systems already possess some capabilities not found in humans, such as LiDAR perception and native image generation". I don't know about them, but I can natively generate images in my mind.

andsoitis 1 day ago |

> Perception: extracting and processing sensory information from the environment

> Generation: producing outputs such as text, speech and actions

> Attention: focusing cognitive resources on what matters

> Learning: acquiring new knowledge through experience and instruction

> Memory: storing and retrieving information over time

> Reasoning: drawing valid conclusions through logical inference

> Metacognition: knowledge and monitoring of one's own cognitive processes

> Executive functions: planning, inhibition and cognitive flexibility

> Problem solving: finding effective solutions to domain-specific problems

> Social cognition: processing and interpreting social information and responding appropriately in social situations

--------------------

I prefer:

a) working memory (hold & manipulate information in mind simultaneously)

b) processing speed (how quickly & efficiently execute basic cognitive operations, leaving more resources for complex tasks)

c) fluid intelligence (ability to reason through novel problems without relying on prior knowledge)

d) crystallized intelligence (accumulated knowledge and ability to apply learned skills)

e) attentional control / executive function (focus, suppress irrelevant information, switch between tasks, inhibit impulsive responses)

f) long-term memory and retrieval (ability to form strong associations and retrieve them fluently)

g) spatial / visuospatial reasoning (mental rotation, visualization, navigating abstract spatial relationships)

h) pattern recognition & inductive reasoning (this is the most primitive and universal expression of intelligence across species, the ability to extract regularities from noisy data, to generalized from examples to rules)

mrkstu about 22 hours ago |

To me, a lot of what makes us sentient is our continuity. I even (briefly) remember my dreams when I wake up, and my dreams are influenced by my state of mind as I enter it.

LLMs 'turn on' when given a question and essentially 'die' immediately after answering a question.

What kind of work is going on with designing an LLM type AI that is continuously 'conscious' and giving it will? The 'claws' seem to be running all the time, but I assume they need rebooting occasionally to clear context.

ianrahman about 9 hours ago |

Altruism would make a good addition to the list. It’s clearly not universal, but most humans would help a fellow human in need. Or even (and in some cases more so) an animal need. Even if it didn’t directly benefit the actor.

There are other changes and additions which could be made to this list, but altruism may be the most important.

swagv1 about 5 hours ago |

AGI? As if all value that humans offer can be compressed and transmitted in binary form?

You'd have a more serious debate about antigravity.

wcgan7 1 day ago |

Cool that we are at a stage where it is meaningful to start measuring progress toward AGI. Something I am wondering on the philosophical side: are we ever going to be able to tell if the system really "understands" and "perceives" the world?

lvoudour 1 day ago |

Social cognition: processing and interpreting social information and responding appropriately in social situations

Is social cognition really a measure of intelligence for non-social entities?

qsort 1 day ago |

Those are crowdsourced benchmarks. We're calling them "cognitive" and "AGI" now, though. It's similar to when they made a benchmark and called it "GDP".

To be clear, I think we've seen very fast progress, certainly faster than I would have expected, I'm not trying to peddle some "wall" rhetoric here, but I struggle to see how this isn't just the SWE-bench du jour.

baggachipz about 23 hours ago |

This is a long way to say "let's crowdsource the shifting of our goalposts".

wewewedxfgdf about 24 hours ago |

AGI feels like a vanity project.

Who cares about AGI? Honestlky what's the gain.

Maybe Google could actually make Gemini good instead of being about 10 miles behind Claude instead of trying to make AGI because of - well some reason - cause they want to be famous.

Havoc about 23 hours ago |

Measuring something you can’t define or quantify seems somewhat dubious

gibsonf1 about 13 hours ago |

Its interesting that they don't even mention the key to human intelligence, concepts, in this list.

1970-01-01 about 23 hours ago |

Way too much framework. The A in AGI is for artificial. Have it build its own test harness instead of outsourcing it via hackathon. If you cannot trust that output, you're nowhere near AGI.

hbarka 1 day ago |

The two guys from Google get to set the rules?

How will they measure wisdom or common sense (ability to make an exception)?

https://youtu.be/lA-zdh_bQBo

righthand about 9 hours ago |

Google’s next cognitive framework will be for AGI Pro after we reach whatever productized, socially-accepted definition they cook up for AGI.

zug_zug about 24 hours ago |

I'm sorry what even is this? Giving $10k rewards for significant advancements toward "AGI"?

What does "making a framework" even mean, it feels like a nothing post.

When I think of what real AGI would be I think:

- Passes the turing test

- Writes a New York Times Bestseller without revealing it was written by AI

- Writes journal articles that pass peer review

- Wins a Nobel Prize

- Writes a successful comedy routine

- Creates a new invention

And no, nobody is going to make an automated kaggle benchmark to verify these. Which is fine, because an LLM will never be AGI. An LLM can't even learn mid-conversation.

boca_honey about 23 hours ago |

Friendly reminder:

Scaling LLMs will not lead to AGI.

cess11 about 23 hours ago |

The belief that there is no fundamental difference between mammals navigating fractal dimensions and imprisoned electrons humming in logic gates has to be considered a religious one.

wslh about 23 hours ago |

AGI may be a prerequisite for true superintelligence, but we're already seeing superhuman performance in narrow domains. We probably need a broader evaluation framework that captures both.

ottah about 22 hours ago |

Can we just focus on real problems, like stable and safe application of existing models? I'm just exhausted with the bullshit.

causalzap about 23 hours ago |

[dead]

drewcboston87 about 11 hours ago |

[dead]

nbnmbnmbnbm about 24 hours ago |

[dead]

AnaPaula15S98 about 12 hours ago |

[dead]

speefers about 24 hours ago |

[dead]

jamesvzb about 21 hours ago |

[flagged]

fnoef about 23 hours ago |

What is it with humans that we tend to speedrun into the extinction of our own race?