Is AGI just a frontier model that is continually running rather than triggered by human interaction?
Is AGI a model that can evolve itself and incorporate new knowledge?
I am not even convinced transformer ANNs are as smart as spiders, and I don't mean the cleverer jumping spiders. Here is my thought experiment: lets say you trained a transformer-powered robo-spider on one gazillion examples of spiderwebs in nature - between rocks, bushes, etc - and verified that in such natural environments you had "superspider" performance (whatever that means). Now test the robo-spider on a pantry, attic, garage, etc. Will the robo-spider be able to reliably spin a functional indoor web as well as a real spider? I doubt it.
I could be wrong of course! Maybe transformers can figure out the underlying geometric/physical principles in a reliable way. But zooming out a bit: despite the success of Be My Eyes / etc, I don't think any of us will live long enough to see an AI replacement for a seeing-eye dog. ("We are very sorry about your mother, there was an edge case where the AI didn't realize that green trucks were dangerous.")
[1] More people should seriously consider that LLMs are similar to Lisp expert systems with an easier user interface, but trading reliability for breadth and ease of development. I use Scheme all the time, clearly Lisp expert systems are useful, as are LLMs. But it is also clear that Lisp will never be a model for human/etc intelligence. See also Drew McDermott's classic paper, "Artificial Intelligence Meets Natural Stupidity" https://dl.acm.org/doi/10.1145/1045339.1045340
> a hypothetical artificial intelligence that matches or exceeds human intelligence — the intelligence of a machine that could successfully perform any intellectual task that a human being can.
Have we achieved this? I don't think so.
Rather, many people nowadays seem to imagine AGI as "a machine that can perform any intellectual task that most human beings can", which is actually a significantly lower bar - if the AI fails at something that a human can do, you must merely establish that some humans would also fail at it, and then the AI still qualifies as AGI by that definition.
From most of my network trying to make products based on LLMs, excluding cost, the biggest hurdles are hallucinations, and seemingly "non-sensical" reasoning or communicating. Subtle choices that just "feel" not quite right. Particularly when the LLM is being constrained for some activity.
Open-ended chat doesn't show these flaws as often.
Maybe you're right that it's self-awareness. The current models seem to have no metacognition, and even the "reasoning" hack isn't quite the same.