this post was submitted on 09 Nov 2023
1 points (100.0% liked)

Machine Learning

1 readers
1 users here now

Community Rules:

founded 10 months ago
MODERATORS
 

Paper: https://arxiv.org/abs/2311.02462

Abstract:

We propose a framework for classifying the capabilities and behavior of Artificial General Intelligence (AGI) models and their precursors. This framework introduces levels of AGI performance, generality, and autonomy. It is our hope that this framework will be useful in an analogous way to the levels of autonomous driving, by providing a common language to compare models, assess risks, and measure progress along the path to AGI. To develop our framework, we analyze existing definitions of AGI, and distill six principles that a useful ontology for AGI should satisfy. These principles include focusing on capabilities rather than mechanisms; separately evaluating generality and performance; and defining stages along the path toward AGI, rather than focusing on the endpoint. With these principles in mind, we propose 'Levels of AGI' based on depth (performance) and breadth (generality) of capabilities, and reflect on how current systems fit into this ontology. We discuss the challenging requirements for future benchmarks that quantify the behavior and capabilities of AGI models against these levels. Finally, we discuss how these levels of AGI interact with deployment considerations such as autonomy and risk, and emphasize the importance of carefully selecting Human-AI Interaction paradigms for responsible and safe deployment of highly capable AI systems.

https://preview.redd.it/64biopsh79zb1.png?width=797&format=png&auto=webp&s=9af1c5085938dac000aaf23aa1b306133b01edb4

you are viewing a single comment's thread
view the rest of the comments
[–] Difficult_Ticket1427@alien.top 1 points 10 months ago (12 children)

I doubt that any model currently is in the “emerging AGI” category (even by there own metric of “general ability and metacognitive abilities like learning new skills”).

The model(s) we currently have are fundamentally unable to update their own weights so they do not “learn new skills”. Also I don’t like how they use “wide range of tasks” as a metric. Yes, LLMs outperform many humans at things like standardized tests, but I have yet to see an LLM who can constantly play tiktaktoe at the level of a 5 year old without a paragraph of “promt engineering”

I’m not the most educated on this topic (still just a student studying machine learning) but imo I think that many researchers are overestimating the abilities of LLMs

[–] ThisIsBartRick@alien.top 1 points 10 months ago

To be fair, 5 year Olds don't have the innate ability to know how tic tac toe works. Someone had to teach them. We just chose not to teach that to llms.

load more comments (11 replies)