this post was submitted on 25 Nov 2023
1 points (100.0% liked)

Machine Learning

1 readers
1 users here now

Community Rules:

founded 11 months ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[โ€“] jugalator@alien.top 1 points 10 months ago (4 children)

Research papers have also observed diminishing returns issues as models grow.

Hell maybe even GPT-4 was hit by this and that's why GPT-4 is not a single giant language model but running a mixture of experts design of eight 220B models trained for subtasks.

But I think even this architecture will run into issues and that it's more like a crutch. I mean, you'll eventually grow each of these subtask models too large and might need to split them as well, but this might mean you run into too small/niche fields per respective model and that sounds like the end of that road to me.

[โ€“] AdoptedImmortal@alien.top 1 points 10 months ago

I mean, that is literally how any form of AGI will work. No one in the field has ever thought one model will be capable of reaching AGI. All these models are highly specialized for the task in which they are trained. Any move towards an AGI will be getting many of these highly specialized AI's to work in conjunction with one another. Much like how our own brains work.

load more comments (3 replies)