bitemenow999

joined 1 year ago
[–] bitemenow999@alien.top 1 points 11 months ago (2 children)

And how is this related to this sub?

[–] bitemenow999@alien.top 1 points 11 months ago

Not sure if that is possible but even if it is possible it will be very inefficient given each model would have to "learn from scratch" like learning grammar, sentence positioning, etc. You could potentially make a central model that is pre-trained on a large corpus dataset and fine-tuned for a specific task but then it is just a standard GPT-like model, which when aggregated like a mixture of experts or ensemble can potentially do what you are saying.

Combining multiple models that are trained independently into one huge model is not possible because the model learns something different for each task and due to the inherently stochastic nature of general LLM (which is desirable to aggregate information), unless you are just looking to purely "retrieve information" what you say is not possible with the current standard training regime.

[–] bitemenow999@alien.top 1 points 11 months ago

GANs have faster inference.

[–] bitemenow999@alien.top 1 points 11 months ago

OpenAI is not a leader in AI, just because you know about chatgpt doesn't make them a leader... There are tons of research labs that are clearly at the forefront of ML

Also what is AGI? there is no clear definition yet, everyone has their own idea of AGI.

[–] bitemenow999@alien.top 1 points 11 months ago

Here is all you need to learn about relu:

relu = max(0,x)

[–] bitemenow999@alien.top 1 points 11 months ago (1 children)

I would rather just learn that from YouTube or free resources because the online MS degree is not worth the paper it will be printed on