this post was submitted on 22 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

Today we run a query, the LLM answers. The evolution of this was going multi-modal, and being able to input or output other forms of data, like video and sound

But still, we query a model, the LLM answers. Same with agents, only then we have multiple queries working in a system, often assisted by function calling and automation in order to put thought into action

We do it this way because it's the most obvious solution. The easiest one to implement, or rather the only one we can implement at this time

At this time

In the future we will get more efficient models. Models that can output in the orders of 100+ t/s. Models that can run so fast that they're essentially able to simulate continuous runtime

And that's not all. With a model this fast you could also implement multi-query systems which runs backend to handle any agent related tasks or even for assisting in reasoning by having lightning fast discussions with the frontend before giving any outputs

tl;dr LLMs are just waiting for us to give them more juice so that they're fast enough to simulate continuous runtime

you are viewing a single comment's thread
view the rest of the comments
[–] autisticit@alien.top 1 points 11 months ago

Maybe nobody is talking about this because it is really obvious? You literally say it in your post title.