this post was submitted on 04 Jun 2024
385 points (91.4% liked)
Technology
59188 readers
2502 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
You still need a massive fleet of these to train those multi-billion parameter models.
On the invocation side, if you have a cloud SaaS service like ChatGPT, hosted Anthropic, or AWS Bedrock, these could answer questions quickly. But they cost a lot to operate at scale. I have a feeling the bean-counters are going to slow down the crazy overspending.
We're heading into a world where edge computing is more cost and energy efficient to operate. It's also more privacy-friendly. I'm more enthused about a running these models on our phones and in-home devices. There, the race will be for TOPS vs power savings.