this post was submitted on 02 Dec 2023
19 points (100.0% liked)
Technology
37720 readers
583 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I don't think we can let the current big AI players regulate themselves, but the ESRB hasn't been too bad at doing its job.
...it is now commonplace to find elements that are considered psychologically equivalent to gambling with real money in games rated E for everyone, therefore recommended for children of all ages.
ESRB may be plenty harsh on violence and sexual content, but it is completely neglecting their job where rating conditioning monetization elements accurately might earn the industry less money.
They were spawned to stop the government from regulating video game content thanks to games like Mortal Kombat and Night Trap.
Yes, I know. My point is that as new needs for self-regulation have come up, they are playing coy. Because as industry representatives it's more profitable to pretend they don't realize there is a new risk, that justifies ratings and warnings for children and their parents. If they will not catch up until the threat of government regulation comes up, they are not doing their job properly.
Ironically they are more harsh at fictional depictions of gambling than at lootboxes with real money, so they always knew there were some risks of this kind.
I get you.