this post was submitted on 17 Feb 2024
371 points (97.4% liked)

Technology

59157 readers
2446 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

OpenAI collapses media reality with Sora AI video generator | If trusting video from anonymous sources on social media was a bad idea before, it's an even worse idea now::Hello, cultural singularity—soon, every video you see online could be completely fake.

top 50 comments
sorted by: hot top controversial new old
[–] gerryflap@feddit.nl 97 points 8 months ago (5 children)

I looked at these videos with very mixed emotions. On the one hand, I marveled at how far we've gotten. In a few years we went from generating sort of okay images in a very confined domain and essentially uncontrollable, to generating high resolution video that on first glance looks real.

But then the sadness struck me. I think we're entering the post-truth era, where the truth is harder and harder to find because all the fake stuff looks so real. We can generate text, images, sound, and now also video of whatever we want in the blink of an eye. Combine this with the tendency of people to accept any "information" that fits their view, and the filter bubbles that already exist, and we can see that humanity will start living in separate bubbles. Every bubble will have their own truth, and even if someone proves that a video or image is fake, that information will probably not even reach them because the truth doesn't generate enough clicks.

I want to stay optimistic, we've overcome so much stuff as a species, maybe we'll right the ship at some point. But with all the shit that is already going on in the world, the last thing we need is the ability to fake videos like this in no time at all. At some point the separate filter bubbles will tear our stable western world as we knew it apart, and we'll see shit like WW II again. The situation is already heating up.

[–] obinice@lemmy.world 32 points 8 months ago (1 children)

Damn, the AI that wrote this is really good!

[–] gerryflap@feddit.nl 12 points 8 months ago

Haha thanks

[–] shneancy@lemmy.world 30 points 8 months ago (1 children)

It's funny that in the human history there will be a gap of around 100 years where photos and video were considered to be solid proof and evidence that could determine the outcome of somebody's future

we're back at square one I guess

[–] General_Effort@lemmy.world 6 points 8 months ago

Naah, that was never a thing. EG: In 1917, two young girls created some photographs of fairies, the Cottingley Fairies. Arthur Conan Doyle, the inventor of Sherlock Holmes, endorsed them as real. When you have eliminated the impossible, whatever remains, however improbable, must be the truth. That quote is terrible advice.

The last 1 or 2 decades were really the golden age of credible evidence. Everyone has a video camera and can upload these videos almost immediately (proving that the videos were not edited later). Yet, at the same time, misinformation has become this huge topic.


We're not back to square 1, either. You can still immediately upload a video (or a hash, or get it certified in some way). Say, you do this with dashcam footage after a collision, ASAP. That makes it almost unassailable as evidence, because you can't have had time to forge it; certainly not in a way that is congruent with independent evidence and testimony.

If several people, upload videos of the same event at about the same time, then they either are all in it together and carefully prepared the videos beforehand, or not.

[–] Dadifer@lemmy.world 10 points 8 months ago

I'm actually glad that AI is making people realize that what they see is likely not real. For the history of media, the default has been for the written word or images or video to be taken as 100% truth, when in reality, it has always been very easy to deceive and manipulate. Now that we will suspect everything, maybe there will finally be critical thinking.

[–] eluvatar@programming.dev 9 points 8 months ago

Honestly I think we've been there for a while. The only difference now is that it's very easy for anyone to fake something, which might actually force us to face it? Or not who knows.

[–] madnness@lemmy.world 4 points 8 months ago

Sounds like what the board of openai thought when they attempted to fire him

[–] 1984@lemmy.today 55 points 8 months ago* (last edited 8 months ago) (2 children)

Another stepping stone to a much worse world. We won't know what is real anymore.

I think it's very cool technology, but in the hands of governments and psyops, it's going to brainwash entire countries.

Want another 9/11? Sure no problem. Blow up a building, tell people you have some random video of what happened, captured by civilians...place evidence in locations where it will be found.

[–] neurogenesis@lemmy.dbzer0.com 25 points 8 months ago (1 children)

We already don't know what is real. This will only make that clearer.

[–] 1984@lemmy.today 5 points 8 months ago (2 children)

I think some governments already had tech like this but not all.

It will be interesting to follow this. Probably lots of fake videos on YouTube as a consequence where events are not real but used to stir up aggression.

[–] neurogenesis@lemmy.dbzer0.com 4 points 8 months ago

Maybe, but I doubt it, only because traditional propaganda has been %100 effective without generative AI.

load more comments (1 replies)
[–] CosmoNova@lemmy.world 7 points 8 months ago (1 children)

We won’t know what is real anymore.

Of all the things, this really scares me. Many people scroll through their socials so quickly they will definitely not be able to tell apart generated clips from real ones. And the generated ones will only get better. One generation later, nobody will believe anything they see on a screen. And no, I don't think regulation can do much here as it will only end up in heavily censoring everything, leading to more distrust in media.

[–] archomrade@midwest.social 5 points 8 months ago

I think it could end up being a good thing if it causes social media to collapse into smaller, better known social groups.

[–] rayquetzalcoatl@lemmy.world 50 points 8 months ago (1 children)

This kind of AI stuff bums me out. You get people legitimately sharing AI images (and potentially videos in the future) and saying "look what I made!". It's totally inauthentic.

My boss loves this shit, on the other hand. Looking forward to the day she can automate our jobs away, I assume.

load more comments (1 replies)
[–] Nacktmull@lemmy.world 37 points 8 months ago (4 children)

Why are they working so hard on making humanity worse?

[–] BrianTheeBiscuiteer@lemmy.world 9 points 8 months ago

I really have to say it?

[–] mods_are_assholes@lemmy.world 3 points 8 months ago

Because the rich profit the most when everyone else is in fear and confusion.

So they generate it as much as possible and reap the rewards of ignorance and knee-jerk policies.

load more comments (2 replies)
[–] Dogyote@slrpnk.net 25 points 8 months ago (3 children)

It's like we're going back to the pre-internet era but it's obviously a little different. Before the internet, there were just a few major media providers on TV plus lots of local newspapers. I would say that, for the most part in the USA, the public trusted TV news sources even though their material interests weren't aligned (regular people vs big media corporations). It felt like there wasn't a reason not to trust them, since they always told an acceptable version of the truth and there wasn't an easy way to find a different narrative (no internet or crazy cable news). Local newspapers were usually very trusted, since they were often locally owned and part of the community.

The internet broke all of those business models. Local newspapers died because why do you need a paper when there are news websites? Major media companies were big enough to weather the storm and could buy up struggling competitors. They consolidated and one in particular started aggressively spinning the news to fit a narrative for ratings and political gain of the ownership class. Other companies followed suit.

This, paired with the thousands of available narratives online, weakened the credibility of the major media companies. Anyone could find the other side of the story or fact check whatever was on TV.

Now what is happening? The internet is being polluted with garbage and lies. It hasn't been good for some time now. Obviously anyone could type up bullshit, but for a minute photos were considered reliable proof (usually). Then photoshopping something became easier and easier, which made videos the new standard of reliable proof (in most cases).

But if anything can be fake now and difficult to identify as fake, then how can you fact check anything? Only those with the means will be able to produce undeniably real news with great difficulty, which I think will return power to major news companies or something equivalent.

I'm probably wrong about what the future holds, so what do you think is going to happen?

[–] kromem@lemmy.world 4 points 8 months ago* (last edited 8 months ago) (2 children)

I don't think you're wrong, I have been thinking the same thing.

Everyone has been worried about "AI misinformation" - but if misinformation becomes so commoditized online that someone convinced the moon landing is fake finds two dozen different AI generated sources agreeing with them but disagreeing with each other (i.e. a video of Orson Wells filming it but also a video of Stanley Kubrick filming it) we may well end up in a world where people just stop paying attention to the bullshit online that has been destroying people's minds for years now.

Couple this with the advances in AI correctly identifying misinformation and live fact checking it with citations to reputable and/or certified sources, combined with things like Elon Musk's 'uncensored' Grok turning around and calling his conservative Twitter fans racist and small minded morons while pointing out why they are wrong, or Gab's literal Adolf Hitler AI telling a user they were disgusting for asking if Jews were vermin - and we may just end up on a narrow path out of the mess we've found ourselves in well before AI was suddenly a thing.

I had been really worried about the AI misinformation angle, but given some recent developments in the past few months I'm actually hopeful about the future of a better informed public for the first time in years.

load more comments (2 replies)
load more comments (2 replies)
[–] SlopppyEngineer@lemmy.world 20 points 8 months ago (4 children)

Every day were getting closer to "The Running Man."

[–] snekerpimp@lemmy.world 6 points 8 months ago

More dancing women in tv???

[–] captsneeze@lemmy.one 5 points 8 months ago

FANTASTIC reference! This movie is so funny and awesome, and it seems to have completely disappeared from pop culture. I never understood why Conan looms so large in our collective memory, but this movie totally vanished.

[–] PipedLinkBot@feddit.rocks 3 points 8 months ago

Here is an alternative Piped link(s):

The Running Man

Piped is a privacy-respecting open-source alternative frontend to YouTube.

I'm open-source; check me out at GitHub.

load more comments (1 replies)
[–] OpenHammer6677@lemmy.world 18 points 8 months ago* (last edited 8 months ago) (4 children)

Genuine question: why do we need this type of thing?

Especially in view of the harm it can cause, what's the point of creating this aside from generating shareholder value?

Sure, creating a video out of text quickly is cool, but is there an actual need for this?

[–] YungOnions@sh.itjust.works 32 points 8 months ago (2 children)

I mean, the ability to generate whatever video you want without having to pay the costs normally associated with filming, location, actors etc is going to be very appealing to people like advertisers. This way you can have a few seconds of a beach for your travel company advert, for example, without having to pay for the stock footage or film it yourself. In fact I can see this transforming stock footage in general. Why bother to pay someone to make a generic video of 'people having a meeting' when an AI can do it for free in half the time. Doesn't even need to be that good if you're only using it briefly in a presentation. Not saying any of this is a good thing, but here we are..

[–] OpenHammer6677@lemmy.world 15 points 8 months ago (8 children)

I get what you're saying, but there isn't really any NEED for that

[–] Corkyskog@sh.itjust.works 9 points 8 months ago

Thats kind of like saying we never needed cars because we had horses.

[–] YungOnions@sh.itjust.works 8 points 8 months ago* (last edited 8 months ago)

I mean, yes in so far that it opens those options up to people who may not have been able to afford it before. Whether that's a 'need' or not depends on your opinion of the company I guess.

Also there are other applications beyond this, of course. Easily made videos could help reduce the costs associated with treating some mental health issues for example.

The you have the ability for novice film makers to make content, a bit like how engines like Unity have made it easier for people to make games. Sure there'll be shit-tier tat, but there'll also be content made by creators that may never have had the chance otherwise.

load more comments (6 replies)
[–] CosmoNova@lemmy.world 8 points 8 months ago (4 children)

This way you can have a few seconds of a beach for your travel company advert, for example, without having to pay for the stock footage or film it yourself.

Advertising holidays at places that do not exist! Exactly what we needed!

[–] I_Has_A_Hat@lemmy.world 7 points 8 months ago (1 children)

So no different than currently? Pictures and videos in ads all have heavy editing and post processing to make them look better.

load more comments (1 replies)
load more comments (3 replies)
[–] I_Has_A_Hat@lemmy.world 18 points 8 months ago (1 children)

Art. Presentations. Visualizations. Porn.

[–] ours@lemmy.world 6 points 8 months ago

Weird music videos. Just this week Youtube had pushed me music videos all done in the weird warpy AI style.

It's kind of cool and simultaneously already feels like a fad that tired with.

[–] BearOfaTime@lemm.ee 6 points 8 months ago

Who said anything about need?

It was created because someone thought of it. How it's used is a measure of the person using it.

People will find ways to utilize whatever someone creates. And usually in ways the creator never envisioned.

"Needing" something comes after a tool becomes ubiquitous. Imagine trying to screw in a Phillips screw with a slot screw driver - you'd need a Phillips driver because those screws are now ubiquitous (and I can't wait for them to go away. All hail our stripped screw saviour Torx!)

load more comments (1 replies)
[–] Mac@mander.xyz 15 points 8 months ago (1 children)

OpenAI singlehandedly breaking the internet. Props, tbh.

[–] Thorny_Insight@lemm.ee 12 points 8 months ago

They really are. I don't know about Google but with DDG when searching for information I feel like most of the top results are articles written by AI. Luckily it's still somewhat easy to recognize but that's not going to be the case for long. It's inevitable though so I don't really blame them. If not OpenAI then it would have just been someone else. I'm just worried about where this is going. I can think of more ways this could go wrong than right.

[–] RootBeerGuy@discuss.tchncs.de 7 points 8 months ago (3 children)

Isn't this a bit over dramatic, seeing as we had deepfake tech for a while now?

[–] M500@lemmy.ml 29 points 8 months ago

I don’t think so, as deep fake stuff was about switching faces and voices. You needed actual footage to train this on.

So if you wanted to stage something, it would take considerable effort, money, time, and manpower.

Now anyone will be able to just type in a prompt and have a video generated.

We saw the Joe Biden deepfake that made calls to tell people not to vote for him. Now just wait until we are having videos of him saying it sent out in mass.

[–] ch00f@lemmy.world 16 points 8 months ago (1 children)

Imagine generating 5,000 videos of different people (likenesses pulled from Facebook) reacting to a fake calamity staged in a certain city.

[–] CosmoNova@lemmy.world 8 points 8 months ago

Imagine seeing it every day to a point you can't see the real calamity coming because you stopped believing in them entirely.

load more comments (1 replies)
load more comments
view more: next ›