this post was submitted on 18 Feb 2024
22 points (100.0% liked)

Technology

37712 readers
539 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 2 years ago
MODERATORS
 

A 4 min 1080p30fps video taken with my phone camera is 518MB, While a 12 min 1080p30fps video ripped from youtube is 341MB, both are using mp4 h.264 as codec and the youtube one isnt of lower quality, so why this big difference?

you are viewing a single comment's thread
view the rest of the comments
[–] dept@lemmy.sdf.org 42 points 8 months ago* (last edited 8 months ago) (1 children)

Compression. YouTube videos usually have low bitrate too.

[–] sarmale@lemmy.zip 5 points 8 months ago (6 children)

Well yes, they have a lower bitrate, but why same quality as a bigger bitrate video? Same resolution, codec, frame rate...

[–] HeartyBeast@kbin.social 47 points 8 months ago* (last edited 8 months ago)

Because they aren’t the same quality? Lossy compression is lossy

[–] tonyn@lemmy.ml 28 points 8 months ago

Because you can compress video without reducing the resolution, codec, or frame rate. When a camera records two green pixels: it records (green pixel) (green pixel). When the video is compressed, it changes to (two green pixels) which takes up less storage space but retains the same information. Compression is computationally expensive, which is why cameras typically don't do it on the fly.

[–] Gormadt@lemmy.blahaj.zone 11 points 8 months ago

A lot of the data in the video file you take isn't that visible but it's there for when you put it in editing software

For example: if you took that lower sized YouTube rip and threw it into editing software to tinker with the brightness (as a simple example) you quickly start getting artifacts whereas the video you took would be able to handle the editing without the artifacts much better as there's more data to work with

Basically the YouTube video has had all the extra data scrapped off that it doesn't need because it's not going to be edited, it's only going to be viewed

[–] UmbraTemporis@lemmy.dbzer0.com 4 points 8 months ago

It' the individual frames that are compressed, essentially the video is unpacked and detail is culled from averages across multiple other frames beside it. So if the top of the video, for example the sky, doesn't change then that part will be kept static.

It's not so much properties about the video, but properties about each frame. I can take a 1080p image and blow it up to 8K in GIMP, but it's got the same detail as a 1080p image.

[–] Helix@feddit.de 4 points 8 months ago

If you do multiple passes you can alleviate some of the downsides of low bitrates. You can always easily spot it in dark areas. I despise watching space movies or shows on streaming services because of the resulting excessive banding artifacts.

[–] BarryZuckerkorn@beehaw.org 2 points 8 months ago

Video encoding has several tradeoffs:

  • Bitrate
  • Resolution/frame rate
  • Perceived quality
  • Computational complexity of encoding
  • Computational complexity of decoding

The cell phone encoding chips for video encoding on device make sacrifices to preserve speed of encoding and preserve battery life (higher computational complexity costs more processing cycles and tends to use more power). So it's simpler encoding, in exchange for inefficient bitrate compression.

YouTube (and all the social media sites) have huge server farms with highly specialized encoding chips for making the videos more efficient with bitrate for quality. That makes sense because videos designed to be watched millions of times could benefit from even a very slight improvement of bitrate in exchange for a one-time cost of complex encoding. It's also why YouTube tends not to convert to AV1 (very efficient in bitrate for quality, but computationally complex to encode) until a video has a few hundred views, because it's not clear whether that tradeoff is worth it until they know a lot of people will be watching it.

Netflix customizes even further for a per-video basis and looks for even more specialized tricks on a scene-by-scene basis, because every single one of its videos only needs to be encoded once for each quality/format but will be watched millions of times.

In other words, it's like any other engineering problem. The engineers choose different tradeoffs based on context, which means that the cell phone applies a different set of tradeoffs compared to the social media site's server farm.