Maxy

joined 2 years ago
[–] Maxy@lemmy.blahaj.zone 2 points 10 months ago (1 children)

It depends what you’re optimising for. If you want a single (relatively small) download to be available on your HDD as fast as possible, then your current setup might be better (optimising for lower latency). However, if you want to be maxing out your internet speeds at all time and increase your HDD speeds by making the copy sequential (optimising for throughput), then the setup with the catch drive will be better. Keep in mind that a HDD’s sequential write performance is significantly higher than its random write performance, so copying a large file in one go will be faster than copying a whole bunch of random chunks in a random order (like torrents do). You can check the difference for yourself by doing a disk benchmark and comparing the sequential vs random writes of your drive.

[–] Maxy@lemmy.blahaj.zone 6 points 10 months ago (6 children)

qBittorrent has exactly the option you’re looking for, I believe it’s called “incomplete download path” in the settings, letting you store incomplete downloads at a temporary path and moving them to their regular location when the download finishes. Aside from the download speed improvement, this will also lead to less fragmentation on your HDD (which might be part of the reason why it is so slow when downloading directly to it). Pre-allocating space could have the same effect, but I would recommend only using one of these two solutions at once (pre-allocating space on your SSD would only waste space)

[–] Maxy@lemmy.blahaj.zone 6 points 11 months ago* (last edited 11 months ago)

It’s possible for a certain hardware/software setup not to support a certain codec. For example, my jellyfin client (Finamp) uses the iOS native decoders (afaik), which means opus files are practically broken. My music library (8000+ songs) contained exactly 1 lossy file, which just so happened to be an opus file. I decided to spend the extra ~20MB to standardise my entire library to flac files, ensuring I could play every song on all my devices.

Edit cause I posted too soon: you are generally correct; only in very specific circumstances will you encounter compatibility issues like this one in the modern world. This is 100% apple being apple, and you can expect pretty much every other (reasonably modern) device to support all codecs you might encounter in the wild.

[–] Maxy@lemmy.blahaj.zone 19 points 11 months ago (2 children)

To add to the audio compression: it isn’t possible to further compress an mp3 file without losing any quality. You can either:

  1. Recompress to a lossy codec (mp3, aac, opus). This will lead to smaller file sizes if you set the bitrate lower than that of the input file, but it will always worsen the quality, no matter the bitrate.
  2. Recompress to a lossless format (flac easily being the best one). Going from a lossy to a lossless format will increase the file size (sometimes by quite a substantial amount), while keeping the same quality. There is very little reason for you to do this
  3. keep the original files (my recommendation)

If you’re willing to spend some extra time learning about audio compression, you can download lossless files and compress those directly to whatever format and bitrate you want. The quality will be better than option 1 above, as the audio is only lossely compressed once instead of twice.

[–] Maxy@lemmy.blahaj.zone 9 points 11 months ago (3 children)

I have about 0 experience with openssl, I just looked at the man page (openssl-enc). It looks like this command doesn’t take a positional argument. I believe the etcBackup.key file isn’t being read, as that command simply doesn’t attempt to read any files without a flag like -in or -out. I could be wrong though, see previously stated inexperience.

[–] Maxy@lemmy.blahaj.zone 6 points 11 months ago

Dutch media are reporting the same thing: https://nos.nl/l/2529468 (liveblog) https://nos.nl/l/2529464 (Normal article)

[–] Maxy@lemmy.blahaj.zone 1 points 1 year ago

That seems like a good edit, and fair enough. Good to know that there is also room for people who want to use their computer in a non-fanatical way, simply minding our own business.

[–] Maxy@lemmy.blahaj.zone 2 points 1 year ago (2 children)

I don’t fit in an of these teams, and neither do literally all Linux users I know. Should we have identity crises, or could this be a giant oversimplification?

[–] Maxy@lemmy.blahaj.zone 5 points 1 year ago

Which compression level are you using? My old server is able to compress flac’s at the highest (and therefore “slowest”) compression level at >50x speed, so bumping the level up shouldn’t be too hard on your CPU.

[–] Maxy@lemmy.blahaj.zone 8 points 1 year ago

I’ve been running some external drives on my server for about a year now. In my experience, hard drives with an external power supply suffer less from random disconnects. The specific PC also makes quite a large difference in reliability. My server is just a regular desktop and has very little problem staying connected and powering my 3 external drives. My seedbox is an old laptop, and has been having almost constant problems with random disconnects and power issues. Maybe test how well your framework does with some external drives before committing to the plan?

[–] Maxy@lemmy.blahaj.zone 7 points 1 year ago (1 children)

To change the ownership of the files, you should only have to run sudo chown -R user:group directory. -R makes chown run recursively, so it will modify the directory and all subdirectories and files. Do note that changing the ownership to plex:plex or something similar would leave your user unable to normally modify the files. My solution to this was to add both my regular user and the plex (in my case jellyfin) user to the same group. That way both users can easily see and modify the files, as long as the group has read/write permissions (the 2nd column of rwx in ls -Al). If necessary, you can add group permissions with sudo chmod -R g+rw directory.

On a side note: have you considered using jellyfin? It’s a completely free alternative to plex, which recently received a truly massive update with tons of new features. Some people prefer plex’ overall experience, but I’ve been running jellyfin with almost no complaints.

Small disclaimer: I’m writing from mobile, so the commands might not be 100% correct. Run at your own risk, and NEVER POINT A CHMOD/CHOWN COMMAND AT SYSTEM DIRECTORIES LIKE / OR /USR. That’s one of the easiest ways to completely break your system.

[–] Maxy@lemmy.blahaj.zone 3 points 1 year ago

Have you tried the official guide from the jellyfin website?

As for the guide this AI generated: it bothers me that they instruct you to use chocolatey for the *arrs, but still advice you to install docker, qbittorrent and jellyfin manually (all of which have chocolatey packages). I disagree with the comment that external storage would be recommended, as internal storage is generally more reliable (depending on a lot of factors of course). Also, I believe the "adding a library"-section of the jellyfin setup is a bit too short to be of any use, and would recommend referring to the jellyfin docs instead.

This guide also doesn't explain how to make jellyfin accessible outside of your LAN. Once again, I'd recommend referring to the jellyfin docs if you want to do this.

I personally have only set up qbittorrent, jellyfin and docker (not the *arr suite), so I can't comment on the completeness of the guide, but I wouldn't trust it too much (seeing the previous oversights).

And finally, as someone who started their selfhosted server journey on windows: don't. There is a reason why almost all guides are written for linux, as it is (in my humble opinion) vastly superior for server usage once you get used to it.

view more: ‹ prev next ›