dataprolet

joined 1 year ago
MODERATOR OF
[–] dataprolet@lemmy.dbzer0.com 2 points 1 month ago

Deezloader and MusicHunter.

[–] dataprolet@lemmy.dbzer0.com 4 points 2 months ago (2 children)

Thanks, but I'm looking for a Desktop Linux client.

[–] dataprolet@lemmy.dbzer0.com 2 points 2 months ago (1 children)

Thanks, but the quality is still subpar and the video settings show 960p.

17
submitted 2 months ago* (last edited 2 months ago) by dataprolet@lemmy.dbzer0.com to c/degoogle@lemmy.ml
 

I'm looking for an alternative Youtube front-end for Linux Desktop but FreeTube for example doesn't support resolutions higher than 1080p. Is there an easy to use client, that supports 1440p and higher out-of-the-box?

[–] dataprolet@lemmy.dbzer0.com 38 points 2 months ago (94 children)

The devs have some problematic views, mainly transphobic and misogynistic.

[–] dataprolet@lemmy.dbzer0.com 2 points 3 months ago

Neo Launcher is pretty nice!

[–] dataprolet@lemmy.dbzer0.com 5 points 3 months ago (3 children)

Isn't this using a lot of computing power?

[–] dataprolet@lemmy.dbzer0.com 1 points 3 months ago

What's costs do you mean? It's free and open source.

[–] dataprolet@lemmy.dbzer0.com 2 points 3 months ago (1 children)

Immich has image and facial recognition by default and a very neat Android app. Also it's running in my home server, which has more power if Immich needs it. In that case I'd say software should serve one purpose and serve that good. Immich is just for picture management and does that very good. Nextcloud is a cloud and the Photos app is just a small extra that can't compete with a full-fledged software. Nextcloud runs fine on my Raspberry Pi 4, but it's only used by me and three friends. It's mainly limited by your network speed and disk speed I'd say. And I'm using an external hard drive without issues.

[–] dataprolet@lemmy.dbzer0.com 9 points 3 months ago (3 children)

Btrfs, but I'm curious about ZFS.

 

Does anybody know whether there is a WebGUI/Docker for Deezloader?

 

I'm looking for a simply solution to monitor all my servers and systems using a single dashboard. I want to see metrics like CPU usage, used RAM and storage to see if something is wrong.
I just set up Node-Exporter, Prometheus and Grafana but haven't found an existing dashboard that shows multiple hosts at once. Now I looked into Checkmk and Zabbix but I feel like both are a little overpowered for what I'm looking for. Do you have any recommendations?

 

I set up Headscale and Tailscale using Docker on a VPS, which I want to use as my public IPv4 and Reverse Proxy to route incoming traffic to my local network and e. g. my home server. I also set up Tailscale using Docker on my home server and connected both to my Headscale server.
I am able to ping on Tailscale container from the other and vice versa and set up --advertise-routes=192.168.178.0/24 on my home server as well as --accept-routes on my VPS, but I can't ping local IP addresses from my VPS. What am I missing?
Both container are connected to the host network, I have opened UDP ports 41641 and 3478 on my VPS.

 

I'm looking for an easy way to upload files from my Android smartphone to my home server. is there a - ideally dockerized - solution for that? Some simple web GUI where I can click on "Upload" and the files will be saved to a certain directory on my home server?

EDIT: I should've added that I want to do this remotely and not in my local network. I want to be able to send files from my Android smartphone from anywhere via the internet to my home server. That's why I thought about a services hosted on my server, which frontend I could access through my smartphone. But I might've answered my question already with the following: https://github.com/zer0tonin/Mikochi

EDIT #2: Thanks guys, I ended up creating my own Docker container running nextcloudcmd inspired by this: https://github.com/juanitomint/nextcloud-client-docker But I built the container from scratch and it's very minimalistic. I can publish it on my Gitlab when it's somewhat ready. Here's a little preview.

Dockerfile

FROM alpine:latest
RUN apk update && apk add nextcloud-client
COPY nc.sh .
RUN chmod +x ./nc.sh
VOLUME /data
CMD ./nc.sh

nc. sh (How can I prevent automatic hyperlinking?)

#!/bin/sh
while true
do
        nextcloudcmd /data https://${username}:${passwort}@${nextcloud-domain}
        sleep 300
done
 

I followed this tutorial to create local certificates for my home server, but now it failed to renew automatically and I have no clue waht to do. Can anybody assist me in debugging, please? https://notthebe.ee/blog/easy-ssl-in-homelab-dns01/

I'm using duckdns.org, added mydomain.duckdns.org and the local IP of my home server. In Nginx-Proxy-Manager I have created the respective wildcard certificate. The log of my NPM container reports the following:

[3/10/2024] [1:55:50 PM] [SSL      ] › ℹ  info      Renewing Let'sEncrypt certificates via DuckDNS for Cert #6: *.mydomain.duckdns.org, mydomain.duckdns.org
[3/10/2024] [1:55:50 PM] [SSL      ] › ℹ  info      Command: certbot renew --force-renewal --config "/etc/letsencrypt.ini" --work-dir "/tmp/letsencrypt-lib" --logs-dir "/tmp/letsencrypt-log" --cert-name "npm-6" --disable-hook-validation --no-random-sleep-on-renew 
[3/10/2024] [1:55:50 PM] [Global   ] › ⬤  debug     CMD: certbot renew --force-renewal --config "/etc/letsencrypt.ini" --work-dir "/tmp/letsencrypt-lib" --logs-dir "/tmp/letsencrypt-log" --cert-name "npm-6" --disable-hook-validation --no-random-sleep-on-renew 
[3/10/2024] [1:55:53 PM] [Express  ] › ⚠  warning   Saving debug log to /tmp/letsencrypt-log/letsencrypt.log
Failed to renew certificate npm-6 with error: The DNS response does not contain an answer to the question: mydomain.duckdns.org. IN TXT
All renewals failed. The following certificates could not be renewed:
  /etc/letsencrypt/live/npm-6/fullchain.pem (failure)
1 renew failure(s), 0 parse failure(s)
 

 

I noticed my home servers SSD running out of space and it ended up being my Jellyfin Docker container which wasn't clearing the directory for transcodes in /var/lib/jellyfin/transcodes correctly.

I simply created a new directory on my media hard drive and bind mounted the above mentioned directory to it. Now Jellyfin got over 1 TB of free space to theoretically clutter. To prevent this I simply created a cronjob to delete old files in case Jellyfin isn't.

@daily /usr/bin/find /path/to/transcodes -mtime +1 -delete

Easy!

 

I got a bunch of self-hosted stuff and use a VPS that has a public IPv4 to access my services because my home network has only DS-Lite. My home server ist connected to the VPS using Wireguard.
Now I want to connect my Smartphone to my VPN to be able to access some local services remotely. I'm able to add a second peer to the Wireguard config on the VPS, but I'm struggeling to configure the AllowedIPs correctly.
The VPS apparently needs AllowedIPs 10.0.0.0/24 and 192.168.178.0/24, but the Smartphone as well for both to redirect request into my home network. But it's not possible to configure the same IP ranges for two peers. What do I do?

EDIT: Solved: https://iliasa.eu/wireguard-how-to-access-a-peers-local-network/

 

I'm running Jellyfin in Docker in my home server for movies and shows. I recently added a music directory and apparently after that I'm getting almost hourly notifications from my Uptime-Kuma instance connected to Gotify that Jellyfin is down with status code 502. It's quickly up again, but I'm wondering what's causing this.
I have Nginx Proxy Manager configured for a local and a public domain pointing to my Jellyfin instance.
Any idea what could be causing this?

 

I have an Intel Core i5-7600K and just passed through my Intel HD 630 iGPU from my Proxmox host to a virtual machine running Debian to be able to use it in a Jellyfin Docker container. Everything worked fine, but I used only the basic configuration that I found which I don't really get. Can someone explain to me whether I'm missing something?

First I followed this tutorial: https://3os.org/infrastructure/proxmox/gpu-passthrough/igpu-passthrough-to-vm/
But I only added intel_iommu=on iommu=pt to my boot parameters and vfio, vfio_iommu_type1, vfio_pci, vfio_virqfd to /etc/modules.

But what are all the other parameters good for?
pcie_acs_override=downstream,multifunction initcall_blacklist=sysfb_init video=simplefb:off video=vesafb:off video=efifb:off video=vesa:off disable_vga=1 vfio_iommu_type1.allow_unsafe_interrupts=1 kvm.ignore_msrs=1 modprobe.blacklist=radeon,nouveau,nvidia,nvidiafb,nvidia-gpu,snd_hda_intel,snd_hda_codec_hdmi,i915"

Then I added the iGPU as a PCIe device to my VM using the Proxmox webUI and added the render device /dev/dri/renderD128 to the Jellyfin Docker container.
I followed the official instructions from Jellyfin: https://jellyfin.org/docs/general/administration/hardware-acceleration/intel/#configure-with-linux-virtualization

But I haven't added the host group ID, what is that good for?
And I also installed the intel-media-va-driver, i965-va-driver and firmware-linux-nonfree as well as firmware-misc-nonfree. Are all of those necessary?
And then I had to add options i915 enable_guc=2 to /etc/modprobe.d/i915.conf to get it to work. This is supposedly only necessary for Low-Power Encoding, but it was necessary to get hardware transcoding to work at all?

I'm happy that it is working now, but I don't really feel like I fully understood what I did. Were some steps unnecessary or did I miss anything?

view more: next ›