If I may ask: how practical is monitoring / administering rootless quadlets? I’m running rootless podman containers via systemd for home use, but splitting the single rootless user into multiple has proven to be quite the pain.
If I may ask: how practical is monitoring / administering rootless quadlets? I’m running rootless podman containers via systemd for home use, but splitting the single rootless user into multiple has proven to be quite the pain.
That’s just the limited resolution if you push it to the limit of how fast it can refresh the picture…
I remember reading somewhere that it would be the farthest manmade object from earth, far outpacing the Voyager spacecrafts, assuming it didn’t vaporize.
With bluray rips, I don’t really see any way to avoid that unfortunately, unless someone else has already added the hashes for your release. Most people use it to scan their encoded releases, which will (in most cases) have already been added to AniDB by the release group. I’m a bit surprised though, that none of your rips are recognized. Have you checked the AniDB pages for your series to see if anyone uploaded hashes for bluray rips?
Grouping seasons into a series folder doesn’t work well in some cases, because that’s not the way they are released in Japan. A new season is (most of the time) effectively an entire new show entry. Show seasons are mostly a north american thing. No matter which software you use, there’s always going to be some minor issues if you group seasons into one entry.
Shoko compares a files ED2K hash against the AniDB database. The filename doesn’t matter for automatic detection. Have a look at the log to see if there are any issues. It’s entirely possible that AniDB just doesn’t have the hashes for the raw BluRay rip. In that case you can either manually link them in Shoko, connecting the AniDB episode id to the file hash, or create new file entries on AniDB with your specific hashes.
Shoko also has rate limits. The problem is that AniDB does rate limiting in an extremely stupid way for a UDP API and doesn’t even have the decency to define clear time limits.
Being free from mind control means seeing idiots wearing tinfoil hats?
Pretty sure that the registry path for official images is “library” (at least it used to be). So it should be “docker.io/library/debian”, though I can’t double check at the moment.
He’s gonna do it instead.
Consider what trace minerals actually means, then think about the tiny amounts of salt used for cooking.
I’ve been hating on the raw food movement for much longer than lemmy even existed. Drinking raw milk is especially stupid.
You mean hiding their public IP? I guess that’s a feature.
That’s what a firewall and a DNS service is for respectively, imho. As long as you get an IPv6 prefix from your ISP, you can expose as many devices or services to the public as you want, by just allowing incoming traffic to a listening port. That was sort of the whole point of having a large enough address space when moving away from v4. Maybe it’s just me but reading stuff about “private AI” on a website where the relation to the product is not immediately obvious, makes me question their legitimacy.
The more I look at their site, the more it reads like a sales pitch for IPv6, which sounds kind of expensive at $6-10 a month.
What problem does this solve? Do ISPs not provide IPv6 prefixes anymore?
That script is a wrapper around a single call to qrencode. I’ve been making qr codes from wireguard config files in the terminal at least since PiVPN existed. There are plenty of guides on how to do this as well.
None of what I said implied anything else.
I’m aware that any of the past attempts to classify body types are extremely pseudo-scientific and I’ve explained as much in a different comment in this thread. The point is that “body type” isn’t just necessarily just a generic way to refer to someone’s body shape. Plenty of people still believe in that made up nonsense.
XML aims to be both human-readable and machine-readable, but manages neither. It’s only really worth it if you actually need the complexity or extensibility, otherwise it’s just a major pain to map XML structures to any sensible type representation. I’ve been forced to work with some of the protocols that people like to present as examples of good XML usage and I hate every single one of them.
Fuck YAML though. That spec is longer and more complex than any other markup language I know of and it doesn’t have a single fully compliant implementation.