• 0 Posts
  • 19 Comments
Joined 4 months ago
cake
Cake day: June 29th, 2024

help-circle


  • Nice, I’ll have to watch this. A quick skim through the YT comments says that it’s AMD drivers which is the only thing I could think of. Linux Mint 21 actually has an “EDGE” iso which has a newer kernel version, and Linux Mint 22 is instead going to track the latest HWE kernels, so my understanding is this type of hardware problem should be a thing of the past at least in Linux Mint’s world. I don’t know if Ubuntu has their own plans or not.



  • “Escape hatch” specifically refers to the speculation that Valve is positioning themselves in a way that they can’t be forced into paying fees for existing on the Windows platform, and that if push comes to shove they can say they only support Linux now. This hasn’t happened yet, but it’s a strategic stance which will likely prevent it from even beginning to happen. This doesn’t have to do with the Steam Deck specifically; it was also part of their intentions with the Steam Machine and etc.


  • Maybe it needs to be more obvious that there are many ways to do things in Linux, and give new users a short “learning to learn” primer on how things operate differently in Linux-land, and where/how to look online for help. There are always first-boot popups but I imagine most people are conditioned to click out of them without even reading; forcing people to confirm a couple times that they want to skip “very helpful reading” may cut down on people that play the search engine lottery on what information they use for their first steps.

    Also semi-related, I hope that mainstream Linux eventually “un-stupids” computers for regular people again. I get the distinct feeling that Microsoft and Apple have, at least somewhat intentionally, imposed ‘learned helplessness’ onto average computer users. “Oh computers are magic no one knows how they work. We are the only wizards that could possibly understand them and we will sell you the solution.” Windows/OSX/iOS/etc are so locked down that people have rightfully learned over time that if they run into a problem, there really is no solution. I suspect that’s permeating into the new user experience on Linux where people will encounter one problem and throw their hands up and say “fucking computers” instead of using basic problem solving to try another approach.


  • Their rough new user experience is concerning though. From what they described I suspect many of their “problems” are not actually “real”, but it doesn’t really matter because they still ended up in a scenario where they thought there were problems. How did they end up thinking that everything must be done with terminal while using Ubuntu? I know in the last ~10 years there’s been a big focus on the new user experience, so what more can be done to prevent this? My gut says there are too many online resources that are confusing new users when they try to onboard themselves - especially resources that are old, written for other distros, or written for people who just want to find the command they can copy-paste to do something.


  • Gaming has been the only pathway to mainstream desktop since forever. I’ve been around for a hot minute and I remember that consistently, the “real Linux users” for years repeated “we don’t need gaming this is an adult OS go back to Windows and play with your toys” and then turned around and whined that no one wanted to use desktop Linux. Valve stepped in and casually created the year of the Linux desktop as a side-effect of just wanting an escape hatch for their business model. Now the casuals and elitists alike will have a better experience via the magic of Marketshare, and all it really took is not listening to people that don’t know what’s good for them.






  • There’s an unnerving lack of substance over on reddit. Recently I decided to look at reddit for the first time since last June, and every post’s comments had 1k people saying absolutely nothing worth reading. It feels like I’m reading AI-written posts that are in the uncanny valley of almost making sense and almost being on-topic. News articles have people that literally only talk about the exact words that were in the headline. Every single post’s top comments must be lame “jokes” or one-liners, and those must have several replies that riff off the joke in decreasingly-funny ways.

    I’ve picked up a strong habit of immediately looking in comment sections for good discussion and TL;DR’s on Lemmy posts and it took me a while to realize that I wasn’t actually reading anything in reddit comment sections. The words pass through my brain and nothing of value is absorbed, over and over. It feels like low-hanging fruit to say “reddit is all bots now” but there’s something seriously wrong about how it feels over there. You only really need ~10-20 top-level replies on a post to get a broad spectrum of answers, and Lemmy comment sections feel solid for the content that’s here. I wish there were more communities here (especially niche ones), but I’m grateful for what we have.




  • I recommend a dead man’s switch like Healthchecks.io, which can be selfhosted for free. Whenever you have something that’s regularly occurring, add an extra callout to your unique Healthchecks callout UUID as part of the automation, and Healthchecks will send you a notification if something misses its callout schedule. You can also attach whatever data (e.g. a log) to the callout so you can look back through the run history. IIRC Borg will give you a non-zero return code if it detects problems, so you can send e.g. https://hc-ping.com/your-uuid-here/$? and a non-zero code will signal a notification as well (more examples here).

    Also, Borgmatic is really easy to use for managing Borg repos. There’s a lot of configuration options (including Healthchecks.io integration) but you can delete like 90% of it for normal usecases.


  • For normal desktop users, yeah Debian Stable + Flatpaks is a winning combo for picking the software that you want to be cutting-edge and leaving the rest to rock-solid stability. Normally Linux distros keep a full ecosystem of packages that interop and depend on each other, but solutions like Flatpak have their own little microcosm of dependencies that can be used independently of the host distro. There are also Debian Backports for when you want native Debian packages that are more cutting-edge but still compiled to work with your older base system. Backports are not available for most packages but sometimes the important ones are available, like the Linux kernel itself. You can also try to compile your own backports, but you’ll be responsible for updating it.



  • I used Proxmox for a couple years and it’s good if you run a lot of VMs or LXCs, but I found that I’m not really the target audience. I ended up only running one Debian VM for my Docker containers. It was fine, but I eventually felt that Proxmox added no value for me, and the end result was sacrificing some memory and performance from using virtio emulations for CPU/GPU/RAM/filesystems. If your machines only have 8-16GB of RAM I don’t think it would be a good idea, as I’ve seen the rule of thumb is to dedicate 2GB for Proxmox’s usage, which is in addition to any guest OS’s requirements. Meanwhile I have a Debian install on a VPS that takes about 450MB of RAM.

    For me, pros:

    • Native ZFS support - invaluable, ZFS is terrific. MergerFS+SnapRAID is a decent replacement but the dodgy tooling and laundry list of footguns makes me nervous to use it on important data. ZFS is idiot-proof, as long as you know what you’re doing during the initial setup. RAIDZ expansion is coming this year and you can still use mixed-size disks in a RAIDZ as long as you accept that all disks are equivalent to the smallest one, so I personally feel ZFS is acceptable for grab-bag disk usage now
    • Separation of bare metal and server environment, which means you can spin up another server VM from scratch without impacting the previous one, then switch with zero downtime. In the end, I replaced Proxmox with Debian on ZFS root (ZFSBootMenu) and wrote a few hundred lines of bash to automate the installation, so when I switched it only took about 30 minutes of downtime start to finish.
    • Isolation of different environments. If my VM gets hacked, it will have a harder time reaching my Proxmox host etc. I run all services in isolated Docker environments anyway so this isn’t that big of a perk for my threat profile.

    Cons:

    • Partitioning RAM for ZFS ARC, Proxmox, and VM leads to inherent inefficiencies at the margins.
    • I usually give my VM n-1 CPU cores, which is still less power than if I had just used the CPU natively.
    • GPU passthroughs to VM can be less efficient, depending on the GPU and how it handles it. My iGPU is less performant when using its ~SR-IOV feature
    • Learning requirement - not a huge learning curve but it’s a lot of knowledge that I will not use now that I’ve stopped using Proxmox
    • Hosting your data pool on the Proxmox host or a dedicated data VM means that your server VM needs to use NFS to access its data, which lacks a handful of features (e.g. inotify) and is a pain
    • Need to maintain two systems for updates, downtimes, etc
    • More points of failure
    • Extra startup time
    • Run by a company that thinks it’s okay to use winrar-style nag popups every time you load the console, and requires you to manually dig through the source to disable that. I understand it’s their business model, it doesn’t change how it affects me the end user who lacks $120/year to spend on disabling a popup