• 0 Posts
  • 58 Comments
Joined 1 year ago
cake
Cake day: June 16th, 2023

help-circle



  • No I thinks is basically right although could be better worded maybe

    /sys is virtual file structure for kernel system info

    /proc is virtual file structure of kernel process info

    My understanding is /proc came first but was abused/free for all and started being used for all sorts of non standard/process kernel access. So /sys was created with stricter rules to make it more standardised.


  • Yea it is user friendly. If you’re using your computer once a week presumably its for things like web browsing or working with documents - these are very easy and straight forward to do in linux.

    The other big benefit is the cost - linux is free and you’ll save £120 on a basic version of Windows which can be used to get get a better PC or just saved.

    Add to that no advertising, much more private and entirely yours to do what you like with. And if you don’t like it you can easily install Windows instead, so its zero risk to try Linux.



  • What you’re asking is if you can run the existing Linux Mint on your drive within Windows running on the same drive?

    It may be feasible if VirtualBox or VMWare are able to access/mount the existing Linux partition as if its a virtual drive, and boot the OS but its likely to be difficult. The main issue is that windows does not easily mount Linux partitions. It is also an edge case use for most users so there will likely not be much guidance on how to achieve this easily.

    It might work more easily the opposite way round - boot Linux and mount Windows within Virtualbox but is not likely to be straightforward. Windows may be less flexible about being booted into a virtual machine with totally different hardware.

    All this may be overkill to the problems you’re trying to solve. You can mount the existing NTFS Windows drive within Linux Mint to access all your windows files without any virtualization. But I’m not clear what “settings” you’re looking at when you boot back in to windows? That seems to be the stumbling block. Is is specific software / tools you’re trying to migrate settings for?

    Another approach may be to launch windows, create a new linux Mint VM in Virtualbox, share a folder between Linux VM and Windows host, create whatever settings you’re trying to migrate in your VM Linux Mint, and when happy copy the home folder / settings folders into the shared folder. Then boot your PC into linux, mount the windows drive and pick up the settings files from the shared folder to migrate into your main Mint system. But whether that is even worth doing depends on what you’re trying to migrate.



  • I get what you’re trying to say but I disagree with this. Software can be a barrier to switching OS but it very much depends on the individual user’s needs - it’s not as easy as substituting open source for closed, and is only part of the difference anyway. For example, I use Outlook at work; Thunderbird is great but it is in no way a substitute for Outlook. Similarly, I use Microsoft Office 365 at work; OnlyOffice is in no way a substitute for an individual user (it can be for a whole business or for personal use, but not if you’re tied in to an organisation or employer using Office). If you’re tied into those platforms with work, then for occasional use you can just use the online versions of Microsoft Office in Linux via a web browser. And if you need to work from home or do more, then realistically you need to have Windows and access to the full suite installed locally.

    But software does not preclude switching to Linux; for example I dual boot between Windows and Linux on my home PC. I have an M.2 drive for Windows and another M.2 drive for Linux. I rarely use Windows at all now, but when I do it’s if for some reason I need to be doing work related stuff from home or rarely if I can’t get a game working in Linux. In Linux I can do all my web browsing, social media, video streaming, music listening, even gaming and I know I’m doing so privately and securely.

    I’d say the best way to switch to Linux is to switch to Linux. New users do not have to be “all in” - they can dual boot between Linux and Windows (or MacOS and Linux), and then have a low level of risk to try out the OS. It can even be beneficial in itself as they can compartmentalise work and free time by OS. And if they don’t want to dual boot, then just try it out by virtualisation.





  • Then I’d definitely set up a test system in a VM on your own PC (I.e. not the actual server machines). Even if you don’t want to use Docker, you can set up a complete version of your new server and practice deploying Jellyfin and Plex, and then test accessing it “remotely” to manage it. You can then decide whether switching away from Win11 is worth it.

    If you’re not familiar with the process of setting up a linux server then I’d actually suggest Debian instead of OpenSuSe. Looking at the Jellyfin guide for example it specifically covers the steps for installing directly onto a Debian host (while OpenSuSE set up means using the Fedora RPM guide). There are also straight forward guides for setting up a Debian server.

    Personally I’m not a fan of Ubuntu (because of Canonical and Snap etc) but there may also be a good choice just because there are so many guides out there for setting up Ubuntu server.


  • Docker is pretty easy to use, and is easy to play with either on your own system (linux or windows) or in VM guest system. The learning curve isn’t that high and Jellyfin for example has a clear set up guide for docker on their wiki.

    But radarr, sonarr etc can be installed directly within linux without docker. The Servarr wiki (that these projects use officially to share information as they’re so similar) has lots of straight forward guides for set up on Linux, Windows, Mac etc as well as Docker.

    I have a Linux guest VM set up with a Radarr, Sonarr etc set up, VPN and torrent set up. It was easy to do and means its network activity is all securely contained away from my host system. The tools let me set naming rules and file preferences. The library is a shared n folder in my host system, and that is included in my Jellyfin library. So all I have to do is subscribe to something i am interested in and it will just appear in my library once downloaded. The servarr tools are extremely convenient and worth looking at if you’re adding to that 30tb library over time.


  • OpenSuSE is a good distro with nice tools like Yast that have a decent CLI interface, and has server releases. The leap edition is stable but relatively up to date.

    But there are lots of viable alternatives, and if you’re going to use Docker then the host distro is probably not as important as you think.

    Simplest route may be to set up a demo server within a VM and see which one chimes the most with your style of use and maintenance. You could have a functioning demo server with docker and deploy both jellyfin and Plex in 20mins.


  • BananaTrifleViolin@lemmy.worldtoLinux@lemmy.mlAm I overthinking it?
    link
    fedilink
    English
    arrow-up
    19
    arrow-down
    3
    ·
    edit-2
    3 months ago

    Atomic systems or rpm-ostree is an interesting concept and may well be the future of distributing linux, but it has a lot of compromises. It may not be the first place to start when leaving windows.

    The problem is all the apps and things you may wish to do with your OS. Flatpak is the preferred method of installing apps as it doesn’t interfere with the OS, but that is a compromise that means more overhead for running apps including memory and disk space, and less integration with the host OS than traditional apps.

    You can overlay native apps but the more you overlay onto the immutable os, the more complex upgrading gets and the risks of breaking stuff.

    I’m not sure I would be starting with an immutable OS when switching away from windows. While it has a lot of theoretical benefits, its a work in progress and with significant compromises at the moment. Your VPN may just be the first of many programmes you find you need to overlay.

    I personally would look at a more traditional install, get it working how you like and if you find Linux works as a permanent home then think about how you might recreate that with an immutable OS base. If your needs a re very simple then maybe it’ll be easy, but if you’re using lots of software and tools (particularly if its not available Flatpak) or custom OS config you may find atomic desktops are not yet quite ready for you.

    It could be frustrating and off putting if you try linux immutable, find loads of problems and attribute that to linux when its actually the immutable OS that’s the cause.


  • Unless you’re specifically wanting to play with a different OS then Debian again. Makes much more sense to be using the same version of Linux and all the software ypu use rather than potentially different versions.

    Also it will be simpler to maintain as everything is the same.

    If you do want to play / test another distro then Mint has a low learning curve. FreeBSD is more different but you could easily try it and switch to something else if you don’t like it. Its different but not so much that linux users would feel totally lost.

    Probably the most confusing thing for linux user trying FreeBSD is that Bash is not installed, and BSD uses sh instead by default. Bash can be easily installed and set as the default shell which will give a lot more familiarity. But otherwise it’ll feel like a familiar modern complete system, and you can use the same desktop environments you’re familiar with already in linux.

    EDIT: You did say “backup” in your title. If that’s the main use case then definitely Debian again. If your laptop breaks or is stolen it makes sense to have a familiar system to pick up. Also important to sync and backup your data so it can be picked up on the other laptop. If backup machine is your focus then I’d say same OS and look more into data retention and retrieval between the two laptops, and ensure your important data is continuously backed up.


  • It kind of makes sense except the vast majority of software in all distros is not being packaged by the developers, its being packaged by volunteers in the relevant project. Most software is being used on trust that it is built off the original code and not interfered with.

    Its very difficult for any distros to actually audit all the code of the software they are distributing. I imagine most time is spent making sure the packages work and don’t conflict with each other.

    The verified tick is good in flatpaks but the “hide anything not verified” seems a little over the top to me. A warning is good but most software is used under trust in Linux - if you’re not building it yourself you don’t know you’re getting unadulterated software. And does this apply to all the shared libraries on flathub? Will thebwarn you if your software is using shared libraries that ate not verified?

    And while Flatpak is a potential vector to a lot of machines if abused, it is also a sandboxed environment unlike the vast majority of software that comes from distros own repos.

    Also given the nature of Flatpaks, any distros could host its own flatpaks but everyone seems to use flathub. If they’re not going to take on the responsibility of maintaining flathub and its software then their probably needs to be some way of “verifying” packages not coming directly from the developers. Otherwise users may lose put on the benefits of a shared distros agnostic library of software.

    I get why mint are doing this but i think its a bit of a false reassurance. Although from mints point of view they would be able to take direct responsibility for the software they distribute in their own repos (as much as you can in a warrentyless “use as your own risk” system)


  • You can keep windows and install Linux next to it.

    The best way would be to add a new ssd or m.2 card to your pc and install Linux on that. Make that the main boot device and Linux normally will detect Windows and give you a boot menu where you can chose between Linux and Windows each time you boot.

    Alternatively you can resize the windows partition and install Linux onto free space on your main drive. This is more fiddly and things can go wrong with this if you don’t know what you’re doing.

    You can also boot Linux on an external USB drive but this will be slower and may guge you a false impression of Linux. You can also try Linux in a virtual machine like Virtualbox but again this will be slower and will give you a false impression of Linux as a daily driver OS.

    I personally run a dual boot system - I have two m.2 nvme drives, one with windows and one with Linux. I barely use the windows partition now but I keep it around for rare work stuff or the rare occasion I have a game I can’t get to run in Linux. And I mean rare - booted Windows maybe 3 times in last 6 months.


  • BananaTrifleViolin@lemmy.worldtoLinux@lemmy.mlWhy does nobody maintain PPAs anymore?
    link
    fedilink
    English
    arrow-up
    73
    arrow-down
    5
    ·
    edit-2
    4 months ago

    PPAs are flawed and limited to the Debian/Ubuntu ecosystem. They’re a security issue as you really need to trust to the person or group who has set up the PPA (yet many people just added PPAs for all sorts of random software based on a Google search). They need to be maintained which is variable depending on the size of the project and for developers they’re only a route to support part of the entire Linux ecosystem. They can also conflict with the main system provided packages and repost which can break entire systems or break upgrades (happened to me on Mint, and I needed to do a complete system reinstall to remove legacy package conflicts).

    They’ve fallen out of fashion and rightly so.

    There are other ways to get software to users. Arch has its AUR which is basically a huge open repo. OpenSuSE has its OBS which is also a huge open repo. These are also not without their risks as it’s hard to curate everything on such an expansive repo. However others can take over packages if the original developer stops updating them, and you can see how the package was built rathe than just download binaries which allays some security concerns. They are also centralised and integrated into the system, while PPAs are a bit of a free for all.

    Flatpaks are a popular alternative now - essentially you download and run software which runs in a sandbox with its own dependencies. Flatpaks share their sandboxed dependencies but it does lead to some bloat as you’ll have system level libraries and separate Flatpak versions of the same libraries both installed and running at the same time. However it does mean software can be run on different systems without breaking the whole system if library dependencies don’t match. There are issues around signing though - flathub allows anyone to maintain software rather than insisting on the original devs doing so. That allows software to be in a Flatpak that might otherwise not happen but adds a potential security risk of bad actors packaging software or not keeping up to date. They do now have a verified tick in Flathub to show if a Flatpak is official.

    Snap is the Canonical alternative to Flatpak - it’s controversial as it’s proprietary and arguably more cumbersome. The backend is closed source and in canonical control. Snaps are also different and for more than just desktop apps and can be used to in servers and other software stacks, while Flatpak is focused only on desktop apps. Canonical arr also forcing Ubuntu users to use it - for example Firefox only comes in a snap on Ubuntu now. It has similar fundamental issues around bloat. It has mostly the same benefits and issues as Flatpak, although Flatpaks are faster to startup.

    Appimage are another alternative way to distribute software - they are basically an all-in-one image. You are essentially “mounting” the image and running the software inside. It includes all the libraries etc within the image and uses those instead of the local libraries. It does and can use local libraries too; the idea is to include specific libraries that are unlikely to be on most target systems. So again it has a bloat associated with it, and also security risks if the Appimage is running insecure older libraries. Appimage can be in a sandbox but doesn’t have to be, unlike Flatpak where sandboxing is mandatory - which is a security concern. Also Appimages are standalone and need to be manually updated individually while Flatpaks and Snaps are usually kept up to date via an update system.

    I used to use PPAs when I was still using Ubuntu and Mint. Now I personally use Flatpak, and rarely Appimages, and occasionally apps from the OBS as I’m on OpenSuSE Tumbleweed. I don’t bother with snaps at all - that’s not to say they don’t have value but it’s not for me.

    Edit: in terms of permissions, with Flatpak you can install Flatseal and manage software’s permissions and access per app. You can give software access to more locations including system level folders should you need to or all devices etc for example. I assume you can do the same with snap but I don’t know how.

    Also you can of course build software form source so it runs natively , if you can’t find it in a repo. I’ve done that a few times - can be fiddly but can also be easy.