I have been wanting to self-host recently I have an old laptop it’s a Toshiba satellite m100-221 sitting around it only has 4gb of ram, but I don’t know what is a good starting point for an OS for my home lab I discovered yunohost but heard mixed opinions about it when searching I would like lemmy’s opinion on a good OS for a beginner wanting to start a home lab I would prefer a simple solution like yunohost but would like it to be configurable it’s fine if it needs a bit of tinkering.

  • sugar_in_your_tea@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    2
    ·
    2 days ago

    That sounds overly complicated, why get VMs involved? Just install Debian or something and get things working.

    Proxmox is good if you know you want multiple VMs running for specialized needs. But multiple VMs isn’t happening on 4GB RAM.

    • downhomechunk@midwest.social
      link
      fedilink
      English
      arrow-up
      1
      ·
      41 minutes ago

      Your approach works too. Something like CasaOS answers OP’s question directly. I was thinking about how I started on this journey. I wanted to play with enterprise level tools at home on repurposed e-waste. So I started with proxmox. But I also came to the table with a couple decades of Linux experience under my belt.

      Those scripts make it so easy. You can paste a command, accept defaults, watch some text scroll by and finish with instructions on how to access the tool you just installed.

      My homelab is low power as well. I’m currently running zero VMs. Everything is done with LXCs. You can run a pi hole on 512 MB RAM.

    • Windex007@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 days ago

      Easily can have multiple LXCs, and being able to take snapshots for backup is probably a nice thing to have if you’re just learning.

      And if they get more hardware, moving VMs to other clustered proxmox instances is a snap.

      • sugar_in_your_tea@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        2 days ago

        If you just want LXCs, use Docker or Podman on whatever Linux distro you’re familiar with. If you get extra hardware, it’s not hard to have one be the trunk and reverse proxy to the other nodes (it’s like 5 lines of config in Caddy or HAProxy).

        If you end up wanting what Proxmox offers, it’s pretty easy to switch, but I really don’t think most people need it unless they’re going to run server grade hardware (i.e. will run multiple VMs). If you’re just running a few services, it’s overkill.

            • drkt@scribe.disroot.org
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              2 days ago

              They use some of the same kernel functions but they are not the same. They are not comparable. LXCs are used to host a whole separate system that shares kernel with its host, docker is used to bundle external requirements and configs for a piece of software for ease of downstream setup. Docker is portable, LXCs much less so.

              • sugar_in_your_tea@sh.itjust.works
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                2
                ·
                2 days ago

                Sure, Docker is more or less an abstraction layer on top of LXC. It’s the same tech underneath, just a different way of interacting with it.

        • Windex007@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          If you’re just running a few services, and will only ever be running a few services, I agree with you.

          The additional burden of starting with proxmox (which is really just debian) is minimal and sets you up for the inevitable deluge of additional services you’ll end up wanting to run in a way that’s extensible and trivially snapshotable.

          I was pretty bullish on “I don’t need a hypervisor” for a long time. I regret not jumping all-in on hypervisors earlier, regardless of the services I plan to run. Is the physical MACHINEs purpose to run services and be headless? Hypervisor. That is my conclusion as for what is the least work overall. I am very lazy.

          • sugar_in_your_tea@sh.itjust.works
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 days ago

            For snapshots, you can use filesystem features, like BTRFS or ZFS snapshots. If you make sure to encapsulate everything in the container, disaster recovery is as simple as putting configs onto the new system and starting services (use specific versions to keep things reasonable.

            I think that’s also really lazy, it’s just a different type of lazy from virtualization.

            My main issue with virtualization is maintenance. Most likely, you’re using system dependencies, and if you upgrade the system, there’s a very real chance of breakage. If you use containers, you can typically upgrade the host without breaking the containers, and you can upgrade containers without touching the host. So upgrades become a lot less scary since I have pretty fine-grained control and can limit breakage to only the part I’m touching, and I get all of that with minimal resource overhead (with VMs, each VM needs the whole host base system, containers don’t).

            Obviously use what works for you, I just think it’s a bit overwhelming for a new user to jump to Proxmox vs a general purpose Linux distro.