Hello everyone,

I would like to get started with selfhost with two projects.

Project A (for me): A NUC with Proxmox installed on it, two VMs including a Home Assistant and a NAS system that I haven’t chosen yet.

The only question I have with this project is:

  • how to access the NAS and HA separately from the outside knowing that my access provider does not offer a static IP and that access to each VM must be differentiated from Proxmox.

Project B (for my uncle): A NUC (with Proxmox or not, I don’t know yet, perhaps simpler for making backups), with HA but especially Frigate. The goal is to use Google Coral to do recognition on 3 video surveillance cameras.

My questions are:

  • is Coral really useful with 3 cameras?
  • do you need a Coral in USB or M.2 version?
  • are there affordable NUCs with free M.2 slots?
  • won’t proxmox add a layer of complexity with Coral/Frigate/a Zigbee dongle?

Thank you in advance for your help and sorry if my post is long.

PS: if you have recommendations for cameras that work with Frigate and are self-powered with solar panels, I’ll take them!

Edit : 8 april 2024

A little update. Thank you everyone for your super quick responses!

Regarding my uncle’s project and after big discussions, he is going to buy Reolink cameras and that’s it. This will be much simpler for maintenance than building a server.

Regarding my project: I chose a Beelink Mini S12 pro with an N100 processor (for its low consumption) with a 2.5 bay for an SSD for my Nextcloud.

I wondered if I wouldn’t take the opportunity to add pihole and that’s where new questions arise…

I see a lot of people installing Pihole on Docker, should I put it on Docker? Or create a VM?

Should Docker be installed on Proxmox or on a VM?

Is Proxmox really useful, shouldn’t I better install HA/Nextcloud/Pihole under Docker directly?

Should I use LXC or Docker?

  • @dai@lemmy.world
    link
    fedilink
    English
    38 months ago

    Yeah there is a workaround for using bind-mounts in Proxmox VMs: https://gist.github.com/Drallas/7e4a6f6f36610eeb0bbb5d011c8ca0be

    If you wanted, and your drives are mounted to the Proxmox host (and not to a VM), try an LXC for the services you are running, if you require a VM then the above workaround would be recommended after backing up your data.

    I’ve got my drives mounted in a container as shown here:

    Basicboi config, but it’s quick and gets the job done.

    I’d originally gone down the same route as you had with VMs and shares, but it’s was all too much after a while.

    I’m almost rid of all my VMs, home assistant is currently the last package I’ve yet to migrate. Migrated my frigate to a docker container under nixos, tailscale exit node under nixos too while the vast majority of other packages are already in LXC.

    • LifeBandit666
      link
      fedilink
      English
      38 months ago

      This all sounds awesome. So eli5 I have all my drives mounted to Proxmox, then passed through to OMV in a VM.

      I can just mount these same drives to containers no issues right now, and I can add them to VMs using your link?

      I would like to get down to LXCs too, but I’ve found VMs so much easier to set up and use. I’ll try your way

      • @dai@lemmy.world
        link
        fedilink
        English
        2
        edit-2
        8 months ago

        I’ve not tested the method linked but yeah it would seem like it’s possible via this method.

        My lone VM doesn’t need a connection to those drives so I’ve not had a point to.

        You could probably run OMV in an LXC and skip the overheads of a VM entirely. LXC are containers, you can just edit the config files for the containers on the host Proxmox and pass drives right through.

        Your containers will need to be privileged, you can also clone a container and make it privileged if you have something setup already as unprivileged!