• 3 Posts
  • 45 Comments
Joined 1 year ago
cake
Cake day: July 3rd, 2023

help-circle
    1. Sure but there’s no reason to openly advertise that yours has open services behind it.
    2. Absolutely. There are countries that I’m never going to travel there so why would I need to allow access to my stuff from there? If you think it’s nonsense then don’t use it, but you do you and I’ll do me.
    3. See point 3.

    We all need to decide for ourselves what we’re comfortable with and what we’re not and then implement appropriate measures to suit. I’m not sure why you’re arguing with me over how I setup my own services for my own use.


  • Yes and no? It’s not quite as black and white as that though. Yes, they can technically decrypt anything that’s been encrypted with a cert that they’ve issued. But they can’t see through any additional encryption layers applied to that traffic (eg. encrypted password vault blobs) or see any traffic on your LAN that’s not specifically passing through the tunnel to or from the outside.

    Cloudflare is a massive CDN provider, trusted to do exactly this sort of thing with the private data of equally massive companies, and they’re compliant with GDPR and other such regulations. Ultimately, the likelihood that they give the slightest jot about what passes through your tunnel as an individual user is minute, but whether you’re comfortable with them handling your data is something only you can decide.

    There’s a decent question and answer about the same thing here: https://community.cloudflare.com/t/what-data-does-cloudflare-actually-see/28660


  • Admittedly I’m paranoid, but I’d be looking to:

    1. Isolate your personal data from any web facing servers as much as possible. I break my own rule here with Immich, but I also…
    2. Use a Cloudflare tunnel instead of opening ports on your router directly. This gets your IP address out of public record.
    3. Use Cloudflare’s WAF features to limit ingress to trusted countries at a minimum.
    4. If you can get your head around it, lock things down more with features like Cloudflare device authentication.
    5. Especially if you don’t do step 4: Integrate Crowdsec into your Nginx setup to block probes, known bot IPs, and common attack vectors.

    All of the above is free, but past step 2 can be difficult to setup. The peace of mind once it is, however, is worth it to me.





  • Synology has Container Manager, which is their GUI frontend for Docker, so if it’ll run in Docker it’ll run on a Syno NAS. I’m running Pihole on mine just fine.

    As for the M.2 drives, you can use non-Synology ones as storage. Don’t quote me on it but I’ve a feeling it “just works” in the EU where they’re not allowed to force you to use specific brands, but if it doesn’t then there’s a script that removes the restriction: https://github.com/007revad/Synology_enable_M2_volume

    You should check their repo as they have other useful scripts. I’m using the one that enables dedupe on non-SSD volumes myself.




  • I’m using plenty of containers, accelerated and otherwise, but I also want a full-blown desktop that I can access from wherever. Even on a wired LAN, streaming that desktop is slow and laggy when it’s hosted on my NAS, which I think is due to the lack of hardware acceleration on that system. I want to move the VM to a host that has that feature (currently running Ubuntu Server) but I need a hypervisor that doesn’t require its own desktop system to be installed in order to manage it.

    Plenty of good replies here to help me though.




  • To be fair the Synology lineup is confusing, but if you get the right model - one with a Ryzen processor and support for 32GB memory (officially; they can take more) - then you’ve got yourself a proper little workhorse with low power consumption, a stable, reliable OS, and super easy expansion thanks to the hot-swap drive bays and their Hybrid RAID option. My 8 bay model is running a couple of full-blown VMs and what must be two dozen or so docker containers while barely breaking a sweat. The DS723+ is the equivalent 2 bay model.

    For things that need some acceleration like Plex and Immich I’ve added a little N100 box (a Beelink S12 Pro) with Ubuntu Server and another Docker instance, and mounted the NAS storage via SMB. This also sips power even when transcoding 4x Plex streams at once.

    All of which is to say you don’t need to do a complex, potentially power hungry and difficult to expand self build to do what you want.






  • I’ll have to have a look when I’m next in the vacinity but I’m pretty sure I have an APC Easy UPS on mine and it works out of the box.

    Let me get back to you…

    Update: It’s an APC Back-UPS 850. No doubt the instructions banged on about requiring Powerchute but I just plugged it into the Syno and it worked fine. You do need to enable UPS support on the NAS itself of course, from Control Panel/Hardware & Power/UPS, and set it to USB UPS.


  • Unless you’re hosting VHDs and need maximum throughput (in which case use NFS), SMB is going to be the easiest to setup and maintain across those 4 platforms.

    The Linux SMB implementation is decent and supports the latest version of the protocol (or close to, at least) whereas NFS in Windows ain’t so great and is a bit of a pig to get working in my experience.