• 0 Posts
  • 7 Comments
Joined 2 years ago
cake
Cake day: June 13th, 2023

help-circle
  • What’s amusing is I’m long time stoner. As such, I have a shit memory. I do not remember writing this comment. Nor do I remember even struggling with this. I do know that I had a bunch of .crypt domains for a while.

    So your comment is hilarious because I frequently find my own comments when I’m struggling through that thing I once did that I don’t remember, documenting what I did.

    I do it to help others, I call it “leaving breadcrumbs for those further back on the path” but those breadcrumbs are great when a server dies and you have to re set it up.

    Kudos for being a great guy and leaving breadcrumbs. Karma likes to remind you that you’re a wonderful person sometimes, so just enjoy it, and don’t let the bastards grind you down.

    Rereading my own comment, I do this, I thank people hoping they’re still active at some point. I really do believe in thanking those that help me, even if they may not see it until 10 months later, if at all. You must have been the post that slotted it all into place.



  • I think I’m a step behind you. I use Uptime Kuma for monitoring and it worked really well. Just have it running on a pi separate from my main machine.

    I worked out how to get it sending me emails when things are down and up, and now my email inbox is a fucking hot mess of notifications.

    So I’ve just this weekend integrated it into Home Assistant and set it to notify me when things are down for 5 minutes or more.

    My next step was going to be finding some way of integrating Portainer into Home Assistant so I can restart stopped containers, and maybe Proxmox so I can reboot VMs from HA. Not sure it’s possible yet though.

    Ultimately I want to have HA send me a notification with actionable buttons with “reboot container” and “reboot VM” which, when pressed, will sort the issue out.

    However this will not help when one of my drives goes down. They’re HDDs plugged in by USB3 which isn’t great and my server is behind the coat rack so sometimes the kids just throw their coats on and it falls onto my server, which then heats up and goes silly.


  • So you’re trying to get 2 instances of qbt behind the same Gluetun vpn container?

    I don’t use Qbt but I certainly have done in the past. Am I correct in remembering that in the gui you can change the port?

    If so, maybe what you could do is set up your stack with 1 instance in, go into the GUI and change the port on the service to 8000 or 8081 or whatever.

    Map that port in your Gluetun config and leave the default port open for QBT, and add a second instance to the stack with a different name and addresses for the config files.

    Restart the stack and have 2 instances.


  • Hello again.

    I’ve gone through your steps outlined in this post now for LAN. I’ve made my own network name .crypt and added *.crypt to Adguard and pointed it at the IP address of Nginx.

    I’ve then gone and mapped my local services in Nginx. So radarr.crypt sonarr.crypt plex.crypt etc and mapped them to ports.

    Now what I enjoyed was that I had to map Adguard to forward to Nginx, but in Nginx I can use the IP address of anything on my network, not just on the host.

    So it’s map Adguard in DNS rewrites to Nginx IP, then map the IP:ports in Proxy Hosts in Nginx.

    Now when I use my Tailscale exit node (that I have from Home Assistant) I can use those addresses outside the house.

    I have noticed it only works for the .crypt domains, and not .local despite being set up as well. I guess because .local is a special address it is harder to map to Tailscale.

    Anyway, it’s working for me after following what you’ve done, I just did less in Tailscale because of the exit node