• 1 Post
  • 26 Comments
Joined 2 years ago
cake
Cake day: June 15th, 2023

help-circle


  • This is one of my pet peeves with containerized services, like why would I want to run three or four instances of mariadb? I get it, from the perspective of the packagers, who want a ‘just works’ solution to distribute, but if I’m trying to run simple services on a 4 GB RPi or a 2 GB VPS, then replicating dbs makes a difference. It took a while, but I did, eventually, get those dockers configured to use a single db backend, but I feel like that completely negated the ‘easy to set up and maintain’ rationale for containers.


  • The services you’ve mentioned are all pretty low compute impact, just bandwidth, so I’d expect your MBP to be fine. Transcoding for jellyfin is the only real wildcard, and that depends on your media and client setups. I run pihole, homeassistant, immich, and kodi on a raspberry pi 4 with plenty of overhead for more services. NAS is nice if your library outgrows a single disk and your storage bandwidth gets choked by USB multiplexing.

    My suggestion is to consider a cheap VPS and vanity domain for external access. Domains cheap as $5/year; fair VPSs cheap as $30/year. Use SSH to forward localhost ports on the VPS to container ports on the MBP, then nginx on thee VPS to reverse-proxy to those forwarded ports. You get unique names for every service, LetsEncrypt certificates, and an offsite location for critical backups. Make sure you are the one paying for VPS & DNS so they don’t get surprise-cancelled.






  • Not exactly the same, but I have an air quality sensor I use to turn the HVAC fan mode on/off to filter. Also a CO2 sensor. Both wired to the RPi I run homeassistant on. The HVAC is controlled via T6 pro Z-wave now, but I started out with a Zooz Zen15 switch to just turn the whole thing on/off.

    The CO2 sensor has been pretty stable for 4(?) years - it has an internal recalibration routine that resets its baseline based on the past week’s data. My readings range from 400-ish with the windows open & fans to 1200+ cooking with gas in the sealed house. Averages around 800 with the AC or 500 with the furnace (which exhausts combustion gasses). The aq sensor has been replaced once after 3-4(?) years. It reads exactly what purpleair says is outside with the windows open, drops to 0-2 µg/m3 with the filters running, spikes to 300+ cooking.



  • I switched from an I3-530, nominal TDP 73W, to an N-100, nominal TDP 7W, and power from the wall didn’t change at all. Even the i3 ran around 0.1 CPU load, except when transcoding, and I’m left with the impression that most of the power goes into HDDs, RAM, maybe fans, and PS losses. My sense is that the best way to decrease homelab power use is to minimize the number of devices. Start with your seyrver at 60W, add a WAP at 10-15W, maybe a switch at 10-15W… Not because of the CPUs, necessarily, but because every CPU every CPU comes with systems to keep the CPU going, keep the power regulated, etc.



  • My ISP seems to use just normal DHCP for assigning addresses and honors re-use requests. The only times my IP addresses have changed has been I’ve changed the MAC or UUID that connects. I’ve been off-line for a week, come back, and been given the same address. Both IPv4 and v6.

    If one really wants their home systems to be publicly accessible, it’s easy enough to get a cheap vanity domain and point it at whatever address. rDNS won’t work, which would probably interfere with email, but most services don’t really need it. It’s a bit more complicated to detect when your IP changes and script a DNS update, but certainly do-able, if (like OP) one is hell bent on avoiding any off-site hardware.





  • It really depends on what your data is and how hard it would be to recreate. I keep a spare HD in a $40/year bank box & rotate it every 3 months. Most of the content is media - pictures, movies, music. Financial records would be annoying to recreate, but if there’s a big enough disaster to force me to go to the off-site backups, I think that’ll be the least of my troubles. Some data logging has a replica database on a VPS.

    My upload speed is terrible, so I don’t want to put a media library in the cloud. If I did any important daily content creation, I’d probably keep that mirrored offsite with rsync, but I feel like the spirit of an offsite backup is offline and asynchronous, so things like ransomware don’t destroy your backups, too.