• 0 Posts
  • 27 Comments
Joined 1 year ago
cake
Cake day: July 9th, 2023

help-circle







  • All the images I used already had x86 variants available. In fact, I was building and pushing my own arm variants for a few images to my own Nexus repository which I’ve stopped since they aren’t necessary anymore.

    If you are using arm only images, you’ll need to build your own x86 variants and host them.

    I created a brand new cluster from scratch and then setup the same storage pv/PVCs and namespaces.

    Then I’d delete the workloads from the old cluster and apply the same yaml to the new cluster, and then update my DNS.

    I used kubectx to swap between them.

    Once I verified the new service was working I’d move to the next. Since the network storage was the same it was pretty seamless. If you’re using something like rook to utilize your nodes disks as network storage that would be much more difficult.

    After everything was moved I powered down the old cluster and waited a few weeks before I wiped the nodes. In case I needed to power it up and reapply a service to it temporarily.

    My old cluster was k8s on raspbian but my new one was all Talos. I also moved from single control plane to 3 machines control plane. (Which is completely unnecessary, but I just wanted to try it). But that had no effect on any services.


  • You can pin the pod to a specific node and pass through the USB device path and that will work. But the whole point of k8s is redundancy and workloads running anywhere.

    Plus for IOT networks like zigbee and zwave, controller position in your house is important. If your server is more centrally located that may not be a concern for you.

    I’ve heard of some using a USB serial over Ethernet device to relocate their controller remotely but i haven’t looked into that. Running this one off rpi for the controller just made more sense for me.


  • You can run it on proxmox if you want to mix non k8s machines onto the same hardware. All my k8s nodes are dedicated to running k8s only though, so there is no reason for me to have that extra step.

    I would not run k8s on proxmox so you can run multiple nodes on the same machine though, the only reason I could really see to do that is if you only had one machine and you really wanted to keep your controller and worker nodes separate.


  • I migrated from a mix of proxmox, hyper-V, bare metal, and Synology hosted docker onto a full k8s cluster.

    It is much easier to manage now, including adding or replacing nodes. Including a rebuild of the cluster from 7 rpis onto 7 elite desk mini PCs. (From arm to x86 and from Debian to Talos)

    But it wasn’t a small process either.

    You’ll have to deploy your k8s cluster, learn how to host the services you want (using a load balancer, dns setup, cluster IPs, etc), and setting up a storage provider (I use NFS to my Synology share, not the fastest or most secure but easiest)

    And then you’ll need to migrate your services off the old hardware onto the cluster one by one… Which means learning docker and k8s and how they work together.

    There are some things that I cannot host on the cluster like zwave2mqtt which requires a physical location centralized in my house and access to a USB zwave adapter. So even then not quite 100% ended up on the cluster, it runs on docker on an rpi though. (Technically you can do this if you pin the container to a single host and pass through the USB device, but I didn’t see a reason for it.)

    But, service upgrades, adding new services now that I’m used to it is very easy… Expanding compute is also pretty easy. So maintenance has gone down a bunch. But it was also a decent amount of work and learning to get there.

    K8s is relatively specialized knowledge compared to the general computer literate population that knows how computers generally work… So in terms of someone being able to take over your work, if they already know k8s, then it would be reasonably easy. If they don’t but are savvy enough to learn it would take a bit but not be too bad. If someone doesn’t already know their way around Linux and a terminal, it would probably not be possible for them to pick it up in a reasonable amount of time though.






  • orb360@lemmy.catoTechnology@lemmy.worldUnsmart a smart TV
    link
    fedilink
    English
    arrow-up
    11
    ·
    7 months ago

    The connection isn’t for you. It’s so the TV can fingerprint the content you watch, and then send that utilization data back to the company.

    You don’t need much bandwidth to do this.

    So with no wifi connection, and a blueray player, if you play Star Wars, they can fingerprint a few frames, send them back to Roku or whoever over sidewalk via your neighbors ring doorbell, and know you played star wars… Even with your completely offline setup