Homelab is the hobby of building and maintaining your own computer systems for the purpose of experimentation or learning about techniques used in managing larger computer systems. In my case, I enjoy managing network, storage, and virtualixation infrastructure at home and dedicate some of my time to building out my homelab.
I get asked a lot about what hardware I recommend for homelabs and home servers. It’s a very difficult question since it depends on what exactly you want to get out of your setup. But, whatever you choose, I’m starting a new series where I’ll setup all of the commonly requested home server software in a single box. Since I want to try this on both used and new hardware, here’s a low cost NAS you can buy brand-new and run your own software on it!
In this video, I expand on the last video of my hyper-converged Proxmox + Ceph cluster to create more custom pool layouts than Proxmox’s GUI allows. This includes setting the storage class (HDD / SSD / NVME), failure domain, and even erasure coding of pools. All of this is then setup as a storage location in Proxmox for RBD (RADOS Block Device), so we can store VM disks on it.
So, you want to send HDMI video around your house? Maybe you want to use your office computer on your living room TV without a proprietary streaming solution like AirPlay or Chromecast? Share a cable or satellite box between your living room and bedroom? Or you’re crazy like me and you want to put all of your computers into the basement, and connect to any of them from any desk in the house?
Short tutorial this week, how to import a VM image you may get to use in Proxmox. I took a vacation for Thanksgiving, so back to the regularly scheduled madness soon. Enjoy!
The Basics: You need to use qm importvm (See the man page here if you want to reference it). The basic command is:
qm importvm <vmid> <source> <storage> <options> Where <vmid> is the number of a VM which alraedy exists, <source> is the path to a file which qemu can convert (qcow2, vmdk,raw img), and <storage> is the name of a storage location in Proxmox.
I’m sure many of you follow me because you use Proxmox. It’s been a staple of my content for some time now. So, while working on the next episode of the Ceph series, I thought it would be good to do a separate segment on networking. So, here you have it. The basics of VLANs, Bridges, and Bonds in Proxmox VE. I’m only covering the native Linux versions, not Open VSwitch and VXLAN.
I have a thin client addiction. You all have seen my 3x Dell Wyse 5060 Hyperconverged Cluster project. And you know that I bought a Dell Wyse 3040. But, I actually bought 3x 3040s, and someone sent me a Wyse 7010, and an HP T620 (yet to be reviewed). And now I bought another. An HP T530.
I’d consider this to be an excellent choice for anyone wanting to run Home Assistant, since it has enough power for, an upgradeable M.
Proxmox has a pretty good backup scheduler, but it relies on the backup destination being mounted as a storage location. This implies that the backup destination needs to be a protocol that Proxmox supports - SMB (CIFS), NFS, … or Proxmox Backup Server. If you want to push your backups to a cloud service, you probably need something a bit more complicated. Thankfully, Proxmox’s backup scheduler thought about this and has a hook feature we can use for this purpose, and we can use any protocol supported on the Debian base system, including things such as FUSE or s3cmd.
Today, I open a new gift to the homelab - an HP MicroServer Gen8. This little chonky cube is full of hard drives and not a whole lot else, making it a perfect test system for ZFS, TrueNAS, Proxmox VE and Proxmox Backup Server, etc. and I’m already planning the videos I want to make with it. So come along as I open it up and see roughly what’s inside, the specs, and use HP iLO (their proprietary IPMI) for the first time.
Casually tearing down the Dell Wyse 7010 (Zx0) that was kindly sent in by a viewer named Tom! Thanks Tom for making this happen. Tom also sent an HP thin client, but you guys only get one treat per video, and that one needed Torx bits that I didn’t have handy on the bench. tl;dr the CPU supports AMD-V (not that you really have enough RAM to think about virtualization, but you can expand it with ordinary DDR3 DIMMs), the GPU is kinda awful, and it has a Realtek NIC.
This is a continuation of my previous article on the Net Booted Thin Client. The instructions got way too long, so I created a new article for the client setup. You need a functional server setup (TFTP, HTTP, iPXE) which I did in my previous post. I could use something like Linux Terminal Server Project, but that’s a bit overkill for this, and I wanted to learn Alpine anyway, so I’ve chosen to use Alpine Linux for the client operating system.