Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.moodmnky.com/llms.txt

Use this file to discover all available pages before exploring further.

Overview

The MOOD MNKY Data Center is built on a five-node Proxmox VE 8.4 cluster designed to host AI workloads, automation stacks, media services, and experimental homelab environments. This page provides a high-level view of the cluster, its hardware profile, and how the rest of the documentation set is organized. It is written for internal SRE and DevOps engineers who already understand Linux, Proxmox, ZFS, and LXC. For deeper detail, refer to:

Network segments (VLAN identity)

Each cluster hypervisor sits on a site /24 behind pfSense; remote access uses NetBird to the hub peer. For the canonical map of CIDRs, management IPs, intent (DATA core vs MOOD public plane vs CODE automation, and so on), and example anchors, read VLAN subnets and identity.

Cluster at a glance

The Proxmox cluster consists of the following nodes:
NodeRole / notesCPUCores / ThreadsRAM (approx)Status
CODE-MNKYPrimary GPU host; LXCs 300/301/3001 + VMs 3055–3056 (live)AMD Ryzen 5 4600G6 / 12125 GiBOnline
CASA-MNKYGeneral-purpose capacityAMD Ryzen 5 4600G6 / 1262 GiBOnline
DATA-MNKYHigh-core-capacity nodeAMD Ryzen 7 5700X8 / 1662.7 GiBOnline
SAGE-MNKYAdditional compute & storage (STUD-zfs)AMD Ryzen 7 5700G8 / 1662 GiBOnline
MOOD-MNKYMedia stack (LXC 120), Intel iGPU / VA-APIIntel Core i9-13900KS~125 GiBOnline
Core characteristics:
  • Hypervisor: Proxmox VE 8.4.17, kernel 6.8.12-19-pve, EFI boot, Secure Boot disabled.
  • Workloads (see Data Center Map and CODE-MNKY LXC inventory):
    • Supabase and n8n as QEMU VMs 3055 and 3056 on CODE-MNKY (canonical production).
    • PegaProx (Proxmox UI helper) on LXC 3001 on CODE-MNKY.
    • LXC 300/301 reserved or legacy: automation and AI tags; no Docker stacks running at last inventory — reprovision or reclaim capacity as needed.
    • Media stack (Jellyfin, *arr, qBittorrent) on MOOD-MNKY LXC 120 (not on CODE-MNKY).
  • Storage:
    • Node-local ZFS pools (CODE-MAIN-zfs, CODE-BKP-zfs, STUD-zfs, rpool on each node).
    • Shared NFS export hyper-mnky-shared mounted across nodes.
  • Networking:
    • Linux bridges for LXC connectivity to the cluster LAN.
    • Integration with TrueNAS for NFS and per-VM datashare.
    • Optional Cloudflare tunnels for external access.
Full node-level hardware snapshots are captured under /root/hardware-snapshots/<node>/<timestamp>/ on each Proxmox host.

Documentation layout

The Data Center documentation is organized as follows:
  • Data Center Map: comprehensive first-principles map of the cluster—every node, VM, LXC, storage, and topology. Use it as the single source of truth for “what runs where.”
  • VLAN subnets and identity: canonical layer-3 segment roles (DATA / MOOD / SAGE / CODE / CASA), management IPs, and exposure model.
  • Data Center Upgrade Plan: living roadmap for phased upgrades (PBS, GitLab, Docker registry, workload rebalance). Updated as phases complete.
  • Overview (this page): global view of the cluster and doc set.
  • Nodes: summarized and per-node hardware/service profiles.
  • CODE-MNKY deep dive: detailed hardware and capabilities of the main GPU node.
  • Storage & network: ZFS, NFS, and network topology across the cluster.
  • Runbooks: repeatable procedures for snapshots, expansion, and incident response.
These pages are intended to be kept in sync with:
  • Terraform and Ansible definitions under proxmox-terraform/ and proxmox-ansible/.
  • The homelab “as code” docs and final report.
  • Actual hardware snapshots collected on the nodes.

Intended usage

Typical scenarios for these docs:
  • Capacity planning before placing new workloads.
  • Verifying hardware details before upgrades or replacements.
  • Using runbooks during node, storage, or GPU incidents.
  • Onboarding new SREs to the internal topology.
Whenever you change hardware (RAM, disks, GPU) or add/remove nodes:
  1. Re-run the hardware snapshot collector on each affected node.
  2. Refresh per-node sections in the Nodes and CODE-MNKY pages.
  3. Update cluster summaries here and in Storage & network.