Files
Sankofa/docs/proxmox/status/CLUSTER_CONFIGURATION.md
defiQUG a8106e24ee Remove obsolete audit and deployment documentation files
- Deleted outdated files related to repository audit and deployment status, including AUDIT_COMPLETE.md, AUDIT_FIXES_APPLIED.md, FINAL_DEPLOYMENT_STATUS.md, and others.
- Cleaned up documentation to streamline the repository and improve clarity for future maintenance.
- Updated README and other relevant documentation to reflect the removal of these files.
2025-12-12 19:42:31 -08:00

2.5 KiB

Proxmox Cluster Configuration: sankofa-sfv-01

Last Updated: 2024-12-19
Cluster Name: sankofa-sfv-01

Cluster Overview

  • Cluster Name: sankofa-sfv-01
  • Node Count: 2
  • Status: Active (verified via node visibility)

Cluster Nodes

Node 1: ML110-01

  • IP: 192.168.11.10
  • FQDN: ml110-01.sankofa.nexus
  • Site: us-sfvalley
  • Status: Online
  • Cluster Role: Primary/First Node

Node 2: R630-01

  • IP: 192.168.11.11
  • FQDN: r630-01.sankofa.nexus
  • Site: us-sfvalley-2
  • Status: Online
  • Cluster Role: Secondary/Second Node

Cluster Verification

Evidence of Cluster Existence

Node Visibility:

  • ML110-01 can see both nodes (r630-01 and ml110-01) in nodes list
  • This indicates cluster membership (standalone nodes only see themselves)

Verification Methods

  1. Proxmox Web UI:

    • Log in to either node
    • Navigate to: Datacenter → Cluster
    • Verify cluster name: sankofa-sfv-01
    • Verify both nodes listed
  2. SSH Commands:

    pvecm status
    pvecm nodes
    
  3. Corosync Configuration:

    cat /etc/pve/corosync.conf
    

Quorum Configuration

For a 2-node cluster, quorum must be explicitly configured:

# On both nodes
pvecm expected 2
pvecm status

Important: Without quorum configuration, a 2-node cluster may have issues if one node goes offline.

Cluster Features

Once cluster is verified, you can:

  • Create VMs on either node
  • Migrate VMs between nodes
  • Manage storage across cluster
  • Use cluster-level operations

Network Configuration

  • Cluster Network: 192.168.11.0/24
  • Corosync Ports: 5404-5405 (should be open between nodes)
  • API Port: 8006 (HTTPS)

Storage Considerations

  • Each node has local storage
  • For shared storage, consider:
    • NFS shares
    • Ceph storage
    • Shared LVM volumes

Monitoring

  • Cluster status: Check via Web UI or pvecm status
  • Node status: Check via Web UI or API
  • Quorum status: Check via pvecm status

Troubleshooting

Cluster Split-Brain

If cluster splits:

# On majority node
pvecm expected 2

# On minority node (if needed)
pvecm expected 1

Node Cannot Join

  1. Check network connectivity
  2. Verify firewall rules
  3. Check corosync service: systemctl status corosync
  4. Review logs: journalctl -u corosync