- Deleted outdated files related to repository audit and deployment status, including AUDIT_COMPLETE.md, AUDIT_FIXES_APPLIED.md, FINAL_DEPLOYMENT_STATUS.md, and others. - Cleaned up documentation to streamline the repository and improve clarity for future maintenance. - Updated README and other relevant documentation to reflect the removal of these files.
2.5 KiB
2.5 KiB
Proxmox Cluster Configuration: sankofa-sfv-01
Last Updated: 2024-12-19
Cluster Name: sankofa-sfv-01
Cluster Overview
- Cluster Name: sankofa-sfv-01
- Node Count: 2
- Status: Active (verified via node visibility)
Cluster Nodes
Node 1: ML110-01
- IP: 192.168.11.10
- FQDN: ml110-01.sankofa.nexus
- Site: us-sfvalley
- Status: Online
- Cluster Role: Primary/First Node
Node 2: R630-01
- IP: 192.168.11.11
- FQDN: r630-01.sankofa.nexus
- Site: us-sfvalley-2
- Status: Online
- Cluster Role: Secondary/Second Node
Cluster Verification
Evidence of Cluster Existence
✅ Node Visibility:
- ML110-01 can see both nodes (r630-01 and ml110-01) in nodes list
- This indicates cluster membership (standalone nodes only see themselves)
Verification Methods
-
Proxmox Web UI:
- Log in to either node
- Navigate to: Datacenter → Cluster
- Verify cluster name: sankofa-sfv-01
- Verify both nodes listed
-
SSH Commands:
pvecm status pvecm nodes -
Corosync Configuration:
cat /etc/pve/corosync.conf
Quorum Configuration
For a 2-node cluster, quorum must be explicitly configured:
# On both nodes
pvecm expected 2
pvecm status
Important: Without quorum configuration, a 2-node cluster may have issues if one node goes offline.
Cluster Features
Once cluster is verified, you can:
- ✅ Create VMs on either node
- ✅ Migrate VMs between nodes
- ✅ Manage storage across cluster
- ✅ Use cluster-level operations
Network Configuration
- Cluster Network: 192.168.11.0/24
- Corosync Ports: 5404-5405 (should be open between nodes)
- API Port: 8006 (HTTPS)
Storage Considerations
- Each node has local storage
- For shared storage, consider:
- NFS shares
- Ceph storage
- Shared LVM volumes
Monitoring
- Cluster status: Check via Web UI or
pvecm status - Node status: Check via Web UI or API
- Quorum status: Check via
pvecm status
Troubleshooting
Cluster Split-Brain
If cluster splits:
# On majority node
pvecm expected 2
# On minority node (if needed)
pvecm expected 1
Node Cannot Join
- Check network connectivity
- Verify firewall rules
- Check corosync service:
systemctl status corosync - Review logs:
journalctl -u corosync