Files
Sankofa/docs/proxmox/status/CLUSTER_CONFIGURATION.md
defiQUG a8106e24ee Remove obsolete audit and deployment documentation files
- Deleted outdated files related to repository audit and deployment status, including AUDIT_COMPLETE.md, AUDIT_FIXES_APPLIED.md, FINAL_DEPLOYMENT_STATUS.md, and others.
- Cleaned up documentation to streamline the repository and improve clarity for future maintenance.
- Updated README and other relevant documentation to reflect the removal of these files.
2025-12-12 19:42:31 -08:00

121 lines
2.5 KiB
Markdown

# Proxmox Cluster Configuration: sankofa-sfv-01
**Last Updated**: 2024-12-19
**Cluster Name**: sankofa-sfv-01
## Cluster Overview
- **Cluster Name**: sankofa-sfv-01
- **Node Count**: 2
- **Status**: Active (verified via node visibility)
## Cluster Nodes
### Node 1: ML110-01
- **IP**: 192.168.11.10
- **FQDN**: ml110-01.sankofa.nexus
- **Site**: us-sfvalley
- **Status**: Online
- **Cluster Role**: Primary/First Node
### Node 2: R630-01
- **IP**: 192.168.11.11
- **FQDN**: r630-01.sankofa.nexus
- **Site**: us-sfvalley-2
- **Status**: Online
- **Cluster Role**: Secondary/Second Node
## Cluster Verification
### Evidence of Cluster Existence
**Node Visibility**:
- ML110-01 can see both nodes (r630-01 and ml110-01) in nodes list
- This indicates cluster membership (standalone nodes only see themselves)
### Verification Methods
1. **Proxmox Web UI**:
- Log in to either node
- Navigate to: Datacenter → Cluster
- Verify cluster name: sankofa-sfv-01
- Verify both nodes listed
2. **SSH Commands**:
```bash
pvecm status
pvecm nodes
```
3. **Corosync Configuration**:
```bash
cat /etc/pve/corosync.conf
```
## Quorum Configuration
For a 2-node cluster, quorum must be explicitly configured:
```bash
# On both nodes
pvecm expected 2
pvecm status
```
**Important**: Without quorum configuration, a 2-node cluster may have issues if one node goes offline.
## Cluster Features
Once cluster is verified, you can:
- ✅ Create VMs on either node
- ✅ Migrate VMs between nodes
- ✅ Manage storage across cluster
- ✅ Use cluster-level operations
## Network Configuration
- **Cluster Network**: 192.168.11.0/24
- **Corosync Ports**: 5404-5405 (should be open between nodes)
- **API Port**: 8006 (HTTPS)
## Storage Considerations
- Each node has local storage
- For shared storage, consider:
- NFS shares
- Ceph storage
- Shared LVM volumes
## Monitoring
- Cluster status: Check via Web UI or `pvecm status`
- Node status: Check via Web UI or API
- Quorum status: Check via `pvecm status`
## Troubleshooting
### Cluster Split-Brain
If cluster splits:
```bash
# On majority node
pvecm expected 2
# On minority node (if needed)
pvecm expected 1
```
### Node Cannot Join
1. Check network connectivity
2. Verify firewall rules
3. Check corosync service: `systemctl status corosync`
4. Review logs: `journalctl -u corosync`
## Related Documentation
- [Cluster Setup Guide](./CLUSTER_SETUP.md)
- [Cluster Status Check](./CLUSTER_STATUS_CHECK.md)
- [Task List](./TASK_LIST.md)