- Deleted outdated files related to repository audit and deployment status, including AUDIT_COMPLETE.md, AUDIT_FIXES_APPLIED.md, FINAL_DEPLOYMENT_STATUS.md, and others. - Cleaned up documentation to streamline the repository and improve clarity for future maintenance. - Updated README and other relevant documentation to reflect the removal of these files.
121 lines
2.5 KiB
Markdown
121 lines
2.5 KiB
Markdown
# Proxmox Cluster Configuration: sankofa-sfv-01
|
|
|
|
**Last Updated**: 2024-12-19
|
|
**Cluster Name**: sankofa-sfv-01
|
|
|
|
## Cluster Overview
|
|
|
|
- **Cluster Name**: sankofa-sfv-01
|
|
- **Node Count**: 2
|
|
- **Status**: Active (verified via node visibility)
|
|
|
|
## Cluster Nodes
|
|
|
|
### Node 1: ML110-01
|
|
- **IP**: 192.168.11.10
|
|
- **FQDN**: ml110-01.sankofa.nexus
|
|
- **Site**: us-sfvalley
|
|
- **Status**: Online
|
|
- **Cluster Role**: Primary/First Node
|
|
|
|
### Node 2: R630-01
|
|
- **IP**: 192.168.11.11
|
|
- **FQDN**: r630-01.sankofa.nexus
|
|
- **Site**: us-sfvalley-2
|
|
- **Status**: Online
|
|
- **Cluster Role**: Secondary/Second Node
|
|
|
|
## Cluster Verification
|
|
|
|
### Evidence of Cluster Existence
|
|
|
|
✅ **Node Visibility**:
|
|
- ML110-01 can see both nodes (r630-01 and ml110-01) in nodes list
|
|
- This indicates cluster membership (standalone nodes only see themselves)
|
|
|
|
### Verification Methods
|
|
|
|
1. **Proxmox Web UI**:
|
|
- Log in to either node
|
|
- Navigate to: Datacenter → Cluster
|
|
- Verify cluster name: sankofa-sfv-01
|
|
- Verify both nodes listed
|
|
|
|
2. **SSH Commands**:
|
|
```bash
|
|
pvecm status
|
|
pvecm nodes
|
|
```
|
|
|
|
3. **Corosync Configuration**:
|
|
```bash
|
|
cat /etc/pve/corosync.conf
|
|
```
|
|
|
|
## Quorum Configuration
|
|
|
|
For a 2-node cluster, quorum must be explicitly configured:
|
|
|
|
```bash
|
|
# On both nodes
|
|
pvecm expected 2
|
|
pvecm status
|
|
```
|
|
|
|
**Important**: Without quorum configuration, a 2-node cluster may have issues if one node goes offline.
|
|
|
|
## Cluster Features
|
|
|
|
Once cluster is verified, you can:
|
|
- ✅ Create VMs on either node
|
|
- ✅ Migrate VMs between nodes
|
|
- ✅ Manage storage across cluster
|
|
- ✅ Use cluster-level operations
|
|
|
|
## Network Configuration
|
|
|
|
- **Cluster Network**: 192.168.11.0/24
|
|
- **Corosync Ports**: 5404-5405 (should be open between nodes)
|
|
- **API Port**: 8006 (HTTPS)
|
|
|
|
## Storage Considerations
|
|
|
|
- Each node has local storage
|
|
- For shared storage, consider:
|
|
- NFS shares
|
|
- Ceph storage
|
|
- Shared LVM volumes
|
|
|
|
## Monitoring
|
|
|
|
- Cluster status: Check via Web UI or `pvecm status`
|
|
- Node status: Check via Web UI or API
|
|
- Quorum status: Check via `pvecm status`
|
|
|
|
## Troubleshooting
|
|
|
|
### Cluster Split-Brain
|
|
|
|
If cluster splits:
|
|
```bash
|
|
# On majority node
|
|
pvecm expected 2
|
|
|
|
# On minority node (if needed)
|
|
pvecm expected 1
|
|
```
|
|
|
|
### Node Cannot Join
|
|
|
|
1. Check network connectivity
|
|
2. Verify firewall rules
|
|
3. Check corosync service: `systemctl status corosync`
|
|
4. Review logs: `journalctl -u corosync`
|
|
|
|
## Related Documentation
|
|
|
|
- [Cluster Setup Guide](./CLUSTER_SETUP.md)
|
|
- [Cluster Status Check](./CLUSTER_STATUS_CHECK.md)
|
|
- [Task List](./TASK_LIST.md)
|
|
|