- Add comprehensive database migrations (001-024) for schema evolution - Enhance API schema with expanded type definitions and resolvers - Add new middleware: audit logging, rate limiting, MFA enforcement, security, tenant auth - Implement new services: AI optimization, billing, blockchain, compliance, marketplace - Add adapter layer for cloud integrations (Cloudflare, Kubernetes, Proxmox, storage) - Update Crossplane provider with enhanced VM management capabilities - Add comprehensive test suite for API endpoints and services - Update frontend components with improved GraphQL subscriptions and real-time updates - Enhance security configurations and headers (CSP, CORS, etc.) - Update documentation and configuration files - Add new CI/CD workflows and validation scripts - Implement design system improvements and UI enhancements
3.2 KiB
3.2 KiB
Ceph-Proxmox Integration Guide
Last Updated: 2024-12-19
Overview
This guide covers integrating Ceph storage with Proxmox VE for distributed storage across the cluster.
Storage Types
RBD (RADOS Block Device)
Use Case: VM disk images, high-performance block storage
Features:
- Thin provisioning
- Snapshots
- Live migration support
- High IOPS
CephFS
Use Case: ISO images, backups, snippets, shared storage
Features:
- POSIX-compliant filesystem
- Shared access
- Snapshots
- Quotas
Integration Steps
Step 1: Verify Ceph Cluster
# Check cluster status
ceph -s
# Verify pools exist
ceph osd pool ls
Step 2: Add RBD Storage
Via CLI:
# On each Proxmox node
pvesm add rbd ceph-rbd \
--pool rbd \
--monhost 192.168.11.10,192.168.11.11 \
--username admin \
--content images,rootdir \
--krbd 1
Via Web UI:
- Datacenter → Storage → Add
- Select "RBD"
- Configure:
- ID:
ceph-rbd - Pool:
rbd - Monitor Host:
192.168.11.10,192.168.11.11 - Username:
admin - Content: Images, Rootdir
- Enable KRBD: Yes
- ID:
Step 3: Add CephFS Storage
Via CLI:
# On each Proxmox node
pvesm add cephfs ceph-fs \
--monhost 192.168.11.10,192.168.11.11 \
--username admin \
--fsname cephfs \
--content iso,backup,snippets
Via Web UI:
- Datacenter → Storage → Add
- Select "CephFS"
- Configure:
- ID:
ceph-fs - Monitor Host:
192.168.11.10,192.168.11.11 - Username:
admin - Filesystem:
cephfs - Content: ISO, Backup, Snippets
- ID:
Using Ceph Storage
Creating VMs with RBD Storage
-
Via Web UI:
- Create VM → Hard Disk
- Storage: Select
ceph-rbd - Configure size and format
-
Via CLI:
qm create 100 --name test-vm --memory 2048 qm disk create 100 ceph-rbd:0 --size 20G
Migrating VMs to Ceph
# Move disk to Ceph storage
qm disk move 100 scsi0 ceph-rbd --delete
# Or via Web UI:
# VM → Hardware → Disk → Move Storage
Monitoring
Ceph Dashboard
# Enable dashboard
ceph mgr module enable dashboard
# Create user
ceph dashboard ac-user-create admin <password> administrator
# Access: https://ml110-01.sankofa.nexus:8443
Proxmox Storage Status
# Check storage status
pvesm status
# Check storage usage
pvesm list
Best Practices
-
Pool Configuration:
- Use separate pools for different workloads
- Configure appropriate PG count
- Set size and min_size appropriately
-
Performance:
- Use SSD for OSD journals (if available)
- Configure network bonding for redundancy
- Monitor OSD performance
-
Backup:
- Use CephFS for backups
- Configure snapshot schedules
- Test restore procedures
Troubleshooting
Storage Not Available
# Check Ceph status
ceph -s
# Check OSD status
ceph osd tree
# Check storage in Proxmox
pvesm status
Performance Issues
# Check OSD performance
ceph osd perf
# Check pool stats
ceph df detail