Separate Pool Configuration for DR Zone in Ceph Multisite

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Team, 



I would like to know if it's possible to configure a separate pool for the DR zone in a Ceph multisite setup. I have two clusters, Zone A and Zone B. Currently, data synced from Zone A (CBE) to Zone B (TUP) resides in the default pool of Zone B. Instead, I want the replicated data from Zone A to be stored in its own separate pool in Zone B. 



Here are the steps I've used to set up the multisite configuration: 




Zone A: 
radosgw-admin realm create --rgw-realm=IN --default
radosgw-admin zonegroup create --rgw-zonegroup=TN --master --default
radosgw-admin zone create --rgw-zone=CBE --rgw-zonegroup=TN --master --default
radosgw-admin user create --uid=replica --display-name="replica_user" --system
radosgw-admin period update --commit 


Zone B: 
radosgw-admin realm pull --url=10.100.1.78 --access-key=xxx--secret=xxxx
radosgw-admin realm default --rgw-realm=IN
radosgw-admin zone create --rgw-zonegroup=TN --rgw-zone=TUP --access-key=xxx--secret=xxxx--endpoints=10.100.1.79
radosgw-admin period update --commit 


Could anyone guide me on how to achieve the separate pool configuration for the replicated data in Zone B? 



Thanks in advance! 




Best Regards, 
Vignesh Varma G 
Cloud Engineer 
www.stackbill.com 
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux