Dear Ceph Users, Can you please address my scenario and suggest me a solution.
Regards Gaurav Goyal On Tue, Aug 16, 2016 at 11:13 AM, Gaurav Goyal <er.gauravgo...@gmail.com> wrote: > Hello > > > I need your help to redesign my ceph storage network. > > As suggested in earlier discussions, i must not use SAN storage. So we > have decided to removed it. > > Now we are ordering Local HDDs. > > My Network would be > > Host1 --> Controller + COmpute --> Local Disk 600GB Host 2--> Compute2 --> > Local Disk 600GB Host 3 --> Compute2 > > Is it right setup for ceph network? For Host1 and Host2 , we are using 1 > 600GB disk for basic filesystem. > > Should we use same size storage disks for ceph environment or i can order > Disks in size of 2TB for ceph cluster? > > Making it > > 2T X 2 on Host1 2T X 2 on Host 2 2T X 2 on Host 3 > > 12TB in total. replication factor 2 should make it 6 TB? > > > Regards > > Gaurav Goyal > > On Thu, Aug 4, 2016 at 1:52 AM, Bharath Krishna <bkris...@walmartlabs.com> > wrote: > >> Hi Gaurav, >> >> There are several ways to do it depending on how you deployed your ceph >> cluster. Easiest way to do it is using ceph-ansible with purge-cluster yaml >> ready made to wipe off CEPH. >> >> https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml >> >> You may need to configure ansible inventory with ceph hosts. >> >> Else if you want to purge manually, you can do it using: >> http://docs.ceph.com/docs/hammer/rados/deployment/ceph-deploy-purge/ >> >> >> Thanks >> Bharath >> >> From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Gaurav >> Goyal <er.gauravgo...@gmail.com> >> Date: Thursday, August 4, 2016 at 8:19 AM >> To: David Turner <david.tur...@storagecraft.com> >> Cc: ceph-users <ceph-users@lists.ceph.com> >> Subject: Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to >> Local Disks >> >> Please suggest a procedure for this uninstallation process? >> >> >> Regards >> Gaurav Goyal >> >> On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal <er.gauravgo...@gmail.com >> <mailto:er.gauravgo...@gmail.com>> wrote: >> >> Thanks for your prompt >> response! >> >> Situation is bit different now. Customer want us to remove the ceph >> storage configuration from scratch. Let is openstack system work without >> ceph. Later on install ceph with local disks. >> >> So I need to know a procedure to uninstall ceph and unconfigure it from >> openstack. >> >> Regards >> Gaurav Goyal >> On 03-Aug-2016 4:59 pm, "David Turner" <david.tur...@storagecraft.com >> <mailto:david.tur...@storagecraft.com>> wrote: >> If I'm understanding your question correctly that you're asking how to >> actually remove the SAN osds from ceph, then it doesn't matter what is >> using the storage (ie openstack, cephfs, krbd, etc) as the steps are the >> same. >> >> I'm going to assume that you've already added the new storage/osds to the >> cluster, weighted the SAN osds to 0.0 and that the backfilling has >> finished. If that is true, then your disk used space on the SAN's should >> be basically empty while the new osds on the local disks should have a fair >> amount of data. If that is the case, then for every SAN osd, you just run >> the following commands replacing OSD_ID with the osd's id: >> >> # On the server with the osd being removed >> sudo stop ceph-osd id=OSD_ID >> ceph osd down OSD_ID >> ceph osd out OSD_ID >> ceph osd crush remove osd.OSD_ID >> ceph auth del osd.OSD_ID >> ceph osd rm OSD_ID >> >> Test running those commands on a test osd and if you had set the weight >> of the osd to 0.0 previously and if the backfilling had finished, then what >> you should see is that your cluster has 1 less osd than it used to, and no >> pgs should be backfilling. >> >> HOWEVER, if my assumptions above are incorrect, please provide the output >> of the following commands and try to clarify your question. >> >> ceph status >> ceph osd tree >> >> I hope this helps. >> >> > Hello David, >> > >> > Can you help me with steps/Procedure to uninstall Ceph storage from >> openstack environment? >> > >> > >> > Regards >> > Gaurav Goyal >> ________________________________ >> [cid:image001.jpg@01D1EE42.88EF6E60]<https://storagecraft.com> >> >> David Turner | Cloud Operations Engineer | StorageCraft Technology >> Corporation<https://storagecraft.com> >> 380 Data Drive Suite 300 | Draper | Utah | 84020 >> Office: 801.871.2760 | Mobile: 385.224.2943<tel:385.224.2943> >> >> ________________________________ >> If you are not the intended recipient of this message or received it >> erroneously, please notify the sender and delete it, together with any >> attachments, and be advised that any dissemination or copying of this >> message is prohibited. >> >> ________________________________ >> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com