Dear Ceph Users,

Can you please address my scenario and suggest me a solution.

Regards
Gaurav Goyal

On Tue, Aug 16, 2016 at 11:13 AM, Gaurav Goyal <er.gauravgo...@gmail.com>
wrote:

> Hello
>
>
> I need your help to redesign my ceph storage network.
>
> As suggested in earlier discussions, i must not use SAN storage. So we
> have decided to removed it.
>
> Now we are ordering Local HDDs.
>
> My Network would be
>
> Host1 --> Controller + COmpute --> Local Disk 600GB Host 2--> Compute2 -->
> Local Disk 600GB Host 3 --> Compute2
>
> Is it right setup for ceph network? For Host1 and Host2 , we are using 1
> 600GB disk for basic filesystem.
>
> Should we use same size storage disks for ceph environment or i can order
> Disks in size of 2TB for ceph cluster?
>
> Making it
>
> 2T X 2 on Host1 2T X 2 on Host 2 2T X 2 on Host 3
>
> 12TB in total. replication factor 2 should make it 6 TB?
>
>
> Regards
>
> Gaurav Goyal
>
> On Thu, Aug 4, 2016 at 1:52 AM, Bharath Krishna <bkris...@walmartlabs.com>
> wrote:
>
>> Hi Gaurav,
>>
>> There are several ways to do it depending on how you deployed your ceph
>> cluster. Easiest way to do it is using ceph-ansible with purge-cluster yaml
>> ready made to wipe off CEPH.
>>
>> https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml
>>
>> You may need to configure ansible inventory with ceph hosts.
>>
>> Else if you want to purge manually, you can do it using:
>> http://docs.ceph.com/docs/hammer/rados/deployment/ceph-deploy-purge/
>>
>>
>> Thanks
>> Bharath
>>
>> From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Gaurav
>> Goyal <er.gauravgo...@gmail.com>
>> Date: Thursday, August 4, 2016 at 8:19 AM
>> To: David Turner <david.tur...@storagecraft.com>
>> Cc: ceph-users <ceph-users@lists.ceph.com>
>> Subject: Re: [ceph-users] Fwd: Ceph Storage Migration from SAN storage to
>> Local Disks
>>
>> Please suggest a procedure for this uninstallation process?
>>
>>
>> Regards
>> Gaurav Goyal
>>
>> On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal <er.gauravgo...@gmail.com
>> <mailto:er.gauravgo...@gmail.com>> wrote:
>>
>> Thanks for your  prompt
>> response!
>>
>> Situation is bit different now. Customer want us to remove the ceph
>> storage configuration from scratch. Let is openstack system work without
>> ceph. Later on install ceph with local disks.
>>
>> So I need to know a procedure to uninstall ceph and unconfigure it from
>> openstack.
>>
>> Regards
>> Gaurav Goyal
>> On 03-Aug-2016 4:59 pm, "David Turner" <david.tur...@storagecraft.com
>> <mailto:david.tur...@storagecraft.com>> wrote:
>> If I'm understanding your question correctly that you're asking how to
>> actually remove the SAN osds from ceph, then it doesn't matter what is
>> using the storage (ie openstack, cephfs, krbd, etc) as the steps are the
>> same.
>>
>> I'm going to assume that you've already added the new storage/osds to the
>> cluster, weighted the SAN osds to 0.0 and that the backfilling has
>> finished.  If that is true, then your disk used space on the SAN's should
>> be basically empty while the new osds on the local disks should have a fair
>> amount of data.  If that is the case, then for every SAN osd, you just run
>> the following commands replacing OSD_ID with the osd's id:
>>
>> # On the server with the osd being removed
>> sudo stop ceph-osd id=OSD_ID
>> ceph osd down OSD_ID
>> ceph osd out OSD_ID
>> ceph osd crush remove osd.OSD_ID
>> ceph auth del osd.OSD_ID
>> ceph osd rm OSD_ID
>>
>> Test running those commands on a test osd and if you had set the weight
>> of the osd to 0.0 previously and if the backfilling had finished, then what
>> you should see is that your cluster has 1 less osd than it used to, and no
>> pgs should be backfilling.
>>
>> HOWEVER, if my assumptions above are incorrect, please provide the output
>> of the following commands and try to clarify your question.
>>
>> ceph status
>> ceph osd tree
>>
>> I hope this helps.
>>
>> > Hello David,
>> >
>> > Can you help me with steps/Procedure to uninstall Ceph storage from
>> openstack environment?
>> >
>> >
>> > Regards
>> > Gaurav Goyal
>> ________________________________
>> [cid:image001.jpg@01D1EE42.88EF6E60]<https://storagecraft.com>
>>
>> David Turner | Cloud Operations Engineer | StorageCraft Technology
>> Corporation<https://storagecraft.com>
>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>> Office: 801.871.2760 | Mobile: 385.224.2943<tel:385.224.2943>
>>
>> ________________________________
>> If you are not the intended recipient of this message or received it
>> erroneously, please notify the sender and delete it, together with any
>> attachments, and be advised that any dissemination or copying of this
>> message is prohibited.
>>
>> ________________________________
>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to