Hello,

Awaiting any suggestion please!




Regards

On Wed, Aug 17, 2016 at 9:59 AM, Gaurav Goyal <er.gauravgo...@gmail.com>
wrote:

> Hello Brian,
>
> Thanks for your response!
>
> Can you please elaborate on this.
>
> Do you mean i must use
>
> 4 x 1TB HDD on each nodes rather than 2 x 2TB?
>
> This is going to be a lab environment. Can you please suggest to have best
> possible design for my lab environment.
>
>
>
> On Wed, Aug 17, 2016 at 9:54 AM, Brian :: <b...@iptel.co> wrote:
>
>> You're going to see pretty slow performance on a cluster this size
>> with spinning disks...
>>
>> Ceph scales very very well but at this type of size cluster it can be
>> challenging to get nice throughput and iops..
>>
>> for something small like this either use all ssd osds or consider
>> having more spinning osds per node backed by nvme or ssd journals..
>>
>>
>>
>> On Wed, Aug 17, 2016 at 1:14 PM, Gaurav Goyal <er.gauravgo...@gmail.com>
>> wrote:
>> > Dear Ceph Users,
>> >
>> > Can you please address my scenario and suggest me a solution.
>> >
>> > Regards
>> > Gaurav Goyal
>> >
>> > On Tue, Aug 16, 2016 at 11:13 AM, Gaurav Goyal <
>> er.gauravgo...@gmail.com>
>> > wrote:
>> >>
>> >> Hello
>> >>
>> >>
>> >> I need your help to redesign my ceph storage network.
>> >>
>> >> As suggested in earlier discussions, i must not use SAN storage. So we
>> >> have decided to removed it.
>> >>
>> >> Now we are ordering Local HDDs.
>> >>
>> >> My Network would be
>> >>
>> >> Host1 --> Controller + COmpute --> Local Disk 600GB Host 2--> Compute2
>> -->
>> >> Local Disk 600GB Host 3 --> Compute2
>> >>
>> >> Is it right setup for ceph network? For Host1 and Host2 , we are using
>> 1
>> >> 600GB disk for basic filesystem.
>> >>
>> >> Should we use same size storage disks for ceph environment or i can
>> order
>> >> Disks in size of 2TB for ceph cluster?
>> >>
>> >> Making it
>> >>
>> >> 2T X 2 on Host1 2T X 2 on Host 2 2T X 2 on Host 3
>> >>
>> >> 12TB in total. replication factor 2 should make it 6 TB?
>> >>
>> >>
>> >> Regards
>> >>
>> >> Gaurav Goyal
>> >>
>> >>
>> >> On Thu, Aug 4, 2016 at 1:52 AM, Bharath Krishna <
>> bkris...@walmartlabs.com>
>> >> wrote:
>> >>>
>> >>> Hi Gaurav,
>> >>>
>> >>> There are several ways to do it depending on how you deployed your
>> ceph
>> >>> cluster. Easiest way to do it is using ceph-ansible with
>> purge-cluster yaml
>> >>> ready made to wipe off CEPH.
>> >>>
>> >>> https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml
>> >>>
>> >>> You may need to configure ansible inventory with ceph hosts.
>> >>>
>> >>> Else if you want to purge manually, you can do it using:
>> >>> http://docs.ceph.com/docs/hammer/rados/deployment/ceph-deploy-purge/
>> >>>
>> >>>
>> >>> Thanks
>> >>> Bharath
>> >>>
>> >>> From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of
>> Gaurav
>> >>> Goyal <er.gauravgo...@gmail.com>
>> >>> Date: Thursday, August 4, 2016 at 8:19 AM
>> >>> To: David Turner <david.tur...@storagecraft.com>
>> >>> Cc: ceph-users <ceph-users@lists.ceph.com>
>> >>> Subject: Re: [ceph-users] Fwd: Ceph Storage Migration from SAN
>> storage to
>> >>> Local Disks
>> >>>
>> >>> Please suggest a procedure for this uninstallation process?
>> >>>
>> >>>
>> >>> Regards
>> >>> Gaurav Goyal
>> >>>
>> >>> On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal
>> >>> <er.gauravgo...@gmail.com<mailto:er.gauravgo...@gmail.com>> wrote:
>> >>>
>> >>> Thanks for your  prompt
>> >>> response!
>> >>>
>> >>> Situation is bit different now. Customer want us to remove the ceph
>> >>> storage configuration from scratch. Let is openstack system work
>> without
>> >>> ceph. Later on install ceph with local disks.
>> >>>
>> >>> So I need to know a procedure to uninstall ceph and unconfigure it
>> from
>> >>> openstack.
>> >>>
>> >>> Regards
>> >>> Gaurav Goyal
>> >>> On 03-Aug-2016 4:59 pm, "David Turner"
>> >>> <david.tur...@storagecraft.com<mailto:david.tur...@storagecraft.com>>
>> wrote:
>> >>> If I'm understanding your question correctly that you're asking how to
>> >>> actually remove the SAN osds from ceph, then it doesn't matter what
>> is using
>> >>> the storage (ie openstack, cephfs, krbd, etc) as the steps are the
>> same.
>> >>>
>> >>> I'm going to assume that you've already added the new storage/osds to
>> the
>> >>> cluster, weighted the SAN osds to 0.0 and that the backfilling has
>> finished.
>> >>> If that is true, then your disk used space on the SAN's should be
>> basically
>> >>> empty while the new osds on the local disks should have a fair amount
>> of
>> >>> data.  If that is the case, then for every SAN osd, you just run the
>> >>> following commands replacing OSD_ID with the osd's id:
>> >>>
>> >>> # On the server with the osd being removed
>> >>> sudo stop ceph-osd id=OSD_ID
>> >>> ceph osd down OSD_ID
>> >>> ceph osd out OSD_ID
>> >>> ceph osd crush remove osd.OSD_ID
>> >>> ceph auth del osd.OSD_ID
>> >>> ceph osd rm OSD_ID
>> >>>
>> >>> Test running those commands on a test osd and if you had set the
>> weight
>> >>> of the osd to 0.0 previously and if the backfilling had finished,
>> then what
>> >>> you should see is that your cluster has 1 less osd than it used to,
>> and no
>> >>> pgs should be backfilling.
>> >>>
>> >>> HOWEVER, if my assumptions above are incorrect, please provide the
>> output
>> >>> of the following commands and try to clarify your question.
>> >>>
>> >>> ceph status
>> >>> ceph osd tree
>> >>>
>> >>> I hope this helps.
>> >>>
>> >>> > Hello David,
>> >>> >
>> >>> > Can you help me with steps/Procedure to uninstall Ceph storage from
>> >>> > openstack environment?
>> >>> >
>> >>> >
>> >>> > Regards
>> >>> > Gaurav Goyal
>> >>> ________________________________
>> >>> [cid:image001.jpg@01D1EE42.88EF6E60]<https://storagecraft.com>
>> >>>
>> >>> David Turner | Cloud Operations Engineer | StorageCraft Technology
>> >>> Corporation<https://storagecraft.com>
>> >>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>> >>> Office: 801.871.2760 | Mobile: 385.224.2943<tel:385.224.2943>
>> >>>
>> >>> ________________________________
>> >>> If you are not the intended recipient of this message or received it
>> >>> erroneously, please notify the sender and delete it, together with any
>> >>> attachments, and be advised that any dissemination or copying of this
>> >>> message is prohibited.
>> >>>
>> >>> ________________________________
>> >>>
>> >>
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to