Hello, Awaiting any suggestion please!
Regards On Wed, Aug 17, 2016 at 9:59 AM, Gaurav Goyal <er.gauravgo...@gmail.com> wrote: > Hello Brian, > > Thanks for your response! > > Can you please elaborate on this. > > Do you mean i must use > > 4 x 1TB HDD on each nodes rather than 2 x 2TB? > > This is going to be a lab environment. Can you please suggest to have best > possible design for my lab environment. > > > > On Wed, Aug 17, 2016 at 9:54 AM, Brian :: <b...@iptel.co> wrote: > >> You're going to see pretty slow performance on a cluster this size >> with spinning disks... >> >> Ceph scales very very well but at this type of size cluster it can be >> challenging to get nice throughput and iops.. >> >> for something small like this either use all ssd osds or consider >> having more spinning osds per node backed by nvme or ssd journals.. >> >> >> >> On Wed, Aug 17, 2016 at 1:14 PM, Gaurav Goyal <er.gauravgo...@gmail.com> >> wrote: >> > Dear Ceph Users, >> > >> > Can you please address my scenario and suggest me a solution. >> > >> > Regards >> > Gaurav Goyal >> > >> > On Tue, Aug 16, 2016 at 11:13 AM, Gaurav Goyal < >> er.gauravgo...@gmail.com> >> > wrote: >> >> >> >> Hello >> >> >> >> >> >> I need your help to redesign my ceph storage network. >> >> >> >> As suggested in earlier discussions, i must not use SAN storage. So we >> >> have decided to removed it. >> >> >> >> Now we are ordering Local HDDs. >> >> >> >> My Network would be >> >> >> >> Host1 --> Controller + COmpute --> Local Disk 600GB Host 2--> Compute2 >> --> >> >> Local Disk 600GB Host 3 --> Compute2 >> >> >> >> Is it right setup for ceph network? For Host1 and Host2 , we are using >> 1 >> >> 600GB disk for basic filesystem. >> >> >> >> Should we use same size storage disks for ceph environment or i can >> order >> >> Disks in size of 2TB for ceph cluster? >> >> >> >> Making it >> >> >> >> 2T X 2 on Host1 2T X 2 on Host 2 2T X 2 on Host 3 >> >> >> >> 12TB in total. replication factor 2 should make it 6 TB? >> >> >> >> >> >> Regards >> >> >> >> Gaurav Goyal >> >> >> >> >> >> On Thu, Aug 4, 2016 at 1:52 AM, Bharath Krishna < >> bkris...@walmartlabs.com> >> >> wrote: >> >>> >> >>> Hi Gaurav, >> >>> >> >>> There are several ways to do it depending on how you deployed your >> ceph >> >>> cluster. Easiest way to do it is using ceph-ansible with >> purge-cluster yaml >> >>> ready made to wipe off CEPH. >> >>> >> >>> https://github.com/ceph/ceph-ansible/blob/master/purge-cluster.yml >> >>> >> >>> You may need to configure ansible inventory with ceph hosts. >> >>> >> >>> Else if you want to purge manually, you can do it using: >> >>> http://docs.ceph.com/docs/hammer/rados/deployment/ceph-deploy-purge/ >> >>> >> >>> >> >>> Thanks >> >>> Bharath >> >>> >> >>> From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of >> Gaurav >> >>> Goyal <er.gauravgo...@gmail.com> >> >>> Date: Thursday, August 4, 2016 at 8:19 AM >> >>> To: David Turner <david.tur...@storagecraft.com> >> >>> Cc: ceph-users <ceph-users@lists.ceph.com> >> >>> Subject: Re: [ceph-users] Fwd: Ceph Storage Migration from SAN >> storage to >> >>> Local Disks >> >>> >> >>> Please suggest a procedure for this uninstallation process? >> >>> >> >>> >> >>> Regards >> >>> Gaurav Goyal >> >>> >> >>> On Wed, Aug 3, 2016 at 5:58 PM, Gaurav Goyal >> >>> <er.gauravgo...@gmail.com<mailto:er.gauravgo...@gmail.com>> wrote: >> >>> >> >>> Thanks for your prompt >> >>> response! >> >>> >> >>> Situation is bit different now. Customer want us to remove the ceph >> >>> storage configuration from scratch. Let is openstack system work >> without >> >>> ceph. Later on install ceph with local disks. >> >>> >> >>> So I need to know a procedure to uninstall ceph and unconfigure it >> from >> >>> openstack. >> >>> >> >>> Regards >> >>> Gaurav Goyal >> >>> On 03-Aug-2016 4:59 pm, "David Turner" >> >>> <david.tur...@storagecraft.com<mailto:david.tur...@storagecraft.com>> >> wrote: >> >>> If I'm understanding your question correctly that you're asking how to >> >>> actually remove the SAN osds from ceph, then it doesn't matter what >> is using >> >>> the storage (ie openstack, cephfs, krbd, etc) as the steps are the >> same. >> >>> >> >>> I'm going to assume that you've already added the new storage/osds to >> the >> >>> cluster, weighted the SAN osds to 0.0 and that the backfilling has >> finished. >> >>> If that is true, then your disk used space on the SAN's should be >> basically >> >>> empty while the new osds on the local disks should have a fair amount >> of >> >>> data. If that is the case, then for every SAN osd, you just run the >> >>> following commands replacing OSD_ID with the osd's id: >> >>> >> >>> # On the server with the osd being removed >> >>> sudo stop ceph-osd id=OSD_ID >> >>> ceph osd down OSD_ID >> >>> ceph osd out OSD_ID >> >>> ceph osd crush remove osd.OSD_ID >> >>> ceph auth del osd.OSD_ID >> >>> ceph osd rm OSD_ID >> >>> >> >>> Test running those commands on a test osd and if you had set the >> weight >> >>> of the osd to 0.0 previously and if the backfilling had finished, >> then what >> >>> you should see is that your cluster has 1 less osd than it used to, >> and no >> >>> pgs should be backfilling. >> >>> >> >>> HOWEVER, if my assumptions above are incorrect, please provide the >> output >> >>> of the following commands and try to clarify your question. >> >>> >> >>> ceph status >> >>> ceph osd tree >> >>> >> >>> I hope this helps. >> >>> >> >>> > Hello David, >> >>> > >> >>> > Can you help me with steps/Procedure to uninstall Ceph storage from >> >>> > openstack environment? >> >>> > >> >>> > >> >>> > Regards >> >>> > Gaurav Goyal >> >>> ________________________________ >> >>> [cid:image001.jpg@01D1EE42.88EF6E60]<https://storagecraft.com> >> >>> >> >>> David Turner | Cloud Operations Engineer | StorageCraft Technology >> >>> Corporation<https://storagecraft.com> >> >>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >> >>> Office: 801.871.2760 | Mobile: 385.224.2943<tel:385.224.2943> >> >>> >> >>> ________________________________ >> >>> If you are not the intended recipient of this message or received it >> >>> erroneously, please notify the sender and delete it, together with any >> >>> attachments, and be advised that any dissemination or copying of this >> >>> message is prohibited. >> >>> >> >>> ________________________________ >> >>> >> >> >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com