Hey Lincoln


On Tue, Sep 8, 2015 at 7:26 PM, Lincoln Bryant <linco...@uchicago.edu>
wrote:

> For whatever it’s worth, my problem has returned and is very similar to
> yours. Still trying to figure out what’s going on over here.
>
> Performance is nice for a few seconds, then goes to 0. This is a similar
> setup to yours (12 OSDs per box, Scientific Linux 6, Ceph 0.94.3, etc)
>
>   384      16     29520     29504   307.287      1188 0.0492006  0.208259
>   385      16     29813     29797   309.532      1172 0.0469708  0.206731
>   386      16     30105     30089   311.756      1168 0.0375764  0.205189
>   387      16     30401     30385   314.009      1184  0.036142  0.203791
>   388      16     30695     30679   316.231      1176 0.0372316  0.202355
>   389      16     30987     30971    318.42      1168 0.0660476  0.200962
>   390      16     31282     31266   320.628      1180 0.0358611  0.199548
>   391      16     31568     31552   322.734      1144 0.0405166  0.198132
>   392      16     31857     31841   324.859      1156 0.0360826  0.196679
>   393      16     32090     32074   326.404       932 0.0416869   0.19549
>   394      16     32205     32189   326.743       460 0.0251877  0.194896
>   395      16     32302     32286   326.897       388 0.0280574  0.194395
>   396      16     32348     32332   326.537       184 0.0256821  0.194157
>   397      16     32385     32369   326.087       148 0.0254342  0.193965
>   398      16     32424     32408   325.659       156 0.0263006  0.193763
>   399      16     32445     32429   325.054        84 0.0233839  0.193655
> 2015-09-08 11:22:31.940164 min lat: 0.0165045 max lat: 67.6184 avg lat:
> 0.193655
>   sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg lat
>   400      16     32445     32429   324.241         0         -  0.193655
>   401      16     32445     32429   323.433         0         -  0.193655
>   402      16     32445     32429   322.628         0         -  0.193655
>   403      16     32445     32429   321.828         0         -  0.193655
>   404      16     32445     32429   321.031         0         -  0.193655
>   405      16     32445     32429   320.238         0         -  0.193655
>   406      16     32445     32429    319.45         0         -  0.193655
>   407      16     32445     32429   318.665         0         -  0.193655
>
> needless to say, very strange.
>

Its indeed very strange

( The solution that you gave me in the below email ) Have you tried
restarting all OSD's ?

By the way my problem got fixed ( but i am afraid , it can come back any
time ) by doing

# service ceph restart osd  on all OSD nodes ( this didn't helped )
# set noout,nodown,nobackfill,norecover and then reboot all OSD nodes ( It
worked )  After they all the rados bench write started to work.

[ i know its hilarious , feels like  i am watching *The IT Crowd* ' Hello
IT , Have you tried turning it OFF and ON again ' ]

It would be really helpful if someone provides a real solution.




>
> —Lincoln
>
>
> > On Sep 7, 2015, at 3:35 PM, Vickey Singh <vickey.singh22...@gmail.com>
> wrote:
> >
> > Adding ceph-users.
> >
> > On Mon, Sep 7, 2015 at 11:31 PM, Vickey Singh <
> vickey.singh22...@gmail.com> wrote:
> >
> >
> > On Mon, Sep 7, 2015 at 10:04 PM, Udo Lembke <ulem...@polarzone.de>
> wrote:
> > Hi Vickey,
> > Thanks for your time in replying to my problem.
> >
> > I had the same rados bench output after changing the motherboard of the
> monitor node with the lowest IP...
> > Due to the new mainboard, I assume the hw-clock was wrong during
> startup. Ceph health show no errors, but all VMs aren't able to do IO (very
> high load on the VMs - but no traffic).
> > I stopped the mon, but this don't changed anything. I had to restart all
> other mons to get IO again. After that I started the first mon also (with
> the right time now) and all worked fine again...
> >
> > Thanks i will try to restart all OSD / MONS and report back , if it
> solves my problem
> >
> > Another posibility:
> > Do you use journal on SSDs? Perhaps the SSDs can't write to garbage
> collection?
> >
> > No i don't have journals on SSD , they are on the same OSD disk.
> >
> >
> >
> > Udo
> >
> >
> > On 07.09.2015 16:36, Vickey Singh wrote:
> >> Dear Experts
> >>
> >> Can someone please help me , why my cluster is not able write data.
> >>
> >> See the below output  cur MB/S  is 0  and Avg MB/s is decreasing.
> >>
> >>
> >> Ceph Hammer  0.94.2
> >> CentOS 6 (3.10.69-1)
> >>
> >> The Ceph status says OPS are blocked , i have tried checking , what all
> i know
> >>
> >> - System resources ( CPU , net, disk , memory )    -- All normal
> >> - 10G network for public and cluster network  -- no saturation
> >> - Add disks are physically healthy
> >> - No messages in /var/log/messages OR dmesg
> >> - Tried restarting OSD which are blocking operation , but no luck
> >> - Tried writing through RBD  and Rados bench , both are giving same
> problemm
> >>
> >> Please help me to fix this problem.
> >>
> >> #  rados bench -p rbd 60 write
> >>  Maintaining 16 concurrent writes of 4194304 bytes for up to 60 seconds
> or 0 objects
> >>  Object prefix: benchmark_data_stor1_1791844
> >>    sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg
> lat
> >>      0       0         0         0         0         0         -
>  0
> >>      1      16       125       109   435.873       436  0.022076
> 0.0697864
> >>      2      16       139       123   245.948        56  0.246578
> 0.0674407
> >>      3      16       139       123   163.969         0         -
> 0.0674407
> >>      4      16       139       123   122.978         0         -
> 0.0674407
> >>      5      16       139       123    98.383         0         -
> 0.0674407
> >>      6      16       139       123   81.9865         0         -
> 0.0674407
> >>      7      16       139       123   70.2747         0         -
> 0.0674407
> >>      8      16       139       123   61.4903         0         -
> 0.0674407
> >>      9      16       139       123   54.6582         0         -
> 0.0674407
> >>     10      16       139       123   49.1924         0         -
> 0.0674407
> >>     11      16       139       123   44.7201         0         -
> 0.0674407
> >>     12      16       139       123   40.9934         0         -
> 0.0674407
> >>     13      16       139       123   37.8401         0         -
> 0.0674407
> >>     14      16       139       123   35.1373         0         -
> 0.0674407
> >>     15      16       139       123   32.7949         0         -
> 0.0674407
> >>     16      16       139       123   30.7451         0         -
> 0.0674407
> >>     17      16       139       123   28.9364         0         -
> 0.0674407
> >>     18      16       139       123   27.3289         0         -
> 0.0674407
> >>     19      16       139       123   25.8905         0         -
> 0.0674407
> >> 2015-09-07 15:54:52.694071min lat: 0.022076 max lat: 0.46117 avg lat:
> 0.0674407
> >>    sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg
> lat
> >>     20      16       139       123    24.596         0         -
> 0.0674407
> >>     21      16       139       123   23.4247         0         -
> 0.0674407
> >>     22      16       139       123     22.36         0         -
> 0.0674407
> >>     23      16       139       123   21.3878         0         -
> 0.0674407
> >>     24      16       139       123   20.4966         0         -
> 0.0674407
> >>     25      16       139       123   19.6768         0         -
> 0.0674407
> >>     26      16       139       123     18.92         0         -
> 0.0674407
> >>     27      16       139       123   18.2192         0         -
> 0.0674407
> >>     28      16       139       123   17.5686         0         -
> 0.0674407
> >>     29      16       139       123   16.9628         0         -
> 0.0674407
> >>     30      16       139       123   16.3973         0         -
> 0.0674407
> >>     31      16       139       123   15.8684         0         -
> 0.0674407
> >>     32      16       139       123   15.3725         0         -
> 0.0674407
> >>     33      16       139       123   14.9067         0         -
> 0.0674407
> >>     34      16       139       123   14.4683         0         -
> 0.0674407
> >>     35      16       139       123   14.0549         0         -
> 0.0674407
> >>     36      16       139       123   13.6645         0         -
> 0.0674407
> >>     37      16       139       123   13.2952         0         -
> 0.0674407
> >>     38      16       139       123   12.9453         0         -
> 0.0674407
> >>     39      16       139       123   12.6134         0         -
> 0.0674407
> >> 2015-09-07 15:55:12.697124min lat: 0.022076 max lat: 0.46117 avg lat:
> 0.0674407
> >>    sec Cur ops   started  finished  avg MB/s  cur MB/s  last lat   avg
> lat
> >>     40      16       139       123   12.2981         0         -
> 0.0674407
> >>     41      16       139       123   11.9981         0         -
> 0.0674407
> >>
> >>
> >>
> >>
> >>     cluster 86edf8b8-b353-49f1-ab0a-a4827a9ea5e8
> >>      health HEALTH_WARN
> >>             1 requests are blocked > 32 sec
> >>      monmap e3: 3 mons at {stor0111=
> 10.100.1.111:6789/0,stor0113=10.100.1.113:6789/0,stor011
> >> 5=10.100.1.115:6789/0}
> >>             election epoch 32, quorum 0,1,2 stor0111,stor0113,stor0115
> >>      osdmap e19536: 50 osds: 50 up, 50 in
> >>       pgmap v928610: 2752 pgs, 9 pools, 30476 GB data, 4183 kobjects
> >>             91513 GB used, 47642 GB / 135 TB avail
> >>                 2752 active+clean
> >>
> >>
> >> Tried using RBD
> >>
> >>
> >> # dd if=/dev/zero of=file1 bs=4K count=10000 oflag=direct
> >> 10000+0 records in
> >> 10000+0 records out
> >> 40960000 bytes (41 MB) copied, 24.5529 s, 1.7 MB/s
> >>
> >> # dd if=/dev/zero of=file1 bs=1M count=100 oflag=direct
> >> 100+0 records in
> >> 100+0 records out
> >> 104857600 bytes (105 MB) copied, 1.05602 s, 9.3 MB/s
> >>
> >> # dd if=/dev/zero of=file1 bs=1G count=1 oflag=direct
> >> 1+0 records in
> >> 1+0 records out
> >> 1073741824 bytes (1.1 GB) copied, 293.551 s, 3.7 MB/s
> >> ]#
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >>
> >> _______________________________________________
> >> ceph-users mailing list
> >>
> >> ceph-users@lists.ceph.com
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to