solarflow99 <solarflo...@gmail.com>:
> now this goes against what I thought I learned about ceph fs.  You should be 
> able to RW to/from all OSDs, how can it be limited to only a single OSD??

Clients only connect to the primary OSD of a PG, so technically an OSD
that isn't the primary of any OSD doesn't have to be reachable by
clients.
That doesn't mean that's a good idea to configure a system like this ;)

Paul


>
> On Sat, Oct 6, 2018 at 4:30 AM Christopher Blum <b...@dorf.rwth-aachen.de> 
> wrote:
>>
>> I wouldn't recommend you pursuit this any further, but if this is the only 
>> client that would reside on the same VM as the OSD, one thing you could try 
>> is to decrease the primary affinity to 0 [1] for the local OSD .
>> That way that single OSD would never become a primary OSD ;)
>>
>> Disclaimer: This is more like a hack.
>>
>>
>> [1] https://ceph.com/geen-categorie/ceph-primary-affinity/
>>
>> On Fri, Oct 5, 2018 at 10:23 PM Gregory Farnum <gfar...@redhat.com> wrote:
>>>
>>> On Fri, Oct 5, 2018 at 3:13 AM Marc Roos <m.r...@f1-outsourcing.eu> wrote:
>>>>
>>>>
>>>>
>>>> I guess then this waiting "quietly" should be looked at again, I am
>>>> having load of 10 on this vm.
>>>>
>>>> [@~]# uptime
>>>>  11:51:58 up 4 days,  1:35,  1 user,  load average: 10.00, 10.01, 10.05
>>>>
>>>> [@~]# uname -a
>>>> Linux smb 3.10.0-862.11.6.el7.x86_64 #1 SMP Tue Aug 14 21:49:04 UTC 2018
>>>> x86_64 x86_64 x86_64 GNU/Linux
>>>>
>>>> [@~]# cat /etc/redhat-release
>>>> CentOS Linux release 7.5.1804 (Core)
>>>>
>>>> [@~]# dmesg
>>>> [348948.927734] libceph: osd23 192.168.10.114:6810 socket closed (con
>>>> state CONNECTING)
>>>> [348957.120090] libceph: osd27 192.168.10.114:6802 socket closed (con
>>>> state CONNECTING)
>>>> [349010.370171] libceph: osd26 192.168.10.114:6806 socket closed (con
>>>> state CONNECTING)
>>>> [349114.822301] libceph: osd24 192.168.10.114:6804 socket closed (con
>>>> state CONNECTING)
>>>> [349141.447330] libceph: osd29 192.168.10.114:6812 socket closed (con
>>>> state CONNECTING)
>>>> [349278.668658] libceph: osd25 192.168.10.114:6800 socket closed (con
>>>> state CONNECTING)
>>>> [349440.467038] libceph: osd28 192.168.10.114:6808 socket closed (con
>>>> state CONNECTING)
>>>> [349465.043957] libceph: osd23 192.168.10.114:6810 socket closed (con
>>>> state CONNECTING)
>>>> [349473.236400] libceph: osd27 192.168.10.114:6802 socket closed (con
>>>> state CONNECTING)
>>>> [349526.486408] libceph: osd26 192.168.10.114:6806 socket closed (con
>>>> state CONNECTING)
>>>> [349630.938498] libceph: osd24 192.168.10.114:6804 socket closed (con
>>>> state CONNECTING)
>>>> [349657.563561] libceph: osd29 192.168.10.114:6812 socket closed (con
>>>> state CONNECTING)
>>>> [349794.784936] libceph: osd25 192.168.10.114:6800 socket closed (con
>>>> state CONNECTING)
>>>> [349956.583300] libceph: osd28 192.168.10.114:6808 socket closed (con
>>>> state CONNECTING)
>>>> [349981.160225] libceph: osd23 192.168.10.114:6810 socket closed (con
>>>> state CONNECTING)
>>>> [349989.352510] libceph: osd27 192.168.10.114:6802 socket closed (con
>>>> state CONNECTING)
>>>
>>>
>>> Looks like in this case the client is spinning trying to establish the 
>>> network connections it expects to be available. There's not really much 
>>> else it can do — we expect and require full routing. The monitors are 
>>> telling the clients that the OSDs are up and available, and it is doing 
>>> data IO that requires them. So it tries to establish a connection, sees the 
>>> network fail, and tries again.
>>>
>>> Unfortunately the restricted-network use case you're playing with here is 
>>> just not supported by Ceph.
>>> -Greg
>>>
>>>>
>>>> ..
>>>> ..
>>>> ..
>>>>
>>>>
>>>>
>>>>
>>>> -----Original Message-----
>>>> From: John Spray [mailto:jsp...@redhat.com]
>>>> Sent: donderdag 27 september 2018 11:43
>>>> To: Marc Roos
>>>> Cc: ceph-users@lists.ceph.com
>>>> Subject: Re: [ceph-users] Cannot write to cephfs if some osd's are not
>>>> available on the client network
>>>>
>>>> On Thu, Sep 27, 2018 at 10:16 AM Marc Roos <m.r...@f1-outsourcing.eu>
>>>> wrote:
>>>> >
>>>> >
>>>> > I have a test cluster and on a osd node I put a vm. The vm is using a
>>>> > macvtap on the client network interface of the osd node. Making access
>>>>
>>>> > to local osd's impossible.
>>>> >
>>>> > the vm of course reports that it cannot access the local osd's. What I
>>>>
>>>> > am getting is:
>>>> >
>>>> > - I cannot reboot this vm normally, need to reset it.
>>>>
>>>> When linux tries to shut down cleanly, part of that is flushing buffers
>>>> from any mounted filesystem back to disk.  If you have a network
>>>> filesystem mounted, and the network is unavailable, that can cause the
>>>> process to block.  You can try forcibly unmounting before rebooting.
>>>>
>>>> > - vm is reporting very high load.
>>>>
>>>> The CPU load part is surprising -- in general Ceph clients should wait
>>>> quietly when blocked, rather than spinning.
>>>>
>>>> > I guess this should not be happening not? Because it should choose an
>>>> > other available osd of the 3x replicated pool and just write the data
>>>> > to that one?
>>>>
>>>> No -- writes always go through the primary OSD for the PG being written
>>>> to.  If an OSD goes down, then another OSD will become the primary.  In
>>>> your case, the primary OSD is not going down, it's just being cut off
>>>> from the client by the network, so the writes are blocking indefinitely.
>>>>
>>>> John
>>>>
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> > _______________________________________________
>>>> > ceph-users mailing list
>>>> > ceph-users@lists.ceph.com
>>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users@lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to