Well, I knew it had all the correct information since earlier so gave it a shot 
:)

Anyway, I think it may be just a bad controller as well. New enterprise drives 
shouldn’t be giving read errors this early in deployment tbh.

Cheers,
Josef
> On 16 Feb 2015, at 17:37, Greg Farnum <gfar...@redhat.com> wrote:
> 
> Woah, major thread necromancy! :)
> 
> On Feb 13, 2015, at 3:03 PM, Josef Johansson <jo...@oderland.se 
> <mailto:jo...@oderland.se>> wrote:
>> 
>> Hi,
>> 
>> I skimmed the logs again, as we’ve had more of this kinda errors,
>> 
>> I saw a lot of lossy connections errors,
>> -2567> 2014-11-24 11:49:40.028755 7f6d49367700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.54:0/1011446 pipe(0x19321b80 sd=44 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x110d2b00).accept replacing existing (lossy) channel (new one lossy=1)
>> -2564> 2014-11-24 11:49:42.000463 7f6d51df1700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/1015676 pipe(0x22d6000 sd=204 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x16e218c0).accept replacing existing (lossy) channel (new one lossy=1)
>> -2563> 2014-11-24 11:49:47.704467 7f6d4d1a5700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/3029106 pipe(0x231f6780 sd=158 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x136bd1e0).accept replacing existing (lossy) channel (new one lossy=1)
>> -2562> 2014-11-24 11:49:48.180604 7f6d4cb9f700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/2027138 pipe(0x1657f180 sd=254 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x13273340).accept replacing existing (lossy) channel (new one lossy=1)
>> -2561> 2014-11-24 11:49:48.808604 7f6d4c498700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/2023529 pipe(0x12831900 sd=289 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x12401600).accept replacing existing (lossy) channel (new one lossy=1)
>> -2559> 2014-11-24 11:49:50.128379 7f6d4b88c700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.53:0/1023180 pipe(0x11cb2280 sd=309 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x1280a000).accept replacing existing (lossy) channel (new one lossy=1)
>> -2558> 2014-11-24 11:49:52.472867 7f6d425eb700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/3019692 pipe(0x18eb4a00 sd=311 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x10df6b00).accept replacing existing (lossy) channel (new one lossy=1)
>> -2556> 2014-11-24 11:49:55.100208 7f6d49e72700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/3021273 pipe(0x1bacf680 sd=353 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x164ae2c0).accept replacing existing (lossy) channel (new one lossy=1)
>> -2555> 2014-11-24 11:49:55.776568 7f6d49468700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/3024351 pipe(0x1bacea00 sd=20 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x1887ba20).accept replacing existing (lossy) channel (new one lossy=1)
>> -2554> 2014-11-24 11:49:57.704437 7f6d49165700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/1023529 pipe(0x1a32ac80 sd=213 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0xfe93b80).accept replacing existing (lossy) channel (new one lossy=1)
>> -2553> 2014-11-24 11:49:58.694246 7f6d47549700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/3017204 pipe(0x102e5b80 sd=370 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0xfb5a000).accept replacing existing (lossy) channel (new one lossy=1)
>> -2551> 2014-11-24 11:50:00.412242 7f6d4673b700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.52:0/3027138 pipe(0x1b83b400 sd=250 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x12922dc0).accept replacing existing (lossy) channel (new one lossy=1)
>> -2387> 2014-11-24 11:50:22.761490 7f6d44fa4700  0 -- 
>> 10.168.7.23:6840/4010217 >> 10.168.7.25:0/27131 pipe(0xfc60c80 sd=300 :6840 
>> s=0 pgs=0 cs=0 l=1 c=0x1241d080).accept replacing existing (lossy) channel 
>> (new one lossy=1)
>> -2300> 2014-11-24 11:50:31.366214 7f6d517eb700  0 -- 
>> 10.168.7.23:6840/4010217 >> 10.168.7.22:0/15549 pipe(0x193b3180 sd=214 :6840 
>> s=0 pgs=0 cs=0 l=1 c=0x10ebbe40).accept replacing existing (lossy) channel 
>> (new one lossy=1)
>> -2247> 2014-11-24 11:50:37.372934 7f6d4a276700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/1013890 pipe(0x25d4780 sd=112 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x10666580).accept replacing existing (lossy) channel (new one lossy=1)
>> -2246> 2014-11-24 11:50:37.738539 7f6d4f6ca700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/3026502 pipe(0x1338ea00 sd=230 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x123f11e0).accept replacing existing (lossy) channel (new one lossy=1)
>> -2245> 2014-11-24 11:50:38.390093 7f6d48c60700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/2026502 pipe(0x16ba7400 sd=276 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x7d4fb80).accept replacing existing (lossy) channel (new one lossy=1)
>> -2242> 2014-11-24 11:50:40.505458 7f6d3e43a700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.53:0/1012682 pipe(0x12a53180 sd=183 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x10537080).accept replacing existing (lossy) channel (new one lossy=1)
>> -2198> 2014-11-24 11:51:14.273025 7f6d44ea3700  0 -- 
>> 10.168.7.23:6865/5010217 >> 10.168.7.25:0/30755 pipe(0x162bb680 sd=327 :6865 
>> s=0 pgs=0 cs=0 l=1 c=0x16e21600).accept replacing existing (lossy) channel 
>> (new one lossy=1)
>> -1881> 2014-11-29 00:45:42.247394 7f6d5c155700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(949861 rbd_data.1c56a792eb141f2.0000000000006200 
>> [stat,write 2228224~12288] ondisk = 0) v4 remote, 10.168.7.54:0/1025735, 
>> failed lossy con, dropping message 0x1bc00400
>> -976> 2015-01-05 07:10:01.763055 7f6d5c155700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(11034565 
>> rbd_data.1cc69562eb141f2.00000000000003ce [stat,write 1925120~4096] ondisk = 
>> 0) v4 remote, 10.168.7.54:0/2007323, failed lossy con, dropping message 
>> 0x12989400
>> -855> 2015-01-10 22:01:36.589036 7f6d5b954700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(727627 rbd_data.1cc69413d1b58ba.0000000000000055 
>> [stat,write 2289664~4096] ondisk = 0) v4 remote, 10.168.7.54:0/1007323, 
>> failed lossy con, dropping message 0x24f68800
>> -819> 2015-01-12 05:25:06.229753 7f6d3646c700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.53:0/2019809 pipe(0x1f0e9680 sd=460 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x13090420).accept replacing existing (lossy) channel (new one lossy=1)
>> -818> 2015-01-12 05:25:06.581703 7f6d37534700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.53:0/1025252 pipe(0x1b67a780 sd=71 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x16311e40).accept replacing existing (lossy) channel (new one lossy=1)
>> -817> 2015-01-12 05:25:21.342998 7f6d41167700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.53:0/1025579 pipe(0x114e8000 sd=502 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x16310160).accept replacing existing (lossy) channel (new one lossy=1)
>> -808> 2015-01-12 16:01:35.783534 7f6d5b954700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(752034 rbd_data.1cc69413d1b58ba.0000000000000055 
>> [stat,write 2387968~8192] ondisk = 0) v4 remote, 10.168.7.54:0/1007323, 
>> failed lossy con, dropping message 0x1fde9a00
>> -515> 2015-01-25 18:44:23.303855 7f6d5b954700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(46402240 
>> rbd_data.4b8e9b3d1b58ba.0000000000000471 [read 1310720~4096] ondisk = 0) v4 
>> remote, 10.168.7.51:0/1017204, failed lossy con, dropping message 0x250bce00
>> -303> 2015-02-02 22:30:03.140599 7f6d5c155700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(17710313 
>> rbd_data.1cc69562eb141f2.00000000000003ce [stat,write 4145152~4096] ondisk = 
>> 0) v4 remote, 10.168.7.54:0/2007323, failed lossy con, dropping message 
>> 0x1c5d4200
>> -236> 2015-02-05 15:29:04.945660 7f6d3d357700  0 -- 10.168.7.23:6819/10217 
>> >> 10.168.7.51:0/1026961 pipe(0x1c63e780 sd=203 :6819 s=0 pgs=0 cs=0 l=1 
>> c=0x11dc8dc0).accept replacing existing (lossy) channel (new one lossy=1)
>>  -66> 2015-02-10 20:20:36.673969 7f6d5b954700  0 -- 10.168.7.23:6819/10217 
>> submit_message osd_op_reply(11088 rbd_data.10b8c82eb141f2.0000000000004459 
>> [stat,write 749568~8192] ondisk = 0) v4 remote, 10.168.7.55:0/1005630, 
>> failed lossy con, dropping message 0x138db200
>> 
>> Could this have lead to the data being erroneous, or is the -5 return code 
>> just a sign of a broken hard drive?
>> 
> 
> These are the OSDs creating new connections to each other because the 
> previous ones failed. That's not necessarily a problem (although here it's 
> probably a symptom of some kind of issue, given the frequency) and cannot 
> introduce data corruption of any kind.
> I’m not seeing any -5 return codes as part of that messenger debug output, so 
> unless you were referring to your EIO from last June I’m not sure what that’s 
> about? (If you do mean EIOs, yes, they’re still a sign of a broken hard drive 
> or local FS.)
> 
>> Cheers,
>> Josef
>> 
>>> On 14 Jun 2014, at 02:38, Josef Johansson <jo...@oderland.se> wrote:
>>> 
>>> Thanks for the quick response.
>>> 
>>> Cheers,
>>> Josef
>>> 
>>> Gregory Farnum skrev 2014-06-14 02:36:
>>>> On Fri, Jun 13, 2014 at 5:25 PM, Josef Johansson <jo...@oderland.se> wrote:
>>>>> Hi Greg,
>>>>> 
>>>>> Thanks for the clarification. I believe the OSD was in the middle of a 
>>>>> deep
>>>>> scrub (sorry for not mentioning this straight away), so then it could've
>>>>> been a silent error that got wind during scrub?
>>>> Yeah.
>>>> 
>>>>> What's best practice when the store is corrupted like this?
>>>> Remove the OSD from the cluster, and either reformat the disk or
>>>> replace as you judge appropriate.
>>>> -Greg
>>>> 
>>>>> Cheers,
>>>>> Josef
>>>>> 
>>>>> Gregory Farnum skrev 2014-06-14 02:21:
>>>>> 
>>>>>> The OSD did a read off of the local filesystem and it got back the EIO
>>>>>> error code. That means the store got corrupted or something, so it
>>>>>> killed itself to avoid spreading bad data to the rest of the cluster.
>>>>>> -Greg
>>>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>>>> 
>>>>>> 
>>>>>> On Fri, Jun 13, 2014 at 5:16 PM, Josef Johansson <jo...@oderland.se>
>>>>>> wrote:
>>>>>>> Hey,
>>>>>>> 
>>>>>>> Just examing what happened to an OSD, that was just turned off. Data has
>>>>>>> been moved away from it, so hesitating to turned it back on.
>>>>>>> 
>>>>>>> Got the below in the logs, any clues to what the assert talks about?
>>>>>>> 
>>>>>>> Cheers,
>>>>>>> Josef
>>>>>>> 
>>>>>>> -1 os/FileStore.cc: In function 'virtual int FileStore::read(coll_t,
>>>>>>> const
>>>>>>> hobject_t&, uint64_t, size_t, ceph::bufferlist&, bool)' thread 7fdacb88
>>>>>>> c700 time 2014-06-11 21:13:54.036982
>>>>>>> os/FileStore.cc: 2992: FAILED assert(allow_eio || !m_filestore_fail_eio
>>>>>>> ||
>>>>>>> got != -5)
>>>>>>> 
>>>>>>> ceph version 0.67.7 (d7ab4244396b57aac8b7e80812115bbd079e6b73)
>>>>>>> 1: (FileStore::read(coll_t, hobject_t const&, unsigned long, unsigned
>>>>>>> long,
>>>>>>> ceph::buffer::list&, bool)+0x653) [0x8ab6c3]
>>>>>>> 2: (ReplicatedPG::do_osd_ops(ReplicatedPG::OpContext*,
>>>>>>> std::vector<OSDOp,
>>>>>>> std::allocator<OSDOp> >&)+0x350) [0x708230]
>>>>>>> 3: (ReplicatedPG::prepare_transaction(ReplicatedPG::OpContext*)+0x86)
>>>>>>> [0x713366]
>>>>>>> 4: (ReplicatedPG::do_op(std::tr1::shared_ptr<OpRequest>)+0x3095)
>>>>>>> [0x71acb5]
>>>>>>> 5: (PG::do_request(std::tr1::shared_ptr<OpRequest>,
>>>>>>> ThreadPool::TPHandle&)+0x3f0) [0x812340]
>>>>>>> 6: (OSD::dequeue_op(boost::intrusive_ptr<PG>,
>>>>>>> std::tr1::shared_ptr<OpRequest>, ThreadPool::TPHandle&)+0x2ea) 
>>>>>>> [0x75c80a]
>>>>>>> 7: (OSD::OpWQ::_process(boost::intrusive_ptr<PG>,
>>>>>>> ThreadPool::TPHandle&)+0x198) [0x770da8]
>>>>>>> 8: (ThreadPool::WorkQueueVal<std::pair<boost::intrusive_ptr<PG>,
>>>>>>> std::tr1::shared_ptr<OpRequest> >, boost::intrusive_ptr<PG>
>>>>>>>> ::_void_process(void*, ThreadPool::TPHandle&)+0xae) [0x7a89
>>>>>>> ce]
>>>>>>> 9: (ThreadPool::worker(ThreadPool::WorkThread*)+0x68a) [0x9b5dea]
>>>>>>> 10: (ThreadPool::WorkThread::entry()+0x10) [0x9b7040]
>>>>>>> 11: (()+0x6b50) [0x7fdadffdfb50]
>>>>>>> 12: (clone()+0x6d) [0x7fdade53b0ed]
>>>>>>> NOTE: a copy of the executable, or `objdump -rdS <executable>` is
>>>>>>> needed to
>>>>>>> interpret this.
>>>>>>> 
>>>>>>> 
>>>>>>> _______________________________________________
>>>>>>> ceph-users mailing list
>>>>>>> ceph-users@lists.ceph.com
>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> 
>>> 
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to