Hi Greg,

(I'm a colleague of Ana), Thank you for your reply


On 10/17/2017 11:57 PM, Gregory Farnum wrote:
>
>
> On Tue, Oct 17, 2017 at 9:51 AM Ana Aviles <a...@greenhost.nl 
> <mailto:a...@greenhost.nl>> wrote:
>
>     Hello all,
>
>     We had an inconsistent PG on our cluster. While performing PG repair
>     operation the OSD crashed. The OSD was not able to start again anymore,
>     and there was no hardware failure on the disk itself. This is the log 
> output
>
>     2017-10-17 17:48:55.771384 7f234930d700 -1 log_channel(cluster) log
>     [ERR] : 2.2fc repair 1 missing, 0 inconsistent objects
>     2017-10-17 17:48:55.771417 7f234930d700 -1 log_channel(cluster) log
>     [ERR] : 2.2fc repair 3 errors, 1 fixed
>     2017-10-17 17:48:56.047896 7f234930d700 -1
>     /build/ceph-12.2.1/src/osd/PrimaryLogPG.cc: In function 'virtual void
>     PrimaryLogPG::on_local_recover(const hobject_t&, const
>     ObjectRecoveryInfo&, ObjectContextRef, bool, ObjectStore::Transaction*)'
>     thread 7f234930d700 time 2017-10-17 17:48:55.924115
>     /build/ceph-12.2.1/src/osd/PrimaryLogPG.cc: 358: FAILED assert(p !=
>     recovery_info.ss.clone_snaps.end())
>
>
> Hmm. The OSD got a push op containing a snapshot it doesn't think should 
> exist. I also see that there's a comment "// hmm, should we warn?" on that 
> assert.


We catched also those log entries, which indeed point to a clone/snapshot 
problem:

 -9877> 2017-10-17 17:46:16.044077 7f234db16700 10 log_client  will send 
2017-10-17 17:46:13.367842 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
483 : cluster [ERR] 2.2fc shard 78 missing 
2:3f72b543:::rbd_data.332d5a836bcc485.000000000000fcf6:466a7
 -9876> 2017-10-17 17:46:16.044105 7f234db16700 10 log_client  will send 
2017-10-17 17:46:13.368026 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
484 : cluster [ERR] repair 2.2fc 
2:3f72b543:::rbd_data.332d5a836bcc485.000000000000fcf6:466a7 is an unexpected 
clone
 -9868> 2017-10-17 17:46:16.324112 7f2354b24700 10 log_client  logged 
2017-10-17 17:46:13.367842 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
483 : cluster [ERR] 2.2fc shard 78 missing 
2:3f72b543:::rbd_data.332d5a836bcc485.000000000000fcf6:466a7
 -9867> 2017-10-17 17:46:16.324128 7f2354b24700 10 log_client  logged 
2017-10-17 17:46:13.368026 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
484 : cluster [ERR] repair 2.2fc 
2:3f72b543:::rbd_data.332d5a836bcc485.000000000000fcf6:466a7 is an unexpected 
clone
   -36> 2017-10-17 17:48:55.771384 7f234930d700 -1 log_channel(cluster) log 
[ERR] : 2.2fc repair 1 missing, 0 inconsistent objects
   -35> 2017-10-17 17:48:55.771417 7f234930d700 -1 log_channel(cluster) log 
[ERR] : 2.2fc repair 3 errors, 1 fixed
    -4> 2017-10-17 17:48:56.046071 7f234db16700 10 log_client  will send 
2017-10-17 17:48:55.771390 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
485 : cluster [ERR] 2.2fc repair 1 missing, 0 inconsistent objects
    -3> 2017-10-17 17:48:56.046088 7f234db16700 10 log_client  will send 
2017-10-17 17:48:55.771419 osd.78 osd.78 [XXXX:XXXX:XXXX:XXXX::203]:6880/9116 
486 : cluster [ERR] 2.2fc repair 3 errors, 1 fixed

>
> Can you take a full log with "debug osd = 20" set, post it with 
> ceph-post-file, and create a ticket on tracker.ceph.com 
> <http://tracker.ceph.com>?

We will submit the ticket tomorrow (we are in CEST), We want to have more pair 
of eyes on it when we start the OSD again.

After this crash the OSD was marked as out by us. The cluster rebalanced 
itself, unfortunately, the same issue appear on another OSD (same pg), after 
several crashes of this OSD, the OSD came back up, but now with one PG down. I 
assume the cluster decided it 'finished' the ceph pg repair command and removed 
the 'repair' state, but now with a broken pg. If you have any hints on how we 
can get the PG online again, we would be very grateful, so we can work on that 
tomorrow.


Thanks,

Mart




Some general info about this cluster:

- all OSD runs the same version, also monitors are all 12.2.1 (ubuntu xenial)

- the cluster is a backup cluster and has min/size 1 and replication 2, so only 
2 copies.

- the cluster was recently upgraded from jewel to luminous (3 weeks ago)

- the cluster was recently upgraded from straw to straw2 (1 week ago)

- it was in HEALTH_OK till this happend.

- we use filestore only

- the cluster was installed with hammer originally. upgraded to infernalis, 
jewel and now luminous



health:
(noup/noout set on purpose while we trying to recover)


$ ceph -s
  cluster:
    id:     xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxxx
    health: HEALTH_WARN
            noup,noout flag(s) set
            Reduced data availability: 1 pg inactive, 1 pg down
            Degraded data redundancy: 2892/31621143 objects degraded (0.009%), 
2 pgs unclean, 1 pg degraded, 1 pg undersized
 
  services:
    mon: 3 daemons, quorum ds2-mon1,ds2-mon2,ds2-mon3
    mgr: ds2-mon1(active)
    osd: 93 osds: 92 up, 92 in; 1 remapped pgs
         flags noup,noout
    rgw: 1 daemon active
 
  data:
    pools:   13 pools, 1488 pgs
    objects: 15255k objects, 43485 GB
    usage:   119 TB used, 126 TB / 245 TB avail
    pgs:     0.067% pgs not active
             2892/31621143 objects degraded (0.009%)
             1483 active+clean
             2    active+clean+scrubbing+deep
             1    active+undersized+degraded+remapped+backfilling
             1    active+clean+scrubbing
             1    down
 
  io:
    client:   340 B/s rd, 14995 B/s wr, 1 op/s rd, 2 op/s wr
    recovery: 9567 kB/s, 2 objects/s
 


$ ceph health detail
HEALTH_WARN noup,noout flag(s) set; Reduced data availability: 1 pg inactive, 1 
pg down; Degraded data redundancy: 2774/31621143 objects degraded (0.009%), 2 
pgs unclean, 1 pg degraded, 1 pg undersized
OSDMAP_FLAGS noup,noout flag(s) set
PG_AVAILABILITY Reduced data availability: 1 pg inactive, 1 pg down
    pg 2.2fc is down, acting [69,93]
PG_DEGRADED Degraded data redundancy: 2774/31621143 objects degraded (0.009%), 
2 pgs unclean, 1 pg degraded, 1 pg undersized
    pg 2.1e9 is stuck undersized for 23741.295159, current state 
active+undersized+degraded+remapped+backfilling, last acting [41]
    pg 2.2fc is stuck unclean since forever, current state down, last acting 
[69,93]



>
> Are all your OSDs running that same version?
> -Greg
>  
>
>
>      ceph version 12.2.1 (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) luminous
>     (stable)
>      1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
>     const*)+0x102) [0x56236c8ff3f2]
>      2: (PrimaryLogPG::on_local_recover(hobject_t const&, ObjectRecoveryInfo
>     const&, std::shared_ptr<ObjectContext>, bool,
>     ObjectStore::Transaction*)+0xd63) [0x56236c476213]
>      3: (ReplicatedBackend::handle_pull_response(pg_shard_t, PushOp const&,
>     PullOp*, std::__cxx11::list<ReplicatedBackend::pull_complete_info,
>     std::allocator<ReplicatedBackend::pull_complete_info> >*,
>     ObjectStore::Transaction*)+0x693) [0x56236c60d4d3]
>      4:
>     
> (ReplicatedBackend::_do_pull_response(boost::intrusive_ptr<OpRequest>)+0x2b5)
>     [0x56236c60dd75]
>      5:
>     
> (ReplicatedBackend::_handle_message(boost::intrusive_ptr<OpRequest>)+0x20c)
>     [0x56236c61196c]
>      6: (PGBackend::handle_message(boost::intrusive_ptr<OpRequest>)+0x50)
>     [0x56236c521aa0]
>      7: (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
>     ThreadPool::TPHandle&)+0x55d) [0x56236c48662d]
>      8: (OSD::dequeue_op(boost::intrusive_ptr<PG>,
>     boost::intrusive_ptr<OpRequest>, ThreadPool::TPHandle&)+0x3a9)
>     [0x56236c3091a9]
>      9: (PGQueueable::RunVis::operator()(boost::intrusive_ptr<OpRequest>
>     const&)+0x57) [0x56236c5a2ae7]
>      10: (OSD::ShardedOpWQ::_process(unsigned int,
>     ceph::heartbeat_handle_d*)+0x130e) [0x56236c3307de]
>      11: (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884)
>     [0x56236c9041e4]
>      12: (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x56236c907220]
>      13: (()+0x76ba) [0x7f2366be96ba]
>      14: (clone()+0x6d) [0x7f2365c603dd]
>      NOTE: a copy of the executable, or `objdump -rdS <executable>` is
>     needed to interpret this.
>
>     Thanks!
>
>     Ana
>
>
>     _______________________________________________
>     ceph-users mailing list
>     ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Attachment: signature.asc
Description: OpenPGP digital signature

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to