Hmm.. I hope I don't really need any thing from osd.0. =P
# ceph-objectstore-tool --op export --pgid 2.35 --data-path 
/var/lib/ceph/osd/ceph-0 --journal-path /var/lib/ceph/osd/ceph-0/journal --file 
2.35.exportFailure to read OSD superblock: (2) No such file or directory# 
ceph-objectstore-tool --op export --pgid 2.2f --data-path 
/var/lib/ceph/osd/ceph-0 --journal-path /var/lib/ceph/osd/ceph-0/journal --file 
2.2f.exportFailure to read OSD superblock: (2) No such file or directory
Regards,Hong 

    On Monday, September 4, 2017 2:29 AM, hjcho616 <hjcho...@yahoo.com> wrote:
 

 Ronny,
While letting cluster replicate, looks like this might be a while, I decided to 
look in to where those pgs are missing.. From the "ceph health detail" I found 
pgs that are unfound.  Then found the directories that had that pgs, pasted on 
the right of that detail message below..pg 2.35 is active+recovering+degraded, 
acting [11,6], 29 unfound, ceph-0/current/2.35_head, ceph-8/current/2.35_headpg 
2.2f is active+recovery_wait+degraded+remapped, acting [9,10], 24 unfound, 
ceph-0/current/2.2f_head, ceph-2/current/2.2f_headpg 2.27 is 
active+recovery_wait+degraded, acting [11,6], 19 unfound, 
ceph-4/current/2.27_headpg 2.1a is active+recovery_wait+degraded, acting 
[2,11], 29 unfound, ceph-0/current/2.1a_head, ceph-3/current/2.1a_head 
ceph-4/current/2.1a_headpg 2.1f is active+recovery_wait+degraded, acting 
[10,2], 20 unfound, ceph-3/current/2.1f_head, ceph-4/current/2.1f_headpg 2.3c 
is active+recovery_wait+degraded, acting [10,2], 26 unfound, 
ceph-0/current/2.3c_head, ceph-4/current/2.3c_head
Basically, I just went to look at the pg directories with rb.* files in it.  I 
noticed that there are more than 1 of those directories throughout osds.  
Should it matter which one of them I export?  Or do I need both? I am seeing 
all of them can be found outside of ceph-0, I'll probably grab from non ceph-0 
OSDs if I can grab from any.  
One thing strange I notice is... pg 2.2f, that one has rb.* files in active 
node ceph-2 but still marked unfound?  Maybe that means I need to export both 
and import both?  If I have to get both, is there a need to merge the two 
before importing?  Or would the tool know how to handle this?
Regards,Hong 

    On Monday, September 4, 2017 1:20 AM, hjcho616 <hjcho...@yahoo.com> wrote:
 

 Thank you Ronny.  I've added two OSDs to OSD2, 2TB each.  I hope that would be 
enough. =)  I've changed min_size and size to 2.  OSDs are busy balancing 
again.  I'll try those you recommended and will get back to you with more 
questions! =) 
# ceph osd treeID WEIGHT   TYPE NAME      UP/DOWN REWEIGHT PRIMARY-AFFINITY-1 
19.87198 root default-2  8.12239     host OSD1 1  1.95250         osd.1       
up  1.00000          1.00000 0  1.95250         osd.0     down        0         
 1.00000 7  0.31239         osd.7       up  1.00000          1.00000 6  1.95250 
        osd.6       up  1.00000          1.00000 2  1.95250         osd.2       
up  1.00000          1.00000-3 11.74959     host OSD2 3  1.95250         osd.3  
   down        0          1.00000 4  1.95250         osd.4     down        0    
      1.00000 5  1.95250         osd.5     down        0          1.00000 8  
1.95250         osd.8     down        0          1.00000 9  0.31239         
osd.9       up  1.00000          1.0000010  1.81360         osd.10      up  
1.00000          1.0000011  1.81360         osd.11      up  1.00000          
1.00000
Regards,Hong 

    On Sunday, September 3, 2017 6:56 AM, Ronny Aasen 
<ronny+ceph-us...@aasen.cx> wrote:
 

  I would not even attempt to connect a recovered drive to ceph, especially not 
one that have had xfs errors and corruption.  
 
 your pg's that are undersized lead me to belive you still need to either 
expand, with more disks, or nodes. or that you need to set 
 osd crush chooseleaf type = 0 
 to let ceph pick 2 disks on the same node as a valid object placement.  
(temporary until you get 2 balanced nodes) generally let ceph selfheal as much 
as possible (no misplaced or degraded objects)  this require that ceph have 
space for the recovery. 
 i would run with size=2 min_size=2  
 
 you should also look at the 7 shrub errors. they indicate that there can be 
other drives with issues, you want to locate where those inconsistent objects 
are, and fix them. read this page about fixing scrub errors. 
http://ceph.com/geen-categorie/ceph-manually-repair-object/
 
 then you would sit with the 103 unfound objects, and those you should try to 
recover from the recovered drive. 
 by using the ceph-objectstore-tool export/import  to try and export pg's 
missing objects  to a dedicated temporary added import drive.
 the import drive does not need to be very large. since you can do one and one 
pg at the time. and you should only recover pg's that contain unfound objects. 
there is realy only 103 unfound objects that you need to recover. 
 once the recovery is compleate you can wipe the functioning recovery drive, 
and install it as a new osd to the cluster.
 
 
 
 kind regards
 Ronny Aasen
 
 
 On 03.09.2017 06:20, hjcho616 wrote:
  
  I checked with ceph-2, 3, 4, 5 so I figured it was safe to assume that 
superblock file is the same.  I copied it over and started OSD.  It still fails 
with the same error message.  Looks like when I updated to 10.2.9, some osd 
needs to be updated and that process is not finding the data it needs?  What 
can I do about this situation? 
  2017-09-01 22:27:35.590041 7f68837e5800  1 
filestore(/var/lib/ceph/osd/ceph-0) upgrade 2017-09-01 22:27:35.590149 
7f68837e5800 -1 filestore(/var/lib/ceph/osd/ceph-0) could not find 
#-1:7b3f43c4:::osd_superblock:0# in index: (2) No such file or directory 
  Regards, Hong 
 
      On Friday, September 1, 2017 11:10 PM, hjcho616 <hjcho...@yahoo.com> 
wrote:
  
 
     Just realized there is a file called superblock in the ceph directory.  
ceph-1 and ceph-2's superblock file is identical, ceph-6 and ceph-7 are 
identical, but not between the two groups.   When I originally created the 
OSDs, I created ceph-0 through 5.  Can superblock file be copied over from 
ceph-1 to ceph-0? 
  Hmm.. it appears to be doing something in the background even though osd.0 is 
down.  ceph health output is changing! # ceph health HEALTH_ERR 40 pgs are 
stuck inactive for more than 300 seconds; 14 pgs backfill_wait; 21 pgs 
degraded; 10 pgs down; 2 pgs inconsistent; 10 pgs peering; 3 pgs recovering; 2 
pgs recovery_wait; 30 pgs stale; 21 pgs stuck degraded; 10 pgs stuck inactive; 
30 pgs stuck stale; 45 pgs stuck unclean; 16 pgs stuck undersized; 16 pgs 
undersized; 2 requests are blocked > 32 sec; recovery 221826/2473662 objects 
degraded (8.968%); recovery 254711/2473662 objects misplaced (10.297%); 
recovery 103/2251966 unfound (0.005%); 7 scrub errors; mds cluster is degraded; 
no legacy OSD present but  'sortbitwise' flag is not set 
  Regards, Hong 
 
       On Friday, September 1, 2017 10:37 PM, hjcho616 <hjcho...@yahoo.com> 
wrote:
  
 
     Tried connecting recovered osd.  Looks like some of the files in the 
lost+found are super blocks.   Below is the log.  What can I do about this? 
  2017-09-01 22:27:27.634228 7f68837e5800  0 set uid:gid to 1001:1001 
(ceph:ceph) 2017-09-01 22:27:27.634245 7f68837e5800  0 ceph version 10.2.9 
(2ee413f77150c0f375ff6f10edd6c8f9c7d060d0),  process ceph-osd, pid 5432 
2017-09-01 22:27:27.635456 7f68837e5800  0 pidfile_write: ignore empty 
--pid-file 2017-09-01 22:27:27.646849 7f68837e5800  
0filestore(/var/lib/ceph/osd/ceph-0) backend xfs (magic 0x58465342) 2017-09-01 
22:27:27.647077 7f68837e5800  
0genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: FIEMAP 
ioctl is disabled via 'filestore fiemap' config option 2017-09-01 
22:27:27.647080 7f68837e5800  
0genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: 
SEEK_DATA/SEEK_HOLE is disabled  via 'filestore seek data hole' config option 
2017-09-01 22:27:27.647091 7f68837e5800  
0genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: splice is 
supported 2017-09-01 22:27:27.678937 7f68837e5800  
0genericfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_features: syncfs(2) 
syscall fully supported (by glibc and kernel) 2017-09-01 22:27:27.679044 
7f68837e5800  0xfsfilestorebackend(/var/lib/ceph/osd/ceph-0) detect_feature: 
extsize is disabled by conf 2017-09-01 22:27:27.680718 7f68837e5800  1 leveldb: 
Recovering log #28054 2017-09-01 22:27:27.804501 7f68837e5800  1 leveldb: 
Delete type=0 #28054 
  2017-09-01 22:27:27.804579 7f68837e5800  1 leveldb: Delete type=3 #28053 
  2017-09-01 22:27:35.586725 7f68837e5800  0filestore(/var/lib/ceph/osd/ceph-0) 
mount: enabling WRITEAHEAD journal mode: checkpoint is not enabled 2017-09-01 
22:27:35.587689 7f68837e5800  1 journal _open /var/lib/ceph/osd/ceph-0/journal 
fd 18: 9998729216 bytes, block size 4096 bytes, directio = 1, aio = 1 
2017-09-01 22:27:35.589631 7f68837e5800  1 journal _open 
/var/lib/ceph/osd/ceph-0/journal fd 18: 9998729216 bytes, block size 4096 
bytes, directio = 1, aio = 1 2017-09-01 22:27:35.590041 7f68837e5800  
1filestore(/var/lib/ceph/osd/ceph-0) upgrade 2017-09-01 22:27:35.590149 
7f68837e5800 -1filestore(/var/lib/ceph/osd/ceph-0) could not find 
#-1:7b3f43c4:::osd_superblock:0# in index: (2) No such file or directory 
2017-09-01 22:27:35.590158 7f68837e5800 -1 osd.0 0 OSD::init() : unable to  
read osd superblock 2017-09-01 22:27:35.590547 7f68837e5800  1 journal close 
/var/lib/ceph/osd/ceph-0/journal 2017-09-01 22:27:35.611595 7f68837e5800 -1 
^[[0;31m ** ERROR: osd init  failed: (22) Invalid argument^[[0m 
  Recovered drive is mounted on /var/lib/ceph/osd/ceph-0. # df Filesystem      
1K-blocks      Used  Available Use% Mounted on udev                10240        
 0      10240   0% /dev tmpfs             1584780      9172    1575608   1% 
/run /dev/sda1        15247760   9319048    5131120  65% / tmpfs             
3961940         0    3961940   0% /dev/shm tmpfs                5120         0  
     5120   0% /run/lock tmpfs             3961940         0    3961940   0% 
/sys/fs/cgroup /dev/sdb1      1952559676 634913968 1317645708  33% 
/var/lib/ceph/osd/ceph-0 /dev/sde1      1952559676 640365952 1312193724  33% 
/var/lib/ceph/osd/ceph-6 /dev/sdd1      1952559676 712018768 1240540908  37% 
/var/lib/ceph/osd/ceph-2 /dev/sdc1      1952559676 755827440 1196732236  39% 
/var/lib/ceph/osd/ceph-1 /dev/sdf1       312417560  42538060  269879500  14% 
/var/lib/ceph/osd/ceph-7 tmpfs              792392         0     792392   0% 
/run/user/0 # cd /var/lib/ceph/osd/ceph-0 # ls activate.monmap  current  
journal_uuid  magic          superblock  whoami active           fsid     
keyring       ready          sysvinit ceph_fsid        journal  lost+found    
store_version  type 
  Regards, Hong 
 
       On Friday, September 1, 2017 2:59 PM,  hjcho616 <hjcho...@yahoo.com> 
wrote:
  
 
     Found the partition, wasn't able to  mount the partition right away... Did 
a xfs_repair on  that drive.   
  Got bunch of messages like this.. =( 
entry"100000a89fd.00000000__head_AE319A25__0" in shortform directory 845908970  
references non-existent inode 605294241                 junking entry 
"100000a89fd.00000000__head_AE319A25__0"  in directory inode 845908970          
   
  Was able to mount.  lost+found has lots of files there. =P   Running du seems 
to show OK files in current  directory. 
  Will it be safe to attach this one  back to the cluster?  Is there a way to 
specify to use  this drive if the data is missing? =)  Or am I being paranoid?  
Just plug it? =) 
  Regards, Hong 
 
       On Friday, September 1,  2017 9:01 AM, hjcho616 <hjcho...@yahoo.com> 
wrote:
  
 
     Looks like it has been  rescued... Only 1 error as we saw  before in the 
smart log! # ddrescue -f /dev/sda  /dev/sdc ./rescue.log GNU ddrescue 1.21 
Press Ctrl-C to interrupt      ipos:    1508 GB, non-trimmed:        0 B,  
current rate:       0 B/s      opos:    1508 GB, non-scraped:        0 B,  
average rate:  88985 kB/s non-tried:        0 B,     errsize:     4096 B,      
run time:  6h 14m 40s   rescued:    2000 GB,      errors:        1,  remaining 
time:         n/a percent rescued:  99.99%      time since last successful  
read:         39s Finished                        
  Still missing partition in the new drive.  =P  I found this util called  
testdisk for broken partition  tables.  Will try that tonight. =P 
  Regards, Hong 
  
 
       On Wednesday, August 30,  2017 9:18 AM, Ronny Aasen 
<ronny+ceph-us...@aasen.cx> wrote:
  
 
    On 30.08.2017 15:32, Steve  Taylor wrote:
  
 
   I'm not familiar with dd_rescue, but  I've just been reading about it. I'm 
not seeing any features that  would be beneficial in this scenario that aren't 
also available  in dd. What specific features give  it "really a far better  
chance of restoring a copy of your disk" than dd? I'm always  interested in 
learning about new recovery  tools. 
 i see i wrote dd_rescue from  old habit, but the package one should use on 
debian is gddrescue or  also called gnu ddrecue. 
 
 this page have some details  on the differences on dd vs the  ddrescue 
variants. 
 http://www.toad.com/gnu/sysadmin/index.html#ddrescue
 
 kind regards
 Ronny Aasen
 
 
 
     
|    |  Steve Taylor | Senior Software Engineer | StorageCraft Technology 
Corporation
 380 Data Drive Suite 300 | Draper | Utah | 84020
 Office: 801.871.2799 |   |

  
| If you are not the intended recipient of this message or received it  
erroneously, please notify the sender and  delete it, together with  any 
attachments, and be advised  that any dissemination or copying of this message 
is prohibited. |

  
 On Tue, 2017-08-29 at  21:49 +0200, Willem Jan Withagen  wrote: 
 On 29-8-2017 19:12, Steve Taylor wrote:
 
 Hong, Probably your  best chance at recovering any data without  special, 
expensive,  forensic procedures is to perform a  dd from /dev/sdb to somewhere 
else large enough to hold a full  disk image and attempt to repair that. You'll 
want to use  'conv=noerror' with your dd command  since your disk is  failing. 
Then you could either  re-attach the OSD from the new  source or attempt to  
retrieve objects from the filestore  on it. 
 Like somebody else already  pointed out In problem "cases like  disk, use 
dd_rescue.  It has really a far better chance of  restoring a copy of your disk 
--WjW 
 I have actually done  this before by creating an RBD that  matches the disk 
size,  performing the dd, running xfs_repair,  and eventually adding it back  
to the cluster as an OSD. RBDs as OSDs  is certainly a temporary arrangement 
for repair only, but I'm  happy to report that it worked  flawlessly in my 
case. I was  able to weight the OSD to 0, offload all of  its data, then remove 
it for  a full recovery, at which  point I just deleted the  RBD. The 
possibilities  afforded by Ceph inception are endless. ☺  Steve Taylor | Senior 
 Software Engineer | StorageCraft  Technology Corporation 380 Data Drive Suite 
300 | Draper | Utah | 84020 Office:  801.871.2799 | If you are not the intended 
recipient of this message  or received it erroneously, please notify the sender 
and delete it,  together with any attachments,  and be advised that any  
dissemination or copying of this message  is prohibited. On Mon,  2017-08-28 at 
23:17 +0100, Tomasz  Kusmierz wrote: 
 Rule of thumb with batteries  is: - more “proper temperature”  you run them at 
the more  life you get out of them  - more battery is overpowered for your 
application the longer it  will survive.  Get your self a LSI 94**  controller 
and use it as HBA and you will  be fine. but get  MORE DRIVES !!!!! …  
 On 28 Aug 2017, at  23:10, hjcho616 <hjcho...@yahoo.com> wrote: Thank you  
Tomasz and Ronny.  I'll have to order some hdd soon  and try these out.  Car 
battery idea is nice!  I may try that.. =)  Do they last longer?  Ones that fit 
the UPS original  battery spec didn't last very long... part of the reason why 
I  gave up on them.. =P  My wife probably won't like the idea  of car battery 
hanging out though ha! The OSD1 (one with mostly ok  OSDs, except that smart 
failure)  motherboard doesn't have  any additional SATA connectors  available.  
Would it be safe to add another OSD  host? Regards, Hong  On Monday, August 28, 
 2017 4:43 PM, Tomasz Kusmierz <tom.kusmierz@g mail.com> wrote: Sorry for  
being brutal … anyway  1. get the battery for  UPS ( a car battery will do as 
well,  I’ve moded on ups  in the past with truck battery and it  was working 
like a charm :D ) 2. get spare drives and put  those in because your cluster 
CAN  NOT get out of  error due to lack of space 3. Follow  advice of Ronny 
Aasen on hot to recover data from hard drives  4 get cooling to drives or  you 
will loose more !  
 On 28 Aug 2017, at  22:39, hjcho616 <hjcho...@yahoo.com> wrote: Tomasz,  Those 
machines are behind a surge  protector.  Doesn't appear to be a good one!   I 
do have a UPS... but it is my fault...  no battery.  Power was pretty reliable 
for a  while... and UPS was just beeping  every chance it had,  disrupting some 
sleep.. =P  So running on surge  protector only.  I am running this in home  
environment.   So far, HDD failures have  been very rare for this environment.  
=)  It just doesn't get loaded as  much!  I am not sure what to  expect, seeing 
that "unfound" and just a  feeling of possibility of  maybe getting OSD back 
made me excited  about it. =) Thanks for  letting me know what should be the  
priority.  I just lack experience and  knowledge in this. =) Please do  
continue to guide me  though this.  Thank you for the decode of  that smart 
messages!  I do agree that looks like it  is on its way out.  I would like to 
know how to get  good portion of it back if possible. =) I think I just set the 
size  and min_size to 1. # ceph osd  lspools 0 data,1  metadata,2 rbd, # ceph 
osd  pool set rbd size 1 set pool 2  size to 1 # ceph osd  pool set rbd 
min_size 1 set pool 2  min_size to 1 Seems to be doing some backfilling work.  
# ceph health HEALTH_ERR 22 pgs are stuck  inactive for more than 300  seconds; 
2 pgs backfill_toofull;  74 pgs backfill_wait;  3 pgs backfilling; 108 pgs  
degraded; 6 pgs down; 6 pgs  inconsistent; 6 pgs peering;  7 pgs recovery_wait; 
16 pgs stale;  108 pgs stuck degraded; 6  pgs stuck inactive; 16  pgs stuck 
stale; 130 pgs stuck unclean;  101 pgs stuck  undersized; 101 pgs undersized; 1 
 requests are blocked 
 32 sec; recovery  1790657/4502340 objects degraded  (39.772%); 
 recovery 641906/4502340  objects misplaced (14.257%);  recovery 147/2251990 
unfound (0.007%); 50 scrub errors;  mds cluster is degraded; no legacy OSD 
present but 'sortbitwise'  flag is not set Regards,  Hong On Monday, August 28, 
2017 4:18 PM, Tomasz  Kusmierz <tom.kusmierz @gmail.com> wrote: So to decode  
few things about your disk:    1 Raw_Read_Error_Rate    0x002f  100  100  051   
 Pre-fail  Always      -      37 37 read erros and only one  sector marked as 
pending - fun disk  :/  181 Program_Fail_Cnt_Total  0x0022  099  099  000    
Old_age  Always      -      35325174 So firmware has quite few  bugs, that’s 
nice 191  G-Sense_Error_Rate      0x0022  100  100  000    Old_age  Always      
-      2855 disk was thrown around  while operational even more  nice. 194 
Temperature_Celsius    0x0002  047  041  000    Old_age  Always      -      53 
(Min/Max 15/59)  if your disk passes 50 you should not  consider using it, high 
 temperatures demagnetise plate layer  and you will see more errors in very 
near future.  197 Current_Pending_Sector  0x0032  100  100  000    Old_age  
Always      -      1 as mentioned before :)  200 Multi_Zone_Error_Rate  0x002a  
100  100  000    Old_age  Always      -      4222 your heads keep missing  
tracks … bent ? I don’t even know how to comment here. generally fun  drive 
you’ve got there … rescue as much as you can and throw it  away !!! 
 
 
_______________________________________________ ceph-users mailing list 
ceph-users@lists.ceph.com 
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com 
 
 
  
 _______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
  
  
     _______________________________________________
 ceph-users mailing list
 ceph-users@lists.ceph.com
 http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
  
 
         
 
         
 
         
 
         
 
      
 
  _______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


   

   

   
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to