svetljo wrote:

> Hi people, could anybody help with advice
>
>
>>  
>>
>> > svetljo wrote:
>> >
>> >> a little note
>> >> i  patch the xfs kernel but mkfs.reiserfs segfaults again
>> >>
>> >> svetljo wrote:
>> >>
>> >>> it's me again and out there is no one
>> >>> so
>> >>> well linux-2.4.10-pre2-xfs LVM-1.0.1rc2 works with XFS and JFS on
>> >>> the LV
>> >>> ext2 and reiserfs segfaults as befor
>> >>> i've compiled the other kernel's with JFS 1.0.4 but i don't see the
>> >>> point to test them
>> >>> i'm shure that JFS will work on the LV and the story with ext2 and
>> >>> reiserfs will be the same
>> >>> but i'll probably try the ReiserFS patches  for 2.4.9 ( after
>> >>> www.namesys.com they already sent them to Linus )
>> >>>
>> >>> i'm using e2fsprogs-1.2.4 and reiserfsprogs-3.x.0j
>> >>>
>> >>> svetljo wrote:
>> >>>
>> >>> > well i did the tests
>> >>> >
>> >>> > clean Linus kernel-2.4.9  LVM-1.0.1rc1                    ext2 and
>> >>> > reiserfs segfaults
>> >>> > clean            linux-2.4.9-ac5 LVM-1.0                        
>> ext2
>> >>> > and reiserfs segfaults
>> >>> >                   | linux-2.4.10-pre2-xfs LVM-1.0.1rc2      
>> ext2 and
>> >>> > reiserfs segfaults, but xfs seems to work
>> >>> >                   | from SGI's cvs tree  linux-2.4-xfs taken 
>> today in
>> >>> > the early morning
>> >>> >
>> >>> > i'll try to ad IBM's JFS and try it again and i'll try bonnie 
>> on them
>> >>> > can you give me ideas how to stres the FS to find out whether it
>> >>> realy
>> >>> > works with XFS and JFS
>> >>> >
>> >>> > what could be wrong with ext2 and reiserfs as it works with 
>> JFS-1.0.3
>> >>> > and with the latest cvs XFS
>> >>> > svetljo wrote:
>> >>> >
>> >>> >> Hi
>> >>> >>
>> >>> >> a small adition i've tried to format the LV with ext2 and 
>> reiserfs,
>> >>> >> but it didn't worked : mkfs segfaults
>> >>> >> a strange one : i'm able to format with IBM JFS , and i can work
>> >>> >> without a problem with the LV everything just to be fine with JFS
>> >>> >>
>> >>> >> i'm currently building:
>> >>> >>  clean 2.4.9-linus with LVM-1.0.1rc1
>> >>> >>       2.4.9-ac5 with LVM-1.0 ( i couldn't do it with LVM-1.0.1rc1
>> >>> & rc2)
>> >>> >>       2.4.10-pre2-xfs-cvs with LVM-1.0.1rc2
>> >>> >> to find out what is going on with ext2 reiserfs XFS ,
>> >>> >> is the problem coming from the XFS kernel changes
>> >>> >>
>> >>> >> >Hi
>> >>> >> >i'm having a serios trouble with creating
>> >>> >> >a LVM over software linear RAID
>> >>> >> >well i created it, formated it with XFS
>> >>> >> >but every time i try to mount the LV mount segfaults
>> >>> >> >and then i can not mount any other file system ( partition, 
>> CD, ..
>> >>> >> >until i reboot, when i try to mount smth mount simple stop to
>> >>> respond
>> >>> >> >without an error and blocks the console
>> >>> >> >
>> >>> >> >i'm using XFS cvs kernel 2.4.9 and LVM-1.0.1rc1
>> >>> >> >on ABIT's BP6 2xCelleron 533 512Mb RAM
>> >>> >> >the drives are on onboard HPT366 controler 2xWD 30Gb 1xMaxtor 
>> 40Gb
>> >>> >> >
>> >>> >> >the LV is striped over the 3 devices of the VG
>> >>> >> >the VG is /dev/hdh10 /dev/md6 /dev/md7
>> >>> >> >/dev/md6 is linear software RAID /dev/hde6 /dev/hde12
>> >>> >> >/dev/md7 is linear software RAID /dev/hdg1 /dev/hdg5 /dev/hdg6
>> >>> >> >dev/hdg11
>> >>> >> >
>> >>> >> >i posted  to the LVM-lists and there i was told
>> >>> >> >to try  "dmesg | ksymoops"
>> >>> >> >
>> >>> >> >and i became the folowing answer
>> >>> >> >
>> >>> >> > > >>EIP; e29c0266 <[linear]linear_make_request+36/f0> <=====
>> >>> >> > >
>> >>> >> > >> Trace; c023fa12 <__make_request+412/6d0>
>> >>> >> > >> Trace; c0278dcd <md_make_request+4d/80>
>> >>> >> > >> Trace; c027fa0f <lvm_make_request_fn+f/20>
>> >>> >> > >> Trace; c023fd89 <generic_make_request+b9/120>
>> >>> >> > >
>> >>> >> > >OK, so the oops is inside the RAID layer, but it may be that
>> >>> it is
>> >>> >> > >being fed bogus data from a higher layer.  Even so, it 
>> should not
>> >>> >> > >oops in this case.  Since XFS changes a lot of the kernel 
>> code, I
>> >>> >> > >would either suggest asking the XFS folks to look at this 
>> oops,
>> >>> >> > >or maybe on the MD RAID mailing list, as they will know more
>> >>> about
>> >>> >> >it.
>> >>> >> >
>> >>> >> >this is the full "dmesg | ksymoops" , i'll try to use other FS
>> >>> to find
>> >>> >> >out whether it's a problem with XFS, but i wish me not to have
>> >>> to use
>> >>> >> >other FS, i realy love XFS
>> >>> >>
>> >>> >>
>> >>> >> Using defaults from ksymoops -t elf32-i386 -a i386
>> >>> >> EFLAGS: 00010247
>> >>> >> eax: 004ac1ab   ebx: 004ac1ab   ecx: 00000000   edx: 00000000
>> >>> >> esi: d54eb320   edi: c188b928   ebp: 00958357   esp: d4eb3670
>> >>> >> ds: 0018   es: 0018   ss: 0018
>> >>> >> Process mount (pid: 5536, stackpage=d4eb3000)
>> >>> >> Stack: d54eb3e0 c023fa12 00000907 d54eb320 00000000 01c02000
>> >>> c0278dcd
>> >>> >> dcec43c0
>> >>> >>         00000000 d54eb320 d54eb320 00000000 01c02000 c027fa0f
>> >>> 00000001
>> >>> >> d54eb320
>> >>> >>         c023fd89 c03a7254 00000000 d54eb320 00000282 00000021
>> >>> 00000000
>> >>> >> 00000000
>> >>> >> Call Trace: [<c023fa12>] [<c0278dcd>] [<c027fa0f>] [<c023fd89>]
>> >>> >> [<c01a6814>]
>> >>> >>     [<c01a6a85>] [<c01a6fc1>] [<c01a6c47>] [<c01a6990>] 
>> [<c0105dac>]
>> >>> >> [<c0105f1c>]
>> >>> >>     [<c02e2140>] [<c021c10a>] [<c01fe5b8>] [<c01ff2a4>] 
>> [<c01a553e>]
>> >>> >> [<c01feb6f>]
>> >>> >>     [<c01feed8>] [<c01fc322>] [<c0201f40>] [<c01fb8f3>] 
>> [<c0202fdf>]
>> >>> >> [<c02026bf>]
>> >>> >>     [<c01a60be>] [<c02026eb>] [<c021e674>] [<c020b69c>] 
>> [<c020b843>]
>> >>> >> [<c020b871>]
>> >>> >>     [<c021cf48>] [<c01294e0>] [<c0125f0e>] [<c0125d9d>] 
>> [<c013cd72>]
>> >>> >> [<c013d01b>]
>> >>> >>     [<c013dafc>] [<c01131e0>] [<c010724c>] [<c013dd56>] 
>> [<c013dbfc>]
>> >>> >> [<c013de13>]
>> >>> >>     [<c010715b>]
>> >>> >> Code: f7 f9 85 d2 74 24 55 51 68 c0 03 9c e2 e8 58 6c 75 dd 6a 00
>> >>> >>
>> >>> >>   >>EIP; e29c0266 <[linear]linear_make_request+36/f0>   <=====
>> >>> >> Trace; c023fa12 <__make_request+412/6d0>
>> >>> >> Trace; c0278dcd <md_make_request+4d/80>
>> >>> >> Trace; c027fa0f <lvm_make_request_fn+f/20>
>> >>> >> Trace; c023fd89 <generic_make_request+b9/120>
>> >>> >> Trace; c01a6814 <_pagebuf_page_io+1f4/370>
>> >>> >> Trace; c01a6a85 <_page_buf_page_apply+f5/1c0>
>> >>> >> Trace; c01a6fc1 <pagebuf_segment_apply+b1/e0>
>> >>> >> Trace; c01a6c47 <pagebuf_iorequest+f7/160>
>> >>> >> Trace; c01a6990 <_page_buf_page_apply+0/1c0>
>> >>> >> Trace; c0105dac <__down+bc/d0>
>> >>> >> Trace; c0105f1c <__down_failed+8/c>
>> >>> >> Trace; c02e2140 <stext_lock+45b4/99d6>
>> >>> >> Trace; c021c10a <xfsbdstrat+3a/40>
>> >>> >> Trace; c01fe5b8 <xlog_bread+48/80>
>> >>> >> Trace; c01ff2a4 <xlog_find_zeroed+94/1e0>
>> >>> >> Trace; c01a553e <_pagebuf_get_object+3e/170>
>> >>> >> Trace; c01feb6f <xlog_find_head+1f/370>
>> >>> >> Trace; c01feed8 <xlog_find_tail+18/350>
>> >>> >> Trace; c01fc322 <xlog_alloc_log+2a2/2e0>
>> >>> >> Trace; c0201f40 <xlog_recover+20/c0>
>> >>> >> Trace; c01fb8f3 <xfs_log_mount+73/b0>
>> >>> >> Trace; c0202fdf <xfs_mountfs+55f/e20>
>> >>> >> Trace; c02026bf <xfs_readsb+af/f0>
>> >>> >> Trace; c01a60be <pagebuf_rele+3e/80>
>> >>> >> Trace; c02026eb <xfs_readsb+db/f0>
>> >>> >> Trace; c021e674 <kmem_alloc+e4/110>
>> >>> >> Trace; c020b69c <xfs_cmountfs+4bc/590>
>> >>> >> Trace; c020b843 <xfs_mount+63/70>
>> >>> >> Trace; c020b871 <xfs_vfsmount+21/40>
>> >>> >> Trace; c021cf48 <linvfs_read_super+188/270>
>> >>> >> Trace; c01294e0 <filemap_nopage+2c0/410>
>> >>> >> Trace; c0125f0e <handle_mm_fault+ce/e0>
>> >>> >> Trace; c0125d9d <do_no_page+4d/f0>
>> >>> >> Trace; c013cd72 <read_super+72/110>
>> >>> >> Trace; c013d01b <get_sb_bdev+18b/1e0>
>> >>> >> Trace; c013dafc <do_add_mount+1dc/290>
>> >>> >> Trace; c01131e0 <do_page_fault+0/4b0>
>> >>> >> Trace; c010724c <error_code+34/3c>
>> >>> >> Trace; c013dd56 <do_mount+106/120>
>> >>> >> Trace; c013dbfc <copy_mount_options+4c/a0>
>> >>> >> Trace; c013de13 <sys_mount+a3/130>
>> >>> >> Trace; c010715b <system_call+33/38>
>> >>> >> Code;  e29c0266 <[linear]linear_make_request+36/f0>
>> >>> >> 00000000 <_EIP>:
>> >>> >> Code;  e29c0266 <[linear]linear_make_request+36/f0>   <=====
>> >>> >>     0:   f7 f9                     idiv   %ecx,%eax   <=====
>> >>> >> Code;  e29c0268 <[linear]linear_make_request+38/f0>
>> >>> >>     2:   85 d2                     test   %edx,%edx
>> >>> >> Code;  e29c026a <[linear]linear_make_request+3a/f0>
>> >>> >>     4:   74 24                     je     2a <_EIP+0x2a> e29c0290
>> >>> >> <[linear]linear_make_request+60/f0>
>> >>> >> Code;  e29c026c <[linear]linear_make_request+3c/f0>
>> >>> >>     6:   55                        push   %ebp
>> >>> >> Code;  e29c026d <[linear]linear_make_request+3d/f0>
>> >>> >>     7:   51                        push   %ecx
>> >>> >> Code;  e29c026e <[linear]linear_make_request+3e/f0>
>> >>> >>     8:   68 c0 03 9c e2            push   $0xe29c03c0
>> >>> >> Code;  e29c0273 <[linear]linear_make_request+43/f0>
>> >>> >>     d:   e8 58 6c 75 dd            call   dd756c6a 
>> <_EIP+0xdd756c6a>
>> >>> >> c0116ed0 <printk+0/1a0>
>> >>> >> Code;  e29c0278 <[linear]linear_make_request+48/f0>
>> >>> >>    12:   6a 00                     push   $0x0
>> >>> >>
>> >>> >> Andreas Dilger wrote:
>> >>> >>
>> >>> >>  >On Aug 31, 2001  15:08 +0200, svetljo wrote:
>> >>> >>  >
>> >>> >>  >>[root@svetljo mnt]# mount -t xfs /dev/myData/Music music
>> >>> >>  >>Segmentation fault
>> >>> >>  >>
>> >>> >>  >
>> >>> >>  >Generally this is a bad sign.  Either mount is segfaulting
>> >>> (unlikely)
>> >>> >>  >or you are getting an oops in the kernel.  You need do run
>> >>> something
>> >>> >>  >like "dmesg | ksymoops" in order to get some useful data about
>> >>> where
>> >>> >>  >the problem is (could be xfs, LVM, or elsewhere in the kernel).
>> >>> >>  >
>> >>> >>  >Once you have an oops, you are best off rebooting the system,
>> >>> because
>> >>> >>  >your kernel memory may be corrupted, and cause more oopses
>> >>> which do
>> >>> >>  >not mean anything.  If you look in /var/log/messages (or
>> >>> >> /var/log/kern.log
>> >>> >>  >or some other place, depending on where kernel messages go),
>> >>> you can
>> >>> >>  >decode the FIRST oops in the log with ksymoops.  All subsequent
>> >>> >> ones are
>> >>> >>  >useless.
>> >>> >>  >
>> >>> >>  >
>> >>> >>  >>the LV ( lvcreate -i3 -I4 -L26G -nMusic )
>> >>> >>  >>
>> >>> >>  >>the VG -> myData   /dev/hdh10 /dev/linVG1/linLV1
>> >>> /dev/linVG2/linLV2
>> >>> >>  >>
>> >>> >>  >>/dev/hdh10 normal partition 14G
>> >>> >>  >>/dev/linVG1/linLV1 ->  linear LV 14G /dev/hde6 /dev/hde12
>> >>> >>  >>/dev/linVg2/linLV2 ->  linear LV 14G /dev/hdg1 /dev/hdg5
>> >>> /dev/hdg6
>> >>> >> /dev/hdg12
>> >>> >>  >>
>> >>> >>  >
>> >>> >>  >There is absolutely no point in doing this (not that it is
>> >>> possible
>> >>> >> to do
>> >>> >>  >so anyways).  First of all, striping is almost never needed 
>> "for
>> >>> >> performance"
>> >>> >>  >unless you are normally doing very large sequential I/Os, 
>> and even
>> >>> >> so most
>> >>> >>  >disks today have very good sequential I/O rates (e.g. 
>> 15-30MB/s).
>> >>> >> Secondly,
>> >>> >>  >you _should_ be able to just create a single LV that is striped
>> >>> >> across all
>> >>> >>  >of the PVs above.  You would likely need to build it in 
>> steps, to
>> >>> >> ensure
>> >>> >>  >that it is striped across the disks correctly.
>> >>> >>  >
>> >>> >>  >Cheers, Andreas
>> >>> >>  >
>> >>> >>
>> >>> >>
>> >>> >>
>> >>> >>
>> >>> >>
>> >>> >
>> >>> >
>> >>> >
>> >>>
>> >>>
>> >>>
>> >>>
>> >>>
>> >>> _______________________________________________
>> >>> linux-lvm mailing list
>> >>> [EMAIL PROTECTED]
>> >>> http://lists.sistina.com/mailman/listinfo/linux-lvm
>> >>> read the LVM HOW-TO at http://www.sistina.com/lvm/Pages/howto.html
>> >>>
>> >>
>> >>
>> >>
>> >>
>> >> _______________________________________________
>> >> linux-lvm mailing list
>> >> [EMAIL PROTECTED]
>> >> http://lists.sistina.com/mailman/listinfo/linux-lvm
>> >> read the LVM HOW-TO at http://www.sistina.com/lvm/Pages/howto.html
>> >>
>> >
>> >
>> >
>>
>>
>>
>>
>
>



Want to buy your Pack or Services from MandrakeSoft? 
Go to http://www.mandrakestore.com

Reply via email to