[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: [dm-devel] kernel panic with bitmap_startwrite



On Wed, 30 Oct 2013 12:54:33 +0400 Yuriy Vet <4363 list ru> wrote:

> 
> Hello!
> I have an IBM System x3200 M2 with Fibre HBA: Brocade Communications Systems, Inc. 415/815/41B/81B 4Gbps/8Gbps PCIe single port FC HBA (rev 01), on it installed Red Hat Enterprise Linux 5 kernel 2.6.18-238.19.1.
> used native multipath-tools v0.4.7 (03/12, 2006) and mdadm.
> #modinfo bfa
> filename: /lib/modules/2.6.18-238.19.1.el5/kernel/drivers/scsi/bfa/bfa.ko
> version: 2.1.2.0
> author: Brocade Communications Systems, Inc.
> description: Brocade Fibre Channel HBA Driver fcpim
> and IBM DS 3512 with 2 LUN, LUN this raid6
> configured multipath configuration:
> #cat /etc/multipath.conf
> defaults {
> # user_friendly_names no
> # polling_interval 2
> # path_selector "round-robin 0"
> # selector "round-robin 0"
> # polling_interval 2
> # find_multipaths yes
> # flush_on_last_del yes
> # checker_timeout 10
> # failback immediate
> # failback manual
> # path_checker readsector0
> # path_checker rdac
> # path_grouping_policy failover
> # path_grouping_policy multibus
> # rr_min_io 100
> # features 0
> # no_path_retry fail
> # getuid_callout "/sbin/scsi_id -g -u -s /block/%n"
> # getuid_callout "/lib/udev/scsi_id --whitelisted --device=/dev/%n."
> # prio_callout "/sbin/mpath_prio_rdac /dev/%n"
> # dev_loss_tmo 10
> # fast_io_fail_tmo 5
> }#defaults
> ####################
> blacklist {
> devnode "sda"
> wwid *
> }#blacklist
> ####################
> blacklist_exceptions {
> #osnovnoy_1
> wwid 360080e50002ef5ee00000bbe524a8ced
> #rezerv_1
> wwid 360080e50002ef5ee00000bc1524a8d93
> }#blacklist_exceptions
> ####################
> devices {
> device {
> vendor "IBM"
> product "1746 FAStT"
> # path_checker rdac
> # path_selector "round-robin 0"
> }#device
> }#devices
> ####################
> multipaths {
> multipath {
> wwid 360080e50002ef5ee00000bbe524a8ced
> alias osnov_1
> }
> multipath {
> wwid 360080e50002ef5ee00000bc1524a8d93
> alias rezerv_1
> }
> }
> 
> #lsscsi
> [0:2:0:0] disk IBM ServeRAID-MR10k 1.40 /dev/sda
> [3:0:0:0] disk IBM 1746 FAStT 1070 /dev/sdb
> [3:0:0:1] disk IBM 1746 FAStT 1070 /dev/sdf
> [3:0:0:31] disk IBM Universal Xport 1070 -
> [3:0:1:0] disk IBM 1746 FAStT 1070 /dev/sdh
> [3:0:1:1] disk IBM 1746 FAStT 1070 /dev/sdi
> [3:0:1:31] disk IBM Universal Xport 1070 -
> [3:0:2:0] disk IBM 1746 FAStT 1070 /dev/sdj
> [3:0:2:1] disk IBM 1746 FAStT 1070 /dev/sdk
> [3:0:2:31] disk IBM Universal Xport 1070 -
> [3:0:3:0] disk IBM 1746 FAStT 1070 /dev/sdl
> [3:0:3:1] disk IBM 1746 FAStT 1070 /dev/sdm
> [3:0:3:31] disk IBM Universal Xport 1070 -
> [4:0:0:0] disk IBM 1746 FAStT 1070 /dev/sdc
> [4:0:0:1] disk IBM 1746 FAStT 1070 /dev/sdd
> [4:0:0:31] disk IBM Universal Xport 1070 -
> [4:0:1:0] disk IBM 1746 FAStT 1070 /dev/sde
> [4:0:1:1] disk IBM 1746 FAStT 1070 /dev/sdg
> [4:0:1:31] disk IBM Universal Xport 1070 -
> [4:0:2:0] disk IBM 1746 FAStT 1070 /dev/sdn
> [4:0:2:1] disk IBM 1746 FAStT 1070 /dev/sdo
> [4:0:2:31] disk IBM Universal Xport 1070 -
> [4:0:3:0] disk IBM 1746 FAStT 1070 /dev/sdp
> [4:0:3:1] disk IBM 1746 FAStT 1070 /dev/sdq
> [4:0:3:31] disk IBM Universal Xport 1070 -
> 
> #multipath -ll
> osnov_1 (360080e50002ef5ee00000bbe524a8ced) dm-2 IBM,1746 FAStT
> [size=11T][features=3 queue_if_no_path pg_init_retries 50][hwhandler=1 rdac][rw]
> \_ round-robin 0 [prio=400][active]
> \_ 4:0:0:0 sdc 8:32 [active][ready]
> \_ 3:0:1:0 sdh 8:112 [active][ready]
> \_ 3:0:3:0 sdl 8:176 [active][ready]
> \_ 4:0:2:0 sdn 8:208 [active][ready]
> \_ round-robin 0 [prio=0][enabled]
> \_ 3:0:0:0 sdb 8:16 [active][ghost]
> \_ 4:0:1:0 sde 8:64 [active][ghost]
> \_ 3:0:2:0 sdj 8:144 [active][ghost]
> \_ 4:0:3:0 sdp 8:240 [active][ghost]
> rezerv_1 (360080e50002ef5ee00000bc1524a8d93) dm-3 IBM,1746 FAStT
> [size=11T][features=3 queue_if_no_path pg_init_retries 50][hwhandler=1 rdac][rw]
> \_ round-robin 0 [prio=400][active]
> \_ 4:0:0:1 sdd 8:48 [active][ready]
> \_ 3:0:1:1 sdi 8:128 [active][ready]
> \_ 3:0:3:1 sdm 8:192 [active][ready]
> \_ 4:0:2:1 sdo 8:224 [active][ready]
> \_ round-robin 0 [prio=0][enabled]
> \_ 3:0:0:1 sdf 8:80 [active][ghost]
> \_ 4:0:1:1 sdg 8:96 [active][ghost]
> \_ 3:0:2:1 sdk 8:160 [active][ghost]
> \_ 4:0:3:1 sdq 65:0 [active][ghost]
> 
> configured mdadm configuration:
> #mdadm -D /dev/md0
> mdadm: metadata format 1.00 unknown, ignored.
> /dev/md0:
> Version : 1.00
> Creation Time : Tue Oct 22 12:34:01 2013
> Raid Level : raid1
> Array Size : 11718818096 (11175.94 GiB 12000.07 GB)
> Used Dev Size : 11718818096 (11175.94 GiB 12000.07 GB)
> Raid Devices : 2
> Total Devices : 2
> Preferred Minor : 0
> Persistence : Superblock is persistent
> Intent Bitmap : Internal
> Update Time : Thu Oct 24 11:45:38 2013
> State : active
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
> Name : 0
> UUID : cfffde5b:3f232606:f45a5be9:d7a5c4b2
> Events : 21
> Number Major Minor RaidDevice State
> 0 253 4 0 active sync /dev/dm-4
> 1 253 5 1 active sync /dev/dm-5
> 
> #cat /proc/mdstat
> Personalities : [raid1]
> md0 : active raid1 dm-4[0] dm-5[1]
> 11718818096 blocks super 1.0 [2/2] [UU]
> bitmap: 0/11 pages [0KB], 524288KB chunk
> unused devices: <none>
> 
> md0 is formatted in ext3
> #tune2fs -l /dev/md0
> tune2fs 1.39 (29-May-2006)
> Filesystem volume name: <none>
> Last mounted on: <not available>
> Filesystem UUID: 6a2352e4-be78-4c53-80a3-a7e7614cb6c2
> Filesystem magic number: 0xEF53
> Filesystem revision #: 1 (dynamic)
> Filesystem features: has_journal resize_inode dir_index filetype needs_recovery sparse_super large_file
> Default mount options: (none)
> Filesystem state: clean
> Errors behavior: Continue
> Filesystem OS type: Linux
> Inode count: 1464860672
> Block count: 2929704524
> Reserved block count: 29297045
> Free blocks: 2883691408
> Free inodes: 1464860662
> First block: 0
> Block size: 4096
> Fragment size: 4096
> Reserved GDT blocks: 325
> Blocks per group: 32768
> Fragments per group: 32768
> Inodes per group: 16384
> Inode blocks per group: 512
> Filesystem created: Wed Oct 23 09:16:11 2013
> Last mount time: Thu Oct 24 11:45:33 2013
> Last write time: Thu Oct 24 11:45:33 2013
> Mount count: 5
> Maximum mount count: 38
> Last checked: Wed Oct 23 09:16:11 2013
> Check interval: 15552000 (6 months)
> Next check after: Mon Apr 21 09:16:11 2014
> Reserved blocks uid: 0 (user root)
> Reserved blocks gid: 0 (group root)
> First inode: 11
> Inode size: 128
> Journal inode: 8
> Default directory hash: tea
> Directory Hash Seed: 82383c38-be3a-4096-9082-ad295ee3404e
> Journal backup: inode blocks
> md0 mounted on /osnov.
> and using the command:
> #dd if=/dev/zero of=/osnov/test bs=1MB count=1000
> or 
> simply copy from MCvs
> we get:
> Code: 0f 0b 68 ce fc 2c 88 c2 95 04 8d 42 01 66 89 45 00 c7 43 50
> RIP [<ffffffff80218ab2>] bitmap_startwrite+0xb5/0x101
>     RSP<ffff81105fbb9c20>
> <0>Kernel panic - not syncing: Fatal exception
> without bitmap all works!
> You can get rid of this problem without installing new drivers and programs?
> 

This is a problem with md/raid rather than dm, so it would have been best to
send the email to linux-raid vger kernel org   But sometimes I watch dm-devel
too.

There should have been a much longer "oops" messages with that "kernel panic"
we really need to see all of it.

2.6.18 is a fairly ancient kernel.  As you are running Enterprise Linux I
suggest that you should actually go through RedHat support channels rather
than community support channels.

NeilBrown

Attachment: signature.asc
Description: PGP signature


[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]