← Back to team overview

kernel-packages team mailing list archive

[Bug 1190295] Re: 2.6.32-47 kernel update on 10.04 breaks software RAID (+ LVM)


It's possible that the patches which were linked are not related.
They're not in the delta between the two versions which are listed.

Here are some questions to try to narrow down the field of potential

1. You mention that this occurs on various upgrades, could you confirm that these
would be various different kernel versions over time, and not ONLY on the 46 to 47 update?

2. Have you seen this when doing any other reboots (not after system updates)?
(such as might occur if this was a boot race -  which could only shows up because the first boot after an update performs additional operations)

3. When this occurs you mention that you see the 'M' for manual
recovery, do you also see the MD "degraded raid" prompt and if so how do
you respond?

4. On taking the 'M' that option LVM slices are missing which are served
from the RAID, can you provide the following information for missing LVs
(backed by md0):

  A) are the device links in /dev/<vgname>/<lvname> present?
  B) are the LVs listed in the output of 'lvs' and what are their state?
  C) are the volumes present in the 'dmsetup ls' output?
  D) are the PVs which are backed by md0 present in 'pvs' and what are their state?
  E) what is the actual state of md0 as show in 'cat /proc/mdstat'?

You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.

  2.6.32-47 kernel update on 10.04 breaks software RAID (+ LVM)

Status in “linux” package in Ubuntu:

Bug description:
  Been running 10.04 LTS on 8 similar AMD Opteron x86_64 servers for
  several years.  The servers have been kept up-to-date with patches as
  they come out.  These servers have been running 2.6.x kernels. Each
  server has some form of Linux software RAID running on it as well as
  3Ware hardware RAID card using SATA disks.  Software RAID is
  configured as RAID1 for all but one server running software RAID10.
  All servers had software raid configured to use single partitions on
  each disk of types of 0xFD (Linux Software Raid Autodetect).  All
  servers were configured with LVM over the top of /dev/md0.

  In past year, mysterious problems have been happening with software
  RAID after applying system patches.  Upon reboot, server is unable to
  mount LVM partitions on Linux software RAID and boot is interrupted
  with "Continue to wait; or Press S to skip mounting or M for manual
  recovery" requiring intervention from an operator.

  Upon pressing 'M' and logging in as root, the LVM slices on the
  software RAID partition are not mounted and sometimes appear to be
  missing from LVM.  Oftentimes pvs, vgs and lvs will complain about
  "leaking memory". Germane to the issue, LVM will sometimes show the
  problem partitions as "Active" while other times during the login,
  they will simply be gone.  With LVM  and /dev/md0 unstable, there is
  no way to discern the true state of the partitons in question.
  Starting the system from alternate boot media such as CDROM or USB
  drive, sometimes shows the software RAID and LVM in proper state which
  leads to suspicion of a kernel update on the afflicted system.
  Historically and subjectively, best practice in this instance seems to
  be booting from live media and starting the array degraded mode, and
  backing up the array.

To manage notifications about this bug go to: