LSI 1068E instabilities

Bug #605939 reported by nutznboltz
20
This bug affects 4 people
Affects Status Importance Assigned to Milestone
linux (Ubuntu)
Invalid
Undecided
Unassigned

Bug Description

I am unable to deploy Ubuntu 10.04 LTS on servers that have LSI 1068E (including LSI SAS 3442E-R and LSI 3081E-R) controllers because of instabilities. See attached dmesg output, search for sdb errors.

While resyncing a 1 TB RAID1

cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid1 sdb1[2] sda1[0]
      976760768 blocks [2/1] [U_]
      [=======>.............] recovery = 37.4% (366153984/976760768) finish=1512.0min speed=6730K/sec

goes to

Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid1 sdb1[2](F) sda1[0]
      976760768 blocks [2/1] [U_]

and the disk is not bad; it works with Jaunty.

ProblemType: Bug
DistroRelease: Ubuntu 10.04
Package: linux-image-2.6.32-23-server 2.6.32-23.37
Regression: Yes
Reproducible: Yes
ProcVersionSignature: Ubuntu 2.6.32-23.37-server 2.6.32.15+drm33.5
Uname: Linux 2.6.32-23-server x86_64
AlsaDevices: Error: command ['ls', '-l', '/dev/snd/'] failed with exit code 2: ls: cannot access /dev/snd/: No such file or directory
AplayDevices: Error: [Errno 2] No such file or directory
Architecture: amd64
ArecordDevices: Error: [Errno 2] No such file or directory
Date: Thu Jul 15 11:56:20 2010
Lsusb:
 Bus 003 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub
 Bus 002 Device 001: ID 1d6b:0001 Linux Foundation 1.1 root hub
 Bus 001 Device 003: ID 0000:0000
 Bus 001 Device 002: ID 04b4:6560 Cypress Semiconductor Corp. CY7C65640 USB-2.0 "TetraHub"
 Bus 001 Device 001: ID 1d6b:0002 Linux Foundation 2.0 root hub
MachineType: HP ProLiant DL165 G5
PciMultimedia:

ProcCmdLine: BOOT_IMAGE=/boot/vmlinuz-2.6.32-23-server root=/dev/mapper/rootvg-rootlv ro console=ttyS0,9600n8 nomodeset
ProcEnviron:
 PATH=(custom, no user)
 LANG=en_US.UTF-8
 SHELL=/bin/bash
SourcePackage: linux
dmi.bios.date: 11/03/2009
dmi.bios.vendor: HP
dmi.bios.version: O13
dmi.chassis.type: 23
dmi.chassis.vendor: HP
dmi.modalias: dmi:bvnHP:bvrO13:bd11/03/2009:svnHP:pnProLiantDL165G5:pvr:cvnHP:ct23:cvr:
dmi.product.name: ProLiant DL165 G5
dmi.sys.vendor: HP

Revision history for this message
nutznboltz (nutznboltz-deactivatedaccount) wrote :
Revision history for this message
nutznboltz (nutznboltz-deactivatedaccount) wrote :

I have commenced testing this server with:
sudo add-apt-repository ppa:kernel-ppa/ppa
sudo aptitude update
sudo aptitude install linux-image-2.6.35-7-server
sudo reboot

$ uname -rsv
Linux 2.6.35-7-server #11~lucid1-Ubuntu SMP Wed Jul 7 16:32:35 UTC 2010

Drive is resyncing:

Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid1 sdb1[2] sda1[0]
      976760768 blocks [2/1] [U_]
      [>....................] recovery = 0.1% (1580608/976760768) finish=2542.4min speed=6392K/sec

Revision history for this message
nutznboltz (nutznboltz-deactivatedaccount) wrote :

Same error with even with the newer kernel. See attached dmesg output.

Revision history for this message
nutznboltz (nutznboltz-deactivatedaccount) wrote :

I rebuilt the server as Jaunty and the RAID1 has been syncing for about eight hours without any errors.

$ uname -srv
Linux 2.6.28-19-generic #61-Ubuntu SMP Wed May 26 23:32:46 UTC 2010

Not only that, it's syncing much faster. 12349K/sec vs. 6392K/sec

$ cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md1 : active raid1 sda2[0] sdb2[1]
      976470784 blocks [2/2] [UU]
      [========>............] resync = 41.0% (401038656/976470784) finish=776.5min speed=12349K/sec

md0 : active raid1 sda1[0] sdb1[1]
      289024 blocks [2/2] [UU]

Revision history for this message
nutznboltz (nutznboltz-deactivatedaccount) wrote :

With Jaunty installed the RAID1 resynced without errors.

Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md1 : active raid1 sda2[0] sdb2[1]
      976470784 blocks [2/2] [UU]

md0 : active raid1 sda1[0] sdb1[1]
      289024 blocks [2/2] [UU]

unused devices: <none>

Revision history for this message
Ante Karamatić (ivoks) wrote :
Changed in linux (Ubuntu):
status: New → Confirmed
Brad Figg (brad-figg)
tags: added: acpi-method-return
tags: added: acpi
Revision history for this message
penalvch (penalvch) wrote :

Original Reporter account deactivated.

summary: - LSI 1068E (including LSI SAS 3442E-R and LSI 3081E-R) instabilities
+ LSI 1068E instabilities
Changed in linux (Ubuntu):
status: Confirmed → Invalid
To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.