Chip type |
Server or HBA card |
1. LSI1030 |
V440 |
3. LSI1064 |
Sun Fire V215 [TM] Server, Sun Fire V245, Sun Fire T1000 |
4. LSI1064E |
Sun Fire T2000, Sun Fire Netra T2000 |
2. LSI1068 |
V445 |
5. LSI1068E |
Sun Fire T5120/T5220/T5140/T5240/T5440/NetraT5220/NetraT5440 |
6. LSI2008 |
T3-1/T3-2/T3-3/T4-1/T4-2/T4-4 |
7. LSI2108 |
Niwot card |
1. The first generation hard raid Chip LSI1030
Applies to: Sun Fire V440 [TM] Server, Netra[TM] 440 Server
Chip: LSI1030 SCSI Host Adapter
- Determine the type of the hard raid HBA
# iostat -En
e.g.
c1t0d0 LSILOGIC Product: 1030
# cfgadm -alv
e.g. c1t0d0 LSILOGIC 1030 /devices/pci@1f,700000/scsi@2:scsi::dsk/c1t0d0
# prtdiag -v
e.g. scsi-pci1000,30 LSI,1030 /pci@1f,700000/scsi@2
# prtconf -vp
e.g.
LSI1030 FCode Driver: 1.8 03/04/17 scsi: '/pci@1f,700000/scsi@2'firmware-version: '1.03.11.01'
# prtpicl -v
e.g.
LSI1030 FCode Driver: 1.8 03/04/17 devfs-path /pci@1f,700000/scsi@2
- FW or patch for LSI1030 CHIP:
Remember to install the right and best OBP version, because OBP contains the proper driver to communicate with FCode on HBA.
Current firmware/FCode:
lsi1030-fw-1.03.11.01
Above codes can be reached either by factory default implementation, proper OBP-upgrade or by use of following patches
lsi1030-fw-1.03.11.01
patch 115662-01
Hardware/Firmware: LSI1030 U320 SCSI controller
- The latest patch release notes:
Keywords: mpt raidctl firmware lsi1030 sun fire v440 v440
Synopsis: Hardware/Firmware: LSI1030 U320 SCSI controller
Date: Sep/24/2003
Install Requirements: See Special Install Instructions
Additional instructions may be listed below
Solaris Release: 8 9
SunOS Release: 5.8 5.9
Unbundled Product: LSI1030 Firmware
Unbundled Release: N/A
NOTE: Sun Fire V440 onboard LSI1030 controller
Xref:
Topic:
Relevant Architectures: sparc
BugId's fixed with this patch: 4834770 4924126
Changes incorporated in this version: 4924126 4834770
Patches accumulated and obsoleted by this patch:
Patches which conflict with this patch:
Patches required with this patch:
Obsoleted by:
Files included with this patch:
lsi1030-fw-1.03.11.01
Problem Description:
4924126 Connected
Command timeouts seen on Seagate disks running stress tests
4834770 1030 firmware does not clean up IOs to a RAID after a bus reset
Patch Installation Instructions:
1) Copy the patch into
any directory, say
(Note, if the patch ends in a
"tar.Z" suffix, you will need to first
uncompress and untar it.) So, you would type:
% su root
# cd
# uncompress
# tar xf
2) The LSI1030 uses the
'mpt' driver. Run the following 'grep'
command
to find all device paths which this
driver is associated with.
# grep "mpt"
/etc/path_to_inst
"/pci@1f,700000/scsi@2" 0
"mpt"
"/pci@1f,700000/scsi@2,1" 1 "mpt"
3) Run the 'format'
command from the prompt to see what disks you have
in the system that are associated with
these device paths:
# format
Searching for disks...done
AVAILABLE DISK
SELECTIONS:
0. c1t0d0
/pci@1f,700000/scsi@2/sd@0,0
1. c1t1d0
/pci@1f,700000/scsi@2/sd@1,0
2. c1t2d0
/pci@1f,700000/scsi@2/sd@2,0
3. c1t3d0
/pci@1f,700000/scsi@2/sd@3,0
Specify disk (enter its number):
Determine all disks
that are associated with all the device paths
identified in step 2. In this case the
disks on the
"/pci@1f,700000/scsi@2/" LSI1030 mpt driver path are assigned
controller
number 'c1'. This will be controller we
will update using just the
number '1'.
Type ^D to quit out of 'format'
4) Update the firmware on your controller.
Use the '/usr/sbin/raidctl' utility and type:
# /usr/sbin/raidctl -F lsi1030-fw-1.03.11.01 1
'1' in this case is your controller id.
This will ask you if you want to continue updating the firmware.
ROM image contains MPT
firmware version 1.03.11 (w/Integrated Mirroring)
Update flash image on controller 1 (yes/no)?
Type 'yes' and hit enter.
Once you are returned to the prompt the update is complete.
5) Note that the
"/pci@1f,700000/scsi@2,1" instance is a 2nd channel on
the same LSI1030 controller identified by the ",1" portion of the
device
path being appended to the 1st channel path "scsi@2". This does not
need to be explicitly upgraded if disks attached to the 1st channel of
the LSI1030 are already upgraded in step 4.
Repeat steps 3 and 4
for each unique LSI1030 controller identified in
step 2.
6) Reboot the system after all the controllers have been updated.
Special Install Instructions:
-----------------------------
It is preferable that this update be done when disk activity is a minimum.
The system must be
rebooted for the new firmware version to appear in prtconf.
After reboot the firmware version can be verified as follows:
# /usr/sbin/prtconf -vp
| grep 'firmware-version:'
firmware-version: '1.03.11.01'
- The administration guides for LSI1030:
Description
This document provides the procedure to create or delete a hardware raid
mirror in a Sun Fire[TM] V440. Support engineers should check the procedures
in this document when assisting customers who may be asking for
guidance in this process.
Steps to Follow
How to Create a Hardware Disk Mirror
Note:
To create a raid volume, the disk being mirrored must have a t#
lower than the t# of the mirror disk. This requirement is due to the
interaction between the way Solaris[TM] handles the boot disk and the
built-in feature of the hardware raid controller that cannot be
disabled.
Caution Creating or restoring a disk mirror destroys all data previously stored on the
disk drive.
Perform this procedure to create an internal hardware disk mirror configuration on
your system.
Verify which disk drive corresponds with which logical device name and physical
device name. Example: Type
# raidctl
No RAID volumes found.
The example above indicates that no RAID volume exists. In another case:
# raidctl
RAID RAID RAID Disk
Volume Status Disk Status
----------------------------------------------------------------------------------
c1t1d0 DEGRADED c1t1d0 OK
c1t2d0 DEGRADED
The example above indicates a hardware mirror has degraded at disk c1t2d0.
-------------------------------------------------------------------------------------------------------------
Note - The logical device names might appear differently on your system, depending
on the number and type of add-on disk controllers installed.
What to Do
1. Type the following command: # raidctl -c master slave
For example: # raidctl -c c1t0d0 c1t1d0
When you create a RAID mirror, the slave drive (in this case, c1t1d0) disappears from
the Solaris device tree.
2. To check the status of a RAID mirror, type the following command: # raidctl
RAID RAID RAID Disk
Volume Status Disk Status -----------------------------------------------------------------------
c1t0d0 RESYNCING c1t0d0 OK
c1t1d0 OK
The example above indicates that the RAID mirror is still resynchronizing with the
backup drive.
The example below shows that the RAID mirror is completely restored and online.
# raidctl
RAID RAID RAID Disk
Volume Status Disk Status
c1t0d0 OK c1t0d0 OK
c1t1d0 OK
How to Delete a Hardware Disk Mirror
Perform this procedure to remove a hardware disk mirror configuration from your
system.
Verify which disk drive corresponds with which logical device name and physical
device name.
Disk Slot Numbers, Logical Device Names, and Physical Device Names
Disk Slot Number Logical Device Name* Physical Device Name
Slot 0 c1t0d0 /devices/pci@1f,700000/scsi@2/sd@0,0
Slot 1 c1t1d0 /devices/pci@1f,700000/scsi@2/sd@1,0
Slot 2 c1t2d0 /devices/pci@1f,700000/scsi@2/sd@2,0
Slot 3 c1t3d0 / /devices/pci@1f,700000/scsi@2/sd@3,0
______________________________________________________________________
*The logical device names might appear differently on your system, depending on the
number and type of add-on disk controllers installed.
- Determine the name of the mirrored volume. Type the following command:
# raidctl
RAID RAID RAID Disk
Volume Status Disk Status
c1t0d0 OK c1t0d0 OK
c1t1d0 OK
In this example, the mirrored volume is c1t0d0.
Note The logical device names might appear differently on your system,
depending on the number and type of add-on disk controllers installed.
- To delete the volume, type the following command: # raidctl -d mirrored-volume
For example: # raidctl -d c1t0d0
RAID Volume c1t0d0 deleted
- To confirm that you have deleted the RAID array, type the following command:
# raidctl
For example: # raidctl
No RAID volumes found
Note:
LSI1030 HBA supports only 1 RAID volume (IM, Integrated Mirror) on V440
H/W mirroring can be initiated after installation of data onto physical disk.
Backwards, breaking the mirror will give two disks with valid data on each.
2. The first generation hard raid Chip LSI1068
Applies to: Sun Fire V445 [TM] Server
Chip: LSI SAS1068 controller is mounted on PCI-X slot on V445 for up to 8 disks
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd1 Soft Errors: 2 Hard Errors: 0 Transport Errors: 0
Vendor: LSILOGIC Product: Logical Volume Revision: 3000 Serial No:
Size: 73.01GB <73012215808 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
# cfgadm -alv
e.g.
c1 connected configured unknown
unavailable scsi-bus n /devices/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1:scsi
c1::dsk/c1t0d0 connected configured unknown LSILOGIC Logical Volume
unavailable disk n /devices/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1:scsi::dsk/c1t0d0
# prtdiag -v
e.g.
pci 199 MB/PCI4 LSILogic,sas-pci1000,54 (scs+ LSI,1068
okay /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1
# prtconf -vp
e.g.
firmware-version: '1.08.04.00'
mpt-version: '1.05'
assigned-addresses: 81080810.00000000.00002000.00000000.00000100.82080814.00000000.04600000.00000000.00010000.8208081c.00000000.04610000.00000000.00010000.82080830.00000000.04800000.00000000.00200000
compatible: 'pci13e9,54' + 'pci1000,54' + 'AAPL,MPT'
model: 'LSI,1068'
reg: 00080800.00000000.00000000.00000000.00000000.01080810.00000000.00000000.00000000.00000100.03080814.00000000.00000000.00000000.00010000.0308081c.00000000.00000000.00000000.00010000.02080830.00000000.00000000.00000000.00200000
version: '1.00.39'
wide: 00000010
device_type: 'scsi-2'
name: 'LSILogic,sas'
# prtpicl -v
e.g.
:model LSI,1068
:compatible (a30000093fTBL)
| pci13e9,54 |
| pci1000,54 |
| AAPL,MPT |
:assigned-addresses
81 08 08 10 00 00 00 00 00 00 20 00 00 00 00 00 00 00 01 00
82 08 08 14 00 00 00 00 04 60 00 00 00 00 00 00 00 01 00 00
82 08 08 1c 00 00 00 00 04 61 00 00 00 00 00 00 00 01 00 00
82 08 08 30 00 00 00 00 04 80 00 00 00 00 00 00 00 20 00 00
:mpt-version 1.05
:pcix-capid-pointer 0x68
:ddi-intr-weight 0xa
:mpxio-disable yes
:ddi-vhci-class scsi_vhci
:fm-ereport-capable
:fm-errcb-capable
:fm-dmachk-capable
:fm-accchk-capable
:pm-components (a300000930TBL)
| NAME=mpt0 |
| 0=Off (PCI D3 State) |
| 3=On (PCI D0 State) |
:firmware-version 1.8.4.0
- FW or patch for LSI1068 CHIP:
sas1068 on V445
123485-01
Hardware/PROM Sun Fire V445 1068X update 1.0.39
In general HBA fcodes are part of obp code on machines with on-board HBAs.
Every OBP- or FCode upgrade requires a power-cyle, not just a reboot.
- The latest patch release notes:
Patch-ID# 123485-01
NOTE:
***********************************************************************
READ THE TERMS OF THE AGREEMENT ("AGREEMENT") IN THE LEGAL_LICENSE.TXT
FILE CAREFULLY BEFORE USING THIS SOFTWARE. BY USING THE SOFTWARE, YOU
AGREE TO THE TERMS OF THIS AGREEMENT. IF YOU DO NOT AGREE TO ALL OF THE
TERMS, PROMPTLY DESTROY THE UNUSED SOFTWARE.
***********************************************************************
Keywords: sun fire v445 1068x_firmware
Synopsis: Hardware/PROM Sun Fire V445 1068X update 1.0.39
Date: Feb/02/2007
Install Requirements: Additional instructions may be listed below
Solaris Release: 9 10
SunOS Release: 5.9 5.10
Unbundled Product: Sun Fire Hardware 1068X PROM
Unbundled Release: 1.0.39
Xref:
Topic: Sun Fire V445 1068X FCode update 1.0.39
Relevant Architectures: sparc
NOTE:
Adds RAID Migration Support to 1068X FCode by enabling the activate-volume function.
NOTE:
The 1068X SAS Controller is used as the embedded disk controller on the V445. A small number of systems were shipped with 1068X FCode at revisions below 1.0.39. If you are using the embedded 1068X to control hardware RAID volumes and the FCode is below revision 1.0.39 the activate-volume function is not available. The activate-volume function is needed if the RAID Set is moved or the 1068X controller replaced. Please refer to your V445 Server Administration Guide for information on using the 1068X to create and control hardware RAID volumes.
BugId's fixed with this patch: 6518779
Changes incorporated in this version: 6518779
Patches accumulated and obsoleted by this patch:
Patches which conflict with this patch:
Patches required with this patch:
Obsoleted by:
Files included with this patch:
Install.info
LSI1068X.ROM
Problem Description:
6518779 LSI1068X Does not support Activate Volume
Patch Installation Instructions:
==============================
Please refer to the Install.info file for instructions on updating
the 1068X FCode using the files included in this patch.
Special Install Instructions:
-----------------------------
Unzip the patch into a directory and then follow the instructions provided
in the file Install.info.
README -- Last modified date: Friday, January 26, 2007
README -- Last modified date: Friday, February 2, 2007
Patch Installation Instructions:
===============================
ABOUT THIS PATCH
-------------------------
This patch updates the PCI-X 1068 FCode:
To version 1.0.39
TO CHECK THE VERSION OF THE Firmware
--------------------------------------------------
There are two ways to check the PCI-X Firmware version:
o A. From the Solaris Operating System (OS)
o B. From the OpenBoot PROM (OBP)
A. From the Solaris OS (2 steps):
The system prompt is shown as bash-3.00#, your prompt will vary depending on
your individual setup and shell in use.
Step 1. As root, run the prtconf command.
bash-3.00# prtconf -vp | egrep '(model|version)' |\
> egrep -vi '(obp|firmwareversion| mpt-version)'
model: 'SUNW,501-7066'
model: 'SUNW,4.22.11'
sparc-version: 00000009
sparc-version: 00000009
sparc-version: 00000009
sparc-version: 00000009
model: 'm5823'
firmware-version: '1.08.04.00'
model: 'LSI,1068'
version: '1.00.39'
model: 'SUNW,525-2274'
bash-3.00#
Step 2: Check the FCode Version
Locate the lines model: followed by version:, the version data is the FCode
version. In the example shown the version is 1.00.39 and no update is required.
B. From OBP (4 steps):
Check the version as follows, the ok prompt is shown as {1} ok in this
example, your prompt may vary depending on the number of processors installed:
Step 1: Select the LSI1068X Controller
{1} ok select /pci@1e,600000/pci@0/pci@1/pci@0/pci@8/LSILogic,sas@1
Step 2: Display information from the LSI1068X controller
{1} ok .properties
firmware-version 1.08.04.00
mpt-version 1.05
assigned-addresses 81060810 00000000 00000000 00000000 00000100
82060814 00000000 00800000 00000000 00010000
8206081c 00000000 00810000 00000000 00010000
82060830 00000000 00a00000 00000000 00200000
compatible pci13e9,54
pci1000,54
AAPL,MPT
model LSI,1068
reg 00060800 00000000 00000000 00000000 00000000
01060810 00000000 00000000 00000000 00000100
03060814 00000000 00000000 00000000 00010000
0306081c 00000000 00000000 00000000 00010000
02060830 00000000 00000000 00000000 00200000
version 1.00.39
wide 00000010
device_type scsi-2
name LSILogic,sas
fcode-rom-offset 00000000
66mhz-capable
devsel-speed 00000001
latency-timer 00000040
cache-line-size 00000010
max-latency 0000000a
min-grant 00000040
interrupts 00000001
class-code 00010000
subsystem-id 00003031
subsystem-vendor-id 00001000
revision-id 00000001
device-id 00000054
vendor-id 00001000
{1} ok
Step 3: Check the FCode Version
Locate the line starting with version, this is the FCode version. In the
example shown the version is 1.00.39 and no update is required.
Step 4: Un-select the node before booting the server.
{1} ok unselect-dev
{1} ok
TO UPDATE THE 1068X Firmware
------------------------------
Step 1: Boot the Solaris OS.
Step 2: Login as root.
Step 3: Copy the new 1064X Firmware image onto the system.
The new 1064 Firmware image is included in this patch and is named
LSI1068XROM. Copy the image into a directory on the system which you
are going to update or to a network accessible directory. Use this
full path along with the image name in step 5) below.
Step 4. Identify the controller.
Determine the controller number via the format command. In the example
shown the controller is number 0. Exit format with ^C
bash-3.00# format
Searching for disks...done
AVAILABLE DISK SELECTIONS:
0. c0t0d0
/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@0,0
1. c0t2d0
/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@2,0
2. c0t5d0
/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@5,0
3. c0t6d0
/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@6,0
4. c0t7d0
/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@7,0
Specify disk (enter its number): ^C
bash-3.00#
Step 5: Update the Firmware using the raidctl command.
Syntax:
# raidctl -F [full_path_of_image] [controller]
Example:
bash-3.00# raidctl -F LSI1068X.ROM 0
bash-3.00#
Step 6: Reboot the system to use the new FCode
At this point the New FCode is loaded into the 1068X flash device but
the system is still running off of the old image.
Reboot the server to load the new 1064 Firmware.
Syntax:
bash-3.00# reboot
- The administration guides for LSI1068:
LSI 1068 HBA Supported RAID Operations - Support for RAID0 (striping) and RAID1 (mirroring)
RAID1 - mirroring
- A 2-disk primary->secondary configuration
- All write I/Os are committed to both disks before returning to provide for data redundancy
- Read I/Os may be serviced by either disk
- Upon a single disk failure, the data is still available - the volume is in DEGRADED mode. The failed disk may be replaced and resync'd
- If the volume is disabled (mirror is broken) the drives need to be relabeled and all data is lost.
- When creating a RAID volume all data contained on the drives is lost. If mirroring of the boot drive is desired the RAID volume must be created before OS installation.
RAID0 - striping
- A 2 or more disk configuration
- No data is shared. Data is across all disks (concatenated) in exactly one location.
- Write I/Os are committed in a round-robin fashion across all member disks, performance increased in some workload scenarios
- Read I/Os are satisfied at their one location (performs as a single disk)
- Upon a single disk failure, data is lost
- If a RAID0 volume is disabled (broken), data is lost
- When creating a RAID volume all data contained on the drives is lost.
Management of RAID volumes by use of the Solaris raidctl command
- Create a mirror
raidctl -c [-r 1] primary secondary - Create a stripe
raidctl -c -r 0 disk1 disk2 [disk3] ... - Delete a RAID volume
raidctl -d volume - Display RAID volumes
raidctl - Update HBA controller
firmware
raidctl -F file controller
Examples
1.) Creating a mirror
- To create a RAID1 mirror
raidctl -c [-r 1] primary secondary
Example:
# raidctl -c c0t0d0 c0t1d0
- RAID volume c0t0d0 will be created
- Data on all drives will be lost
- No file systems may be mounted on either disk
- Secondary must be as large as primary
- RAID volume has the capacity of one disk
2.) Creating a stripe
- To create a RAID0 stripe
raidctl -c -r 0 disk1 disk2 [disk3] ..
Example:
# raidctl -c -r 0 c0t1d0 c0t2d0 c0t3d0
- RAID volume c0t1d0 will be created
- Data on all drives will be lost
- No file systems may be mounted on any disk
- RAID volume has the capacity of
all disks
combined - smallest common size is used
After RAID volume creation
- Member disk drives are no longer visible, only the RAID volume is presented
- The RAID volume may be addressed as any other LUN in Solaris
- The status of the RAID volume and it's member disks may be retrieved using raidctl(1M)
- Each LSI1064 HBA supports up to 2 RAID volumes at any one time
- Volumes must be labeled after creation before they can be used under Solaris
Labeling Volumes After Creation
After creating a RAID volume by using 'raidctl' or by using OBP level commands the volume must be labeled correctly before it can be used. This can be done using the 'format' command in Solaris:
As the root user run 'format' and choose the newly created volume from the list of drives.
(you may see some messages about the volume having a corrupt label or the block count being incorrect when booting or running format, this is due to the need for a correct label)
After the volume is selected chose "type" from the format menu, this will give you a list of available drive types to chose from.
Chose "0. Auto configure" from the list. This will define the volume with the correct type.
Then write the new label to the volume using the "label" command from the format> prompt. The system will ask you if you are ready to label the disk. type in "yes"
After you perform these steps the drive is ready for use. You can now partition the drive as needed and use the volume within Solaris normally.
Note also that after a RAID volume is deleted the drives that were previously part of the RAID volume should be re-labeled at that time for proper use also.
RAID Volumes Possible Status
OK - volume is optimal
DEGRADED - volume has a failed disk (mirror) or is otherwise functionally minimized
RESYNCING - volume is synchronizing (mirror)
FAILED - volume is offline
Disk Possible Status
OK - disk is operating normally
MISSING - disk is removed or otherwise offline
FAILED - disk is in a failed state and needs service
3. The first generation hard raid Chip LSI1064
Applies to: Sun Fire V215 [TM] Server, Sun Fire V245, Sun Fire T1000
Chip: LSI1064 SCSI Host Adapter
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd1 Soft Errors: 2 Hard Errors: 0 Transport Errors: 1
Vendor: LSILOGIC Product: Logical Volume Revision: 3000 Serial No:
Size: 146.56GB <146561286144 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
# cfgadm -alv
e.g.
c1::dsk/c1t0d0 connected configured unknown LSILOGIC Logical Volume
unavailable disk n /devices/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1:scsi::dsk/c1t0d0
# prtdiag -v
e.g.
pci 188 MB scsi-pci1000,50 (scsi-2) LSI,1064
okay /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
# prtconf -vp
e.g.
firmware-version: '1.08.04.00'
mpt-version: '1.05'
assigned-addresses: 810d0810.00000000.00002000.00000000.00000100.820d0814.00000000.04c00000.00000000.00010000.820d081c.00000000.04c10000.00000000.00010000.820d0830.00000000.04e00000.00000000.00200000
compatible: 'pci1000,50.1000.3020.2' + 'pci1000,50.1000.3020' + 'pci1000,3020' + 'pci1000,50.2' + 'pci1000,50' + 'pciclass,010000' + 'pciclass,0100'
model: 'LSI,1064'
# prtpicl -v
e.g.
:model LSI,1064
- FW or patch for LSI1068 CHIP:
SAS1064 V1.18.05.00IR on V215 /V245
136762-01
LSI SAS1064/1068 HBA supports up to 2 RAID volumes of same RAID-level for machines with more than two harddisks.
Just one RAID is supported on two-disk machines T1000, V215, Ultra25.
sas1064 on T1000
123456-01
Hardware/PROM Sun Fire T1000 1064 Firmware update V1.10.00.00IR
- The latest patch release notes:
Keywords: sun fire v245 1064_firmware
Synopsis: Hardware/PROM Sun Fire V245 1064 Firmware update V1.18.05.00IR
Date: Dec/13/2007
Install Requirements: Additional instructions may be listed below
Solaris Release: 10
SunOS Release: 5.10
Unbundled Product: Hardware/PROM
Unbundled Release: 1.18.05.00IR
Xref:
Topic: Sun Fire V245 1064 Firmware update V1.18.05.00IR
Relevant Architectures: sparc
Note:
Patch 119850-22 is required for S10U1. This patch is included in S10U2 and greater.
BugId's fixed with this patch: 6531216
Changes incorporated in this version: 6531216
Patches accumulated and obsoleted by this patch:
Patches which conflict with this patch:
Patches required with this patch: 119850-22 (or greater)
Obsoleted by:
Files included with this patch:
1.18.05-1064-Seattle.fw
Problem Description:
6531216 When replaced the mother board of the mirror system V215/V245, resyncing time were about 30hours.
Patch Installation Instructions:
===============================
ABOUT THIS PATCH
-------------------------
This patch updates the PCI-X 1064 Firmware on the Sun Fire v215/v245
to version:
1.18.05.00
TO CHECK THE VERSION OF THE FIRMWARE
--------------------------------------------------
There are two ways to check the PCI-X Firmware version:
o From the Solaris Operating System (OS)
o From the OpenBoot PROM (OBP)
From the Solaris OS:
1. As root, run the prtconf command.
Example:
# prtconf -vp | egrep '(model|version)' | \
> egrep -vi '(obp|sparc-version|mpt-version)'
model: 'SUNW,375-3463'
model: 'SUNW,4.25.10'
model: 'm5823'
firmware-version: '1.08.04.00'
model: 'LSI,1064'
version: '1.00.40'
model: 'SUNW,525-2320'
version: '5.10'
#
In this example, the 1064 PCI-X Firmware is version 1.08.04.00, and needs
to be updated.
From the OBP:
1. Identify the node for the PCI-X 1064:
Example:
{0} ok show-disks
a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk <----
b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom
c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk
q) NO SELECTION
Enter Selection, q to quit: q
{0} ok
In this example on the Sun Fire V245, "a" represents
the 1064 node and the disk underneath the node.
Once you identify the node in this list, type q to return
to the ok prompt.
2. Select the node.
Example:
{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
{0} ok
In this example, only the full path to the node is specified
(the path does not include the "disk").
3. Display the node properties.
Example:
{0} ok .properties
firmware-version 1.08.04.00 <----
mpt-version 1.05
assigned-addresses 810d0810 00000000 00002000 00000000 00000100
820d0814 00000000 04c00000 00000000 00010000
820d081c 00000000 04c10000 00000000 00010000
820d0830 00000000 04e00000 00000000 00200000
compatible pci1000,50.1000.3020.2
pci1000,50.1000.3020
pci1000,3020
pci1000,50.2
pci1000,50
pciclass,010000
pciclass,0100
model LSI,1064
reg 000d0800 00000000 00000000 00000000 00000000
010d0810 00000000 00000000 00000000 00000100
030d0814 00000000 00000000 00000000 00010000
030d081c 00000000 00000000 00000000 00010000
020d0830 00000000 00000000 00000000 00200000
version 1.00.40
wide 00000010
device_type scsi-2
name scsi
local-wwid 50 80 02 00 00 00 29 99
devsel-speed 00000001
latency-timer 000000f8
max-latency 0000000a
min-grant 00000040
interrupts 00000001
cache-line-size 00000080
class-code 00010000
subsystem-id 00003020
subsystem-vendor-id 00001000
revision-id 00000002
device-id 00000050
vendor-id 00001000
{0} ok
In this example, the 1064 Firmware is version 1.08.04.00, and needs
to be updated.
4. UN-select the node before booting the server.
Example:
{0}ok unselect-dev
{0}ok
TO UPDATE THE 1064 Firmware
------------------------------
1. Boot the Solaris OS.
2. Login as root.
3. Copy the new 1064 Firmware image onto the system.
The new 1064 Firmware image is included in this patch. Copy the image
into a dir on the system which you are going to update or to
a network accessible dir. Use this full path along with the
image name in step 5) below.
4. Identify the controller.
In the following example, the controller contains the same path as shown
above in the OBP example:
# ls -al /dev/cfg
total 30
drwxr-xr-x 5 root root 512 Sep 25 13:03 .
drwxr-xr-x 17 root sys 7168 Dec 4 16:33 ..
lrwxrwxrwx 1 root root 63 Sep 21 09:51 c0 -> ../../devices/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1:scsi
lrwxrwxrwx 1 root root 57 Sep 25 13:03 pxb_plx5.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@1:pcie0
lrwxrwxrwx 1 root root 57 Sep 25 13:03 pxb_plx6.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@2:pcie0
lrwxrwxrwx 1 root root 57 Sep 25 13:03 pxb_plx7.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@3:pcie0
drwxr-xr-x 2 root root 512 Sep 21 09:51 usb0
drwxr-xr-x 2 root root 512 Sep 21 09:51 usb1
drwxr-xr-x 2 root root 512 Sep 21 09:51 usb2
#
The controller in the above example is 0.
5. Update the firmware using the raidctl command.
Syntax:
# raidctl -F [full_path_of_image] [controller]
Example:
# raidctl -F 1.18.05-1064-Seattle.fw 0
ROM image contains MPT firmware version 1.18.05.00
Update flash image on controller 0 (yes/no)? yes
Flash updated successfully.
#
After the flash update is completed, the Solaris driver resets the 1064 to
put it back into an operational state. The following error(s) are expected
as a result of this reset:
Example:
Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.warning] WARNING: /pci@7c0/pci@0/
pci@8/scsi@2/sd@0,0 (sd0):
Jun 14 08:03:16 wgs57-68 Error for Command: write Error
Level: Retryable
Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice] Requested Block: 11375
Error Block: 11375
Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice] Vendor: SEAGATE
Serial Number: 0517104C7L
Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice] Sense Key: Unit_Attention
Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice] ASC: 0x29 (scsi bus reset
occurred), ASCQ: 0x2, FRU: 0x2
6. At this point the New Firmware image is loaded into the 1064's flash device but
the system is still running off of the old image.
Reboot the server to load the new 1064 Firmware.
Syntax:
# reboot
*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*
Special Install Instructions:
-----------------------------
Note:
Patch 119850-14 (or Greater) is required for S10U1. Patch 119850-14 is included in S10U2.
- The administration guides for LSI1064:
Hardware - LSI SAS1064 4 Phy-wide SAS HBA built into the motherboard.
The kernel module is "mpt".
Each LSI1064 HBA supports up to 2 RAID volumes at any one time.
Supported RAID Operations - Support for RAID0 (striping) and RAID1 (mirroring)
RAID1 - mirroring
- A 2-disk primary->secondary configuration
- All write I/Os are committed to both disks before returning to provide for data redundancy
- Read I/Os may be serviced by either disk
- Upon a single disk failure, the data is still available - the volume is in DEGRADED mode. The failed disk may be replaced and resync'd
- If the volume is disabled (mirror is broken) the drives need to be relabeled and all data is lost.
- When creating a RAID volume all data contained on the drives is lost. If mirroring of the boot drive is desired the RAID volume must be created before OS installation.
RAID0 - striping
- A 2 or more disk configuration
- No data is shared. Data is across all disks (concatenated) in exactly one location.
- Write I/Os are committed in a round-robin fashion across all member disks, performance increased in some workload scenarios
- Read I/Os are satisfied at their one location (performs as a single disk)
- Upon a single disk failure, data is lost
- If a RAID0 volume is disabled (broken), data is lost
- When creating a RAID volume all data contained on the drives is lost.
Management of RAID volumes by use of the Solaris raidctl command
- Create a mirror
raidctl -c [-r 1] primary secondary
- Create a stripe
raidctl -c -r 0 disk1 disk2 [disk3] ...
- Delete a RAID volume
raidctl -d volume - Display RAID volumes
raidctl - Update HBA controller
firmware
raidctl -F file controller
Examples
1.) Creating a mirror
- To create a RAID1 mirror
raidctl -c [-r 1] primary secondary
Example:
# raidctl -c c0t0d0 c0t1d0
- RAID volume c0t0d0 will be created
- Data on all drives will be lost
- No file systems may be mounted on either disk
- Secondary must be as large as primary
- RAID volume has the capacity of one disk
2.) Creating a stripe
- To create a RAID0 stripe
raidctl -c -r 0 disk1 disk2 [disk3] ..
Example:
# raidctl -c -r 0 c0t1d0 c0t2d0 c0t3d0
- RAID volume c0t1d0 will be created
- Data on all drives will be lost
- No file systems may be mounted on any disk
- RAID volume has the capacity of
all disks
combined - smallest common size is used
After RAID volume creation
- Member disk drives are no longer visible, only the RAID volume is presented
- The RAID volume may be addressed as any other LUN in Solaris
- The status of the RAID volume and it's member disks may be retrieved using raidctl(1M)
- Each LSI1064 HBA supports up to 2 RAID volumes at any one time
- Volumes must be labeled after creation before they can be used under Solaris
Labeling Volumes After Creation
After creating a RAID volume by using 'raidctl' or by using OBP level commands the volume must be labeled correctly before it can be used. This can be done using the 'format' command in Solaris:
As the root user run 'format' and choose the newly created volume from the list of drives.
(you may see some messages about the volume having a corrupt label or the block count being incorrect when booting or running format, this is due to the need for a correct label)
After the volume is selected chose "type" from the format menu, this will give you a list of available drive types to chose from.
Chose "0. Auto configure" from the list. This will define the volume with the correct type.
Then write the new label to the volume using the "label" command from the format> prompt. The system will ask you if you are ready to label the disk. type in "yes"
After you perform these steps the drive is ready for use. You can now partition the drive as needed and use the volume within Solaris normally.
Note also that after a RAID volume is deleted the drives that were previously part of the RAID volume should be re-labeled at that time for proper use also.
RAID Volumes Possible Status
OK - volume is optimal
DEGRADED - volume has a failed disk (mirror) or is otherwise functionally minimized
RESYNCING - volume is synchronizing (mirror)
FAILED - volume is offline
Disk Possible Status
OK - disk is operating normally
MISSING - disk is removed or otherwise offline
FAILED - disk is in a failed state and needs service
OBP Level Commands
PLEASE NOTE: OBP Level RAID commands are NOT SUPPORTED and are not intended for normal customer use. Future firmware updates may change command behavior and/or remove access to these commands altogether.
Before running the RAID OBP commands the auto-boot? variable needs to be set to false and the fcode-debug? variable needs to be set to true, then a reset-all is performed. After the reset finishes the controller must be selected. In the example below the "show-disks" command is used to determine the controller path. Note that the "/disk" was left off the path since it is the controller you want to select. (You can also use the probe-scsi-all command to find the path.)
{1} ok setenv auto-boot? false
auto-boot? = false
{1} ok setenv fcode-debug? true
fcode-debug? = true
{1} ok reset-all
:
:
:
{1} ok show-disks
a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk
b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom
c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk
q) NO SELECTION
Enter Selection, q to quit: a
/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk has been selected.
Type ^Y ( Control-Y ) to insert it in the command line.
e.g. ok nvalias mydev ^Y
for creating devalias mydev for /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk
{1} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
The "show-volumes" command will show if there are any existing volumes and their state.
{1} ok show-volumes
Volume 0 Target 0 Type IM (Integrated Mirroring)
Optimal Enabled
2 Members 143112591 Blocks, 73 GB
Disk 1
Primary Online
Target 4 FUJITSU MAY2073RCSUN72G 0401
Disk 0
Secondary Online
Target 1 FUJITSU MAY2073RCSUN72G 0401
{1} ok
If you wanted to create a volume you would select the controller as above and use the "create" command. The following example creates a mirrored volume with disks 0 and 1.
{0} ok 0 1 create-im-volume
Target 0 size is 143243264 Blocks, 73 GB
Target 1 size is 143243264 Blocks, 73 GB
The volume can be any size from 1 MB to 69943 MB
What size do you want? [69943]
Volume size will be 143243264 Blocks, 73 GB
PhysDisk 0 has been created for target 0
PhysDisk 1 has been created for target 1
Volume has been created
{0} ok
This command creates a striped volume using disks 0 and 1
{0} ok 0 1 create-is-volume
Target 0 size is 143243264 Blocks, 73 GB
Target 1 size is 143243264 Blocks, 73 GB
The volume can be any size from 1 MB to 139886 MB
What size do you want? [139886]
Volume size will be 286486528 Blocks, 146 GB
PhysDisk 0 has been created for target 0
PhysDisk 1 has been created for target 1
Volume has been created
{0} ok
To unselect
{0} ok unselect-dev
when you have completed managing the RAID volumes you can set the auto-boot? and fcode-debug? variables back and reset the system:
{1} ok setenv auto-boot? true
auto-boot? = true
{1} ok setenv fcode-debug? false
fcode-debug? = false
{1} ok reset-all
Remember that even if you create a RAID volume at the OBP level you will still need to label the volume using format in Solaris before they can be properly used.
If you have existing RAID volumes and have to replace the system board the
following commands will need to be run to see the volume with the new board
{1} ok show-disks
a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk
b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom
c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk
q) NO SELECTION
Enter Selection, q to quit: q
/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk has been selected.
Type ^Y ( Control-Y ) to insert it in the command line.
e.g. ok nvalias mydev ^Y
for creating devalias mydev for /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk
{1} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
{0} ok show-volumes ( Look for the inactive volume)
{0} ok X activate-volume (X is the volume number)
{0} ok unselect-dev
If activating multiple volumes ensure they are activated in descending order:
{0} ok show-volumes
{0} ok 1 activate-volume
{0} ok 0 activate-volume
{0} ok unselect-dev
{0} ok probe-scsi-all
Failure to do so will result in the following error and will require a reset-all prior to activating the second volume:
{0} ok 0 activate-volume
Volume 0 is now activated
{0} ok 1 activate-volume
RAID volume request failed, ioc-status = 0007
Volume 1 could not be activate
The first volume activated will remain 'Degraded' until both volumes have completed syncing.
NOTE: IF the EEPROM variable "auto-boot?" was set to true you may get the following message.
{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
Base SAS World Wide ID is 0!
This must be fixed immediately using set-sas-wwid
{0} ok
If you do see this message set "auto-boot?" to false and issue a "reset-all"
4. The first generation hard raid Chip LSI1064E
Applies to: Sun Fire T2000, Sun Fire Netra T2000
Chip: LSI1064E SCSI Host Adapter
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
Vendor: LSILOGIC Product: Logical Volume Revision: 3000 Serial No:
Size: 73.34GB <73340551168 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
# cfgadm -alv
e.g.
Ap_Id Receptacle Occupant Condition Information
When Type Busy Phys_Id
c0 connected configured unknown
unavailable scsi-bus n /devices/pci@780/pci@0/pci@9/scsi@0:scsi
c0::dsk/c0t0d0 connected configured unknown LSILOGIC Logical Volume
unavailable disk n /devices/pci@780/pci@0/pci@9/scsi@0:scsi::dsk/c0t0d0
# prtdiag -v
e.g.
========================= IO Configuration =========================
IO
Location Type Slot Path Name Model
----------- ----- ---- --------------------------------------------- ------------------------- ---------
IOBD/NET0 PCIE IOBD /pci@780/pci@0/pci@1/network@0 network-pciex8086,105e
IOBD/NET1 PCIE IOBD /pci@780/pci@0/pci@1/network@0,1 network-pciex8086,105e
IOBD/PCIE PCIE IOBD /pci@780/pci@0/pci@9/scsi@0 scsi-pciex1000,56 LSI,1064E
IOBD/NET2 PCIE IOBD /pci@7c0/pci@0/pci@2/network@0 network-pciex8086,105e
IOBD/NET3 PCIE IOBD /pci@7c0/pci@0/pci@2/network@0,1 network-pciex8086,105e
IOBD/PCIE1 PCIE 1 /pci@7c0/pci@0/pci@8/emlx@0 emlx-pci10df,fe00LPe11000-E
IOBD/PCIE2 PCIE 2 /pci@7c0/pci@0/pci@9/emlx@0 emlx-pci10df,fe00LPe11000-E
IOBD/PCIX PCIX IOBD /pci@7c0/pci@0/pci@1/pci@0/isa@2 isa
IOBD/PCIX PCIX IOBD /pci@7c0/pci@0/pci@1/pci@0/usb@5 usb-pciclass,0c0310
IOBD/PCIX PCIX IOBD /pci@7c0/pci@0/pci@1/pci@0/usb@6 usb-pciclass,0c0310
IOBD/PCIX PCIX IOBD /pci@7c0/pci@0/pci@1/pci@0/ide@8 ide-pci10b9,5229
# prtconf -vp
e.g.
firmware-version: '1.09.00.00'
mpt-version: '1.05'
assigned-addresses: 81070010.00000000.00001000.00000000.00000100.82070014.00000000.00300000.00000000.00010000.8207001c.00000000.00310000.00000000.00010000.82070030.00000000.00400000.00000000.00400000
compatible: 'pciex1000,56.1000.1000.2' + 'pciex1000,56.1000.1000' + 'pciex1000,56.2' + 'pciex1000,56' + 'pciexclass,010000' + 'pciexclass,0100' + 'pci1000,56'
model: 'LSI,1064E'
# prtpicl -v
e.g.
:model LSI,1064E
:compatible (1b6000003e6TBL)
| pciex1000,56.1000.1000.2 |
| pciex1000,56.1000.1000 |
| pciex1000,56.2 |
| pciex1000,56 |
| pciexclass,010000 |
| pciexclass,0100 |
| pci1000,56 |
:assigned-addresses
- FW or patch for LSI1064E CHIP:
sas1064 on T2000 with PCI-X Plugin cards FCode_1_00_40.ROM
122165-02
Hardware/1064_FCode Sun Fire T2000 1064 FCode update
- The latest patch release notes:
Keywords: sun fire t2000 1064_fcode 370-7696-04 lsi3041xl
Synopsis: Hardware/1064_FCode Sun Fire T2000 1064 FCode update
Date: Mar/15/2006
Install Requirements: Additional instructions may be listed below
Solaris Release: 10
SunOS Release: 5.10
Unbundled Product: Hardware/Fcode LSI1064 PCI-X
Unbundled Release: N/A
Xref:
Topic: 1064_FCode 370-7696-04 LSI3041XL update for the Sun Fire T2000
Relevant Architectures: sparc
BugId's fixed with this patch: 6378605 6394464
Changes incorporated in this version: 6394464
Patches accumulated and obsoleted by this patch:
Patches which conflict with this patch:
Patches required with this patch:
Obsoleted by:
Files included with this patch:
370-7696-04_FCode_1_00_40.ROM
Problem Description:
6394464 Volume type is
being referenced instead of volume id in 1.00.39 1064 fcode on PCI-X 1064
plugin card
(from )
6378605 activate_volume FCode command is needed for PCI-X 1064 plugin card
Patch Installation Instructions:
===============================
ABOUT THIS PATCH
-------------------------
This patch updates the PCI-X 1064 FCode:
To version 1.00.40
Before using the RAID feature on your boot drive, ensure that the PCI-X
1064 plug-in card's Fcode has been upgraded to version 1.00.39 or later.
TO CHECK THE VERSION OF THE FCODE
--------------------------------------------------
There are two ways to check the PCI-X FCODE version:
o From the Solaris Operating System (OS)
o From the OpenBoot PROM (OBP)
From the Solaris OS:
1. As root, run the prtconf command.
Example:
# prtconf -vp | egrep '(model|version)' | \
egrep -vi '(obp|firmwareversion| mpt-version)'
model: 'LSI,1064'
version: '1.00.39'
#
In this example, the PCI-X FCODE is version 1.00.39, and needs
to be updated.
From the OBP:
1. Identify the node for the PCI-X 1064 card:
Example:
{0} ok show-disks
a) /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2/disk
b) /pci@7c0/pci@0/pci@1/pci@0/ide@8/cdrom
c) /pci@7c0/pci@0/pci@1/pci@0/ide@8/disk
q) NO SELECTION
Enter Selection, q to quit: q
{0} ok
In this example on the Sun-Fire-T2000, "a" represents
the plugin card. The plugin card has LSILogic in the name
because it is treated as a third party add in card.
Once you identify the node in this list, type q to return
to the ok prompt.
2. Select the node.
Example:
{0} ok select /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2
{0} ok
In this example, only the full path to the node is specified
(the path does not include the "disk").
3. Display the node properties.
Example:
{0} ok .properties
firmware-version 1.06.00.00
mpt-version 1.05
assigned-addresses 81061010 00000000 00001000 00000000 00000100
82061014 00000000 03000000 00000000 00010000
8206101c 00000000 03010000 00000000 00010000
82061030 00000000 03200000 00000000 00200000
compatible pci13e9,50
pci1000,50
AAPL,MPT
model LSI,1064
reg 00061000 00000000 00000000 00000000 00000000
01061010 00000000 00000000 00000000 00000100
03061014 00000000 00000000 00000000 00010000
0306101c 00000000 00000000 00000000 00010000
02061030 00000000 00000000 00000000 00200000
---> version 1.00.39
wide 00000010
device_type scsi-2
name LSILogic,sas
fcode-rom-offset 00009a00
66mhz-capable
devsel-speed 00000001
latency-timer 00000040
cache-line-size 00000010
max-latency 0000000a
min-grant 00000040
interrupts 00000001
class-code 00010000
subsystem-id 000030f0
subsystem-vendor-id 00001000
revision-id 00000002
device-id 00000050
vendor-id 00001000
In this example, the PCI-X FCODE is version 1.00.39, and needs
to be updated.
4. UN-select the node before booting the server.
Example:
{0}ok unselect-dev
{0}ok
TO UPDATE THE FCODE
------------------------------
1. Boot the Solaris OS.
2. Login as root.
3. Copy the new FCODE image onto the system.
The new FCode image is included in this patch. Copy the image
into a dir on the system which you are going to update or to
a network accessible dir. Use this full path along with the
image name in step 5) below.
4. Identify the controller.
In the following example, the controller is identified with
the "LSI" string:
# ls -al /dev/cfg
total 22
drwxr-xr-x 4 root root 512 Jan 19 17:28 .
drwxr-xr-x 16 root sys 4608 Jan 20 10:47 ..
lrwxrwxrwx 1 root root 50 Jan 19 17:28 c0 ->
../../ devices/pci@7c0/pci@0/pci@1/pci@0/ide@8:scsi
---> lrwxrwxrwx 1 root root 61 Jan 19 17:28 c1 ->
../../ devices/pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2:scsi
lrwxrwxrwx 1 root root 45 Jan 19 17:28 c2 ->
../../ devices/pci@780/pci@0/pci@9/scsi@0:scsi
5. Update the FCODE using the raidctl command.
Syntax:
# raidctl -F [full_path_of_image] [controller]
Example:
# raidctl -F 370-7696-04_FCode_1_00_40.ROM 1
Image file contains:
1.00.40
Update flash image on controller 1 (yes/no)? y
Flash updated successfully.
#
6. Reboot the server to load the new FCode.
- The administration guides for LSI1064E:
Brief summary of the RAID operations and command usage:
Create a mirror:
# raidctl -c [-r 1] primary secondary
Create a stripe:
# raidctl -c -r 0 disk1 disk2 [disk3] . . .
Delete a RAID volume:
# raidctl -d volume
Display RAID volumes:
# raidctl
Display the disk controller firmware/fcode version:
# prtconf -vp | egrep '(model|version)' | \
egrep -vi '(obp|firmwareversion| mpt-version)'
Update the disk controller firmware:
# raidctl -F [path_of_image_file] [controller#]
5. The first generation hard raid Chip LSI1068E
Applies to: Sun Fire T5120/T5220/T5140/T5240/T5440/NetraT5220/NetraT5440
Chip: LSI1068E SCSI Host Adapter
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd0 Soft Errors: 2 Hard Errors: 0 Transport Errors: 0
Vendor: LSILOGIC Product: Logical Volume Revision: 3000 Serial No:
Size: 146.56GB <146561286144 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
# cfgadm -alv
e.g.
c1 connected configured unknown
unavailable scsi-bus n /devices/pci@400/pci@0/pci@8/scsi@0:scsi
c1::dsk/c1t0d0 connected configured unknown LSILOGIC Logical Volume
unavailable disk n /devices/pci@400/pci@0/pci@8/scsi@0:scsi::dsk/c1t0d0
# prtdiag -v
e.g.
================================ IO Devices ================================
Slot + Bus Name + Model
Status Type Path
----------------------------------------------------------------------------
MB/SASHBA PCIE scsi-pciex1000,58 LSI,1068E
/pci@400/pci@0/pci@8/scsi@0
# prtconf -vp
e.g.
firmware-version: '1.27.00.00'
mpt-version: '1.05'
assigned-addresses: 81060010.00000000.00000000.00000000.00000100.82060014.00000000.00300000.00000000.00010000.8206001c.00000000.00310000.00000000.00010000.82060030.00000000.00400000.00000000.00400000
compatible: 'pciex1000,58.1000.1000.4' + 'pciex1000,58.1000.1000' + 'pciex1000,58.4' + 'pciex1000,58' + 'pciexclass,010000' + 'pciexclass,0100' + 'pci1000,56'
model: 'LSI,1068E'
reg: 00060000.00000000.00000000.00000000.00000000.01060010.00000000.00000000.00000000.00000100.03060014.00000000.00000000.00000000.00010000.0306001c.00000000.00000000.00000000.00010000.02060030.00000000.00000000.00000000.00400000
version: '1.00.41'
# prtpicl -v
e.g.
:model LSI,1068E
:compatible (b800000b18TBL)
| pciex1000,58.1000.1000.4 |
| pciex1000,58.1000.1000 |
| pciex1000,58.4 |
| pciex1000,58 |
| pciexclass,010000 |
| pciexclass,0100 |
| pci1000,56 |
:assigned-addresses
- FW or patch for LSI1068E CHIP:
Patches affecting all T1000 T2000 T5120/T5220 T5140/T5240 V445 V215 V245
Patch ID |
Description |
Min level |
Notes & Dependencies |
|
Sun OS 5.10 raidctl |
-05 |
This resolves a number of the issues affecting behaviour changed in s10 8/07 (update 4)from prior S10 |
T5120/T5220 Patches:
Patch ID |
Description |
Min level |
Notes & Dependencies |
|
Sun OS 5.10 librcm.so.1 patch |
-01 |
For Solaris 10 1/06 OS only,not required for the Solaris 10 6/06 or later OS. |
|
Sun OS 5.10: mpt radioictl libpiclsbl libpiclfrudr patch |
-14 |
118822-20 |
|
Hardware/PROM Sun Fire T1000 1064 Firmware update V1.10.00.00IR |
-01 |
119850-14 |
Required Solaris patches:
Solaris 8:
SunOS 5.8: mpt driver patch
Solaris 9:
SunOS 5.9: mpt Patch
Solaris 10:
(or later) SunOS 5.10: MPT driver patch
Please note the Required Patches for 143128-xx listed in the Patch Description
In general mpt driver is part of kernel patches of Sol9 and Sol10; kernel patching is required.
Hardware Raid HBA firmware:
SPARC Enterprise T5120 / T5220
Huron System Board 501-7781 / 501-7812 / 501-7813 / 501-7814
Netra T5220
Turgo System Board
540-7755 / 501-7813 + 541-2814 / 501-7781 + 541-2867 / 501-7814
SAS1068E-B1 1.27.00.00 (not ECO released) |
Huron+ System Board 511-1087 / 511-1198 / 511-1200
SAS1068E-B2 1.27.00.00 (not ECO released) 2 |
SPARC Enterprise T5140 / T5240
Maramba System Board 501-7846 / 501-7847 / 501-7848 / 511-1035
SAS1068E-B2 Version 1.27.00.00 2 |
SPARC Enterprise T5440
Batoka System Board 541-2749 / 501-7783
SAS1068E-B2 1.27.00.00 (not ECO released) 1 |
- The latest patch release notes:
Refer the patch readme
- The administration guides for LSI1068E
There is a new version of the "raidctl" utility introduced with Solaris 10 8/07.The usage of the utility has changed dramatically.Two potential areas of concern are:
1.) Previous version of "raidctl" required only entering "raidctl" to get volume and disk status.It now requires "raidctl -l volumename".Concern here that a disk or volume in a non optimal or failed state may go unnoticed.
2.) There are option and features listed that are not supported in many platforms(example:hot spares) that may cause customers to think they are available.
The new version of the 'raidctl' included original function
How to obtain RAID status:
raidctl -l 1 -- Controller 1
raidctl -l c1t2d0 -- Volume c1t2d0
raidctl -l c1t0d0
raidctl -l -g 0.0.0 1 -- Disk 0.0.0
raidctl -l -g 0.1.0 1
raidctl -l -g 0.2.0 1
raidctl -l -g 0.3.0 1
Facts:
- OBP level RAID commands are not supported by Engineering and are not intended for customer use.
- Migrating of disk volumes from one system to another is not supported.
- Creation of any RAID volume will result in the data on ALL disks in the volume being lost.
How to activate a RAID volume on T1000/T2000, V215/V245/V445, T5120/T5220, T5140/T5240/T5440 Servers
This document provides a reference to the procedure for re-activating an existing RAID volume, on Sun Fire T1000/T2000, V215/V245/V445, SRARC Enterprise T5120/T5220, T5140/T5240 and T5440 Servers. If you have an existing RAID volume and have to replace the raid controller on one of these servers, you'll need to run some OBP commands in order to see the volume with the new board. The outlined procedure includes generic instructions, as well as platform specific instructions.
When the RAID controller for internal disks has been replaced, customers may need to re-activate the RAID volume by using OBP commands.
For T2000 systems, if using CPU/mem board part number 501-6843, the controller is a plug-in PCI-X card, part number 370-7696. For all other instances on the T2000, T1000, V215/V245/V445, T5120/T5220, T5140/T5240 and T5440, the controller is on the system board (motherboard).
Generic instructions for T1000, T2000, V215/V245/V445, T5x20 and T5x40:
{0} ok setenv fcode-debug? true
{0} ok setenv auto-boot? false
{0} ok reset-all
Path select instructions for T1000/Erie (dual-disk configuration is required for RAID operations):
{0} ok select /pci@7c0/pci@0/pci@8/scsi@2
Path select instructions for T2000/Ontario (pn 501-6843):
{0} ok select /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2
Path select instructions for T2000+/Ontario+ (pn 501-7501):
{0} ok select /pci@780/pci@0/pci@9/scsi@0
Path select instructions for V215/V245/Seattle:
{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1
Path select instructions for V445/Boston:
{0} ok select /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1
Path select instructions for V445 with 501-7244 SAS controller
{0} ok select /pci@1e,600000/pci@0/pci@2/scsi@0
Path select instructions for T5120/T5220/Huron:
{0} ok select /pci@0/pci@0/pci@2/scsi@0
Path select instructions for T5140/T5240/Maramba:
{0} ok select /pci@400/pci@0/pci@8/scsi@0/
Path select instructions for T5440/Batoka:
{0} ok select /pci@400/pci@0/pci@1/scsi@0/
Generic instructions for T1000, T2000, V215, V245, V445, T5x20, T5x40:
{0} ok show-volumes ( Look for the inactive volume)
{0} ok X activate-volume (X is the volume number)
{0} ok unselect-dev
{0} ok probe-scsi-all (to check for disks)
When dealing with multiple volumes make sure they are activated in descending order:
Generic instructions for T1000, T2000, V215, V245, V445, T5x20, T5x40:
{0} ok show-volumes
{0} ok 1 activate-volume
{0} ok 0 activate-volume
{0} ok unselect-dev
{0} ok probe-scsi-all (to check for disks)
Failure to do so will result in the following error and will require a reset-all prior to
activating the second volume:
{0} ok 0 activate-volume
Volume 0 is now activated
{0} ok 1 activate-volume
RAID volume request failed, ioc-status = 0007
Volume 1 could not be activated
Following activation the first volume will remain 'Degraded' until both volumes have completed syncing.
6. The second generation hard raid Chip LSI2008
Applies to: SPARC T3-1/T3-2/T3-4; T4-1/T4-2/T4-4
Chip: LSI 2008 mpt sas HBA
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd4 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
Vendor: LSI Product: Logical Volume Revision: 3000 Device Id: id1,sd@n600
508e0000000004e09f9a472cafd04
Size: 299.00GB <298999349248 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
sd7 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
Vendor: LSI Product: Logical Volume Revision: 3000 Device Id: id1,sd@n600508e000000000ae093bd052e7bc08
Size: 599.00GB <598999040000 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 2 Predictive Failure Analysis: 0
sd8 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
# cfgadm -alv
e.g.
c7 connected configured unknown
unavailable scsi-sas n /devices/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/
iport@v0:scsi
c7::dsk/c7t34FDCA72A4F9094Ed0 connected configured unknown LSI Logical
Volume
unavailable disk n /devices/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/
iport@v0:scsi::dsk/c7t34FDCA72A4F9094Ed0
# prtdiag -v
e.g.
================================ IO Devices ================================
Slot + Bus Name + Model
Status Type Path
----------------------------------------------------------------------------
/SYS/MB/REM0/SASHBA0PCIE LSI,sas-pciex1000,72 LSI,2008
/pci@400/pci@1/pci@0/pci@0/LSI,sas@0
# prtconf -vp
e.g.
firmware-version: '11.05.02.00'
mpt-version: '2.00'
local-wwid: 500605b0.02925120
vf-assigned-addresses: 83030000.00000041.00080000.00000
000.00004000.83030002.00000041.000c0000.00000000.00040000
assigned-addresses: 81030010.00000000.00000000.00000000
.00000100.83030014.00000041.00000000.00000000.00004000.8303001c.00000041.0004000
0.00000000.00040000.82030030.00000000.00100000.00000000.00080000
vf-reg: 03030000.00000000.00000000.00000000.00004000.03
030002.00000000.00000000.00000000.00040000
compatible: 'pciex1000,72' + 'pci1000,72'
model: 'LSI,2008'
# prtpicl -v
e.g.
:device_type scsi-sas
:wide 0x10
:version 1.00.60
:reg
00 03 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00 00
01 03 00 10 00 00 00 00 00 00 00 00 00 00 00 00 00 00 01 00
03 03 00 14 00 00 00 00 00 00 00 00 00 00 00 00 00 00 40 00
03 03 00 1c 00 00 00 00 00 00 00 00 00 00 00 00 00 04 00 00
02 03 00 30 00 00 00 00 00 00 00 00 00 00 00 00 00 08 00 00
:model LSI,2008
- FW or patch for LSI 2008 CHIP:
SPARC T3-1 LSI 2008 Firmware to support TSST DVD drive |
This version is for SPARC T3-1 LSI 2008 Firmware to support TSST DVD drive. Refer to the README file for instructions on updating the firmware using the files included in this patch. |
|
SPARC T3-1 LSI SAS2 Expander firmware to force zoning on the 16 disk backplane |
This version is for SPARC T3-1 LSI SAS2 Expander firmware to force zoning on the 16 disk backplane. Refer to the README file for instructions on updating the firmware using the files included in this patch. |
- The administration guides for LSI 2008
RAID technology allows for the construction of a logical volume, made up of several physical disks, to provide data redundancy, increased performance, or both. The system's onboard disk controller supports the following RAID configurations:
- Integrated Stripe (IS) volumes - RAID 0
- Integrated Mirror (IM) volumes - RAID 1
- Enhanced mirroring - RAID 1e (not supported in T3-4 & T4-4)
The onboard disk controller(s) of the servers provide for the creation of two hardware RAID volumes per controller. The T3-2 and T4-2 have one onboard controller so there is a max of two raid volumes. The other systems which have two onboard controllers or REM's would support four.
NOTE: It is extremely important to be aware with all of these platforms that when a system board, or REM is replaced,certain procedures have to be followed to have the existing RAID volumes be seen after replacement. Please reference document: 1387771.1.
=============================
Similarities from T5xx0 systems:
- Creating and deleting volumes will both destroy the data on all member disks
- Once a raid volume is created all members disks will no longer be seen by Solaris or OBP, only the raid volume will be visible
- Procedures must be followed to see any configured volumes after a system board or REM is replaced
Differences from T5xx0 systems:
- Unlike T5xxx systems, OBP raid commands are supported for customer use
- The solaris utility raidctl is not supported, in it's place is sas2ircu but it must be downloaded from an LSI website, another is raidconfig which is part of HMP
- Raid level 1E is supported (not T3-4 and T4-4)
Beware of disk Labels:
The volume must be labeled with one of type SMI for boot disks which are limited to 1TB, but can be extended to 2TB. An EFI label is required for all volumes larger than 2TB. More information can be found . ALWAYS ensure that the newly created volume has been relabeled!!!
Three environments for creating and managing Hardware RAID:
- Fcode utility -- This utility consists of a set of special commands to display targets and manage logical volumes on your server. You access these commands through the OBP environment.
- LSI SAS2 2008 RAID Management Utility for SPARC T4 servers -- You can use the sas2ircu commands contained in the LSI SAS2 Integrated Configuration Utility to configure and manage RAID volumes on your server. To use the sas2ircu command set, download and install the SAS2IRCU software from the following location: for T3-1, T3-2, T4-1, & T4-2; for T3-4, T4-4, T3-1B, & T4-1B.
- Oracle Hardware Management Pack 2.2 -- You can use the raidconfig commands contained in this software's Oracle Server CLI Tools component to create and manage RAID volumes on your server. To use these commands, download and install the latest version of the Hardware Management Pack from this location:
- Fcode utility:
The following OBP level commands are supported:
show-children Lists all connected physical drives and logical volumes.
show-volumes Lists all connected logical volumes in detail.
create-raid0-volume Creates a RAID 0 volume (minimum two targets).
create-raid1-volume Creates a RAID 1 volume (exactly two targets).
create-raid1e-volume Creates a RAID 1e volume (minimum three targets).
Delete-volume Deletes a RAID volume.
activate-volume Re-activate a RAID volume after controller has been replaced.
You can configure two global hot spare drives to protect data on mirrored RAID volumes. If one of the drives in a RAID 1 or RAID 1E mirrored volume fails, the onboard RAID controller will replace the failed drive automatically with a hot spare drive and then resynchronize the mirrored data. You must use the the sas2ircu LSI utility or raidconfig to add global hot spare drives.
Example of Console Output
SPARC T3-1, No Keyboard
Copyright (c) 1998, 2011, Oracle and/or its affiliates. All rights reserved.
OpenBoot 4.32.4, 2048 MB memory available, Serial #94789078.
Ethernet address 0:21:28:a6:5d:d6, Host ID: 85a65dd6.
{0} ok
{0} ok select /pci@400/pci@2/pci@0/pci@4/scsi@0
{0} ok show-volumes
No volumes to show
{0} ok show-children
FCode Version 1.00.54, MPT Version 2.00, Firmware Version 5.00.17.00
Target a
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abb2358 SASAddress 5000cca00abb2359 PhyNum 0
Target b
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abba630 SASAddress 5000cca00abba631 PhyNum 1
Target c
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abc5270 SASAddress 5000cca00abc5271 PhyNum 2
Target d
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00aa80228 SASAddress 5000cca00aa80229 PhyNum 3
Target e
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00a82fbfc SASAddress 5000cca00a82fbfd PhyNum 4
Target f
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abcd2b4 SASAddress 5000cca00abcd2b5 PhyNum 5
Target 10
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00aac16e4 SASAddress 5000cca00aac16e5 PhyNum 6
Target 12
Unit 0 Encl Serv device SUN SAS2 X16DBP 0305
SASAddress 50800200000272bd PhyNum 18
{0} ok a b create-raid0-volume
Target a size is 583983104 Blocks, 298 GB
Target b size is 583983104 Blocks, 298 GB
The volume can be any size from 1 MB to 570296 MB
What size do you want? [570296]
Volume size will be 1167966208 Blocks, 597 GB
Enter a volume name: [0 to 15 characters] RAID0
Volume has been created
{0} ok c d e create-raid1e-volume
Target c size is 583983104 Blocks, 298 GB
Target d size is 583983104 Blocks, 298 GB
Target e size is 583983104 Blocks, 298 GB
The volume can be any size from 1 MB to 427722 MB
What size do you want? [427722]
Volume size will be 875974656 Blocks, 448 GB
Enter a volume name: [0 to 15 characters] RAID1E
Volume has been created
{0} ok f 10 create-raid1-volume
Maximum number of volumes already exist
{0} ok show-volumes
Volume 0 Target 388 Type RAID1E (Mirroring Extended)
Name RAID1E WWID 0673b84764732723
Optimal Enabled Consistent Background Init In Progress
3 Members 875974656 Blocks, 448 GB
Disk 3
Member 0 Optimal
Target c HITACHI H103030SCSUN300G A2A8
Disk 4
Member 1 Optimal
Target d HITACHI H103030SCSUN300G A2A8
Disk 5
Member 2 Optimal
Target e HITACHI H103030SCSUN300G A2A8
Volume 1 Target 389 Type RAID0 (Striping)
Name RAID0 WWID 00c3b81bae2324f7
Optimal Enabled
2 Members 1167966208 Blocks, 597 GB
Disk 1
Member 0 Optimal
Target a HITACHI H103030SCSUN300G A2A8
Disk 2
Member 1 Optimal
Target b HITACHI H103030SCSUN300G A2A8
{0} ok reset-all
CONSERVER: XOFF timed out on t3-1-sin06-a.sg (/dev/term/e025)
SPARC T3-1, No Keyboard
Copyright (c) 1998, 2011, Oracle and/or its affiliates. All rights reserved.
OpenBoot 4.32.4, 2048 MB memory available, Serial #94789078.
Ethernet address 0:21:28:a6:5d:d6, Host ID: 85a65dd6.
{0} ok ok select scsi1
{0} ok show-children
FCode Version 1.00.54, MPT Version 2.00, Firmware Version 5.00.17.00
Target a
Unit 0 Removable Read Only device TEAC DV-W28SS-R 1.0C
SATA device PhyNum 6
Target b
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abce7f8 SASAddress 5000cca00abce7f9 PhyNum 8
Target c
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00aaec29c SASAddress 5000cca00aaec29d PhyNum 9
Target d
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00abce514 SASAddress 5000cca00abce515 PhyNum a
Target e
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00a8065f0 SASAddress 5000cca00a8065f1 PhyNum b
Target f
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00a7f8ea8 SASAddress 5000cca00a7f8ea9 PhyNum c
Target 10
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00aaec608 SASAddress 5000cca00aaec609 PhyNum d
Target 11
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00aaf4ab0 SASAddress 5000cca00aaf4ab1 PhyNum e
Target 12
Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB
SASDeviceName 5000cca00a81f620 SASAddress 5000cca00a81f621 PhyNum f
Target 13
Unit 0 Encl Serv device SUN SAS2 X16DBP 0305
SASAddress 50800200000272bd PhyNum 18
{0} ok show-volumes
No volumes to show
{0} ok a b create-raid1-volume
Couldn't get RAID PhysDisk information for target .a
{0} ok b c create-raid1-volume
Target b size is 583983104 Blocks, 298 GB
Target c size is 583983104 Blocks, 298 GB
The volume can be any size from 1 MB to 285148 MB
What size do you want? [285148]
Volume size will be 583983104 Blocks, 298 GB
Enter a volume name: [0 to 15 characters] RAID1
Volume has been created
{0} ok show-volumes
Volume 0 Target 389 Type RAID1 (Mirroring)
Name RAID1 WWID 0d8bbe28ad6b3ae2
Optimal Enabled Background Init In Progress
2 Members 583983104 Blocks, 298 GB
Disk 0
Primary Optimal
Target b HITACHI H103030SCSUN300G A2A8
Disk 1
Secondary Optimal
Target c HITACHI H103030SCSUN300G A2A8
{0} ok
- sas2ircu utility:
The utility is sas2ircu and does not come as part of Solaris. It must be downloaded from the LSI website The raidctl utility is not supported for the T3/T4 platforms. It is recommended to rename this utility to sas2ircu(SAS2 Integrated RAID Configuration Utility)
(removing the version indication) & place this utility in /usr/sbin, as well as giving it root execute permissions. Command syntax & sample commands are:
Note:
SAS2IRCU is a command-line utility that you use to configure the Integrated RAID
functions of LSI SAS2 controllers. SAS2IRCU runs in the DOS, Linux, EFI, and Windows
Pre-Installation (WinPE) environments. It is a minimally interactive program that you
can execute from a command-line prompt or a shell script. When you use SAS2IRCU
commands, the program sends results to the environment through the program status
value that it returns when it exits.
OEMs can use SAS2IRCU to configure Integrated RAID devices on the LSI family of
MPT2.0-based SAS2 controllers. Typical uses for this utility include, but are not limited
to, system configuration on the manufacturing floor, post-delivery configuration by
end users, and ongoing configuration and monitoring of Integrated RAID devices.
sas2ircu
where
where
DISPLAY - display controller, volume and physical device info
LIST - Lists all available LSI adapters (does not need ctlr #>
CREATE - create an IR volume
DELETE - Delete entire RAID configuration on the selected controller
DELETEVOLUME - Delete a specific RAID Volume on the selected controller
HOTSPARE - make drive a hot spare
STATUS - display current volume status info
CONSTCHK - Start Consistency Check operation on the specified IR Volume
ACTIVATE - Activate an Inactive IR volume
LOCATE - Locate a disk drive on an enclosure
LOGIR - Upload or Clear IR Log data
BOOTIR - Select an IR Boot Volume as primary boot device
BOOTENCL - Select an Enclosure/Bay as primary boot device
where
Please see the TOI on the Troubleshooting page for sample output. Obtain the following command output from a system for diagnosis:
# sas2ircu 0 display all
# sas2ircu 0 status
# sas2ircu 1 display all
# sas2ircu 1 status
A sample mirror creation is as follows:
# sas2ircu 1 create raid1 MAX 2:9 2:15
Example of sas2ircu output:
# ./sas2ircu-6.250.02.00
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
SAS2IRCU: No command specified.
sas2ircu
where
Number between 0 and 255
where
DISPLAY - display controller, volume and physical device info
LIST - Lists all available LSI adapters (does not need ctlr #>
CREATE - create an IR volume
DELETE - Delete entire RAID configuration on the selected controller
DELETEVOLUME - Delete a specific RAID Volume on the selected controller
HOTSPARE - make drive a hot spare
STATUS - display current volume status info
CONSTCHK - Start Consistency Check operation on the specified IR Volume
ACTIVATE - Activate an Inactive IR volume
LOCATE - Locate a disk drive on an enclosure
LOGIR - Upload or Clear IR Log data
BOOTIR - Select an IR Boot Volume as primary boot device
BOOTENCL - Select an Enclosure/Bay as primary boot device
where
Command specific
values; enter "sas2ircu
to get command specific help
# ./sas2ircu-6.250.02.00 LIST
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Adapter Vendor Device SubSys SubSys
Index Type ID ID Pci Address Ven ID Dev ID
----- ------------ ------ ------ ----------------- ------ ------
0 SAS2008 1000h 72h 00h:400h:00h:00h 1000h 0072h
Adapter Vendor Device SubSys SubSys
Index Type ID ID Pci Address Ven ID Dev ID
----- ------------ ------ ------ ----------------- ------ ------
1 SAS2008 1000h 72h 00h:b00h:00h:00h 1000h 0072h
SAS2IRCU: Utility Completed Successfully.
# ./sas2ircu-6.250.02.00 0 DISPLAY
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Read configuration has been initiated for controller 0
------------------------------------------------------------------------
Controller information
------------------------------------------------------------------------
Controller type : SAS2008
BIOS version : 0.00.00.00
Firmware version : 5.00.17.00
Channel description : 1 Serial Attached SCSI
Initiator ID : 0
Maximum physical devices : 831
Concurrent commands supported : 1871
Slot : Unknown
Segment : 0
Bus : 1024
Device : 0
Function : 0
RAID Support : Yes
------------------------------------------------------------------------
IR Volume information
------------------------------------------------------------------------
Physical device information
------------------------------------------------------------------------
Initiator at ID #0
Device is a Hard disk
Enclosure # : 2
Slot # : 0
SAS Address : 5000cca-0-0abb-2359
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1031G9XHHE
Protocol : SAS
Drive Type : SAS_HDD
Device is a Sequential access device
Enclosure # : 2
Slot # : 0
SAS Address : 5080020-0-0002-72bd
State : Standby (SBY)
Manufacturer : SUN
Model Number : SAS2 X16DBP
Firmware Revision : 0305
Serial No : x365351
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 1
SAS Address : 5000cca-0-0abb-a631
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1031GA66ME
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 2
SAS Address : 5000cca-0-0abc-5271
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1031GAKP7E
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 3
SAS Address : 5000cca-0-0aa8-0229
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1027GZDB5E
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 4
SAS Address : 5000cca-0-0a82-fbfd
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1017GA14AE
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 5
SAS Address : 5000cca-0-0abc-d2b5
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1031GAV71E
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 6
SAS Address : 5000cca-0-0aac-16e5
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1027G1MY1E
Protocol : SAS
Drive Type : SAS_HDD
Device is a Hard disk
Enclosure # : 2
Slot # : 7
SAS Address : 5000cca-0-0aab-331d
State : Ready (RDY)
Size (in MB)/(in sectors) : 286102/585937499
Manufacturer : HITACHI
Model Number : H103030SCSUN300G
Firmware Revision : A2A8
Serial No : 1027G14STE
Protocol : SAS
Drive Type : SAS_HDD
------------------------------------------------------------------------
Enclosure information
------------------------------------------------------------------------
Enclosure# : 1
Logical ID : 50800200:00d69868
Numslots : 8
StartSlot : 0
Enclosure# : 2
Logical ID : 50800200:000272bf
Numslots : 17
StartSlot : 0
------------------------------------------------------------------------
SAS2IRCU: Command DISPLAY Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
#
# ./sas2ircu-6.250.02.00 1 create
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
SAS2IRCU: Incorrect number of command line parameters for 'CREATE' command!
The format of the CREATE command is
sas2ircu
where
A controller number between 0 and 255.
where
The type of the volume to create and is either RAID1 (or)
RAID1E (or) RAID0 (or) RAID10.
where
The size of the volume to create. It should be given in Mbytes
e.g. 2048 or 'MAX' to use the maximum size possible.
where
A list of Encl:Bay pairs identifying the disk drives you
wish to include in the volume being created. If the volume type is
'RAID1', the first drive will be selected as the primary and the
second as the secondary drive.
For a type 'RAID1' volume exactly 2 disks must be specified.
For a type 'RAID1E' volume min of 3 disks must be specified.
For a type 'RAID0' volume min of 2 disks must be specified.
For a type 'RAID10' volume min of 4 disks must be specified.
where [Volume Name] is an optional argument that can be used
to identify a Volume with a user specified Alpha-numeric string
where noprompt is an optional argument that eliminates
warnings and prompts
# ./sas2ircu-6.250.02.00 1 create RAID1 MAX 2:10 2:11
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
You are about to create an IR volume.
WARNING: Proceeding with this operation may cause data loss or data
corruption. Are you sure you want to proceed (YES/NO)? yes
WARNING: This is your last chance to abort this operation. Do you wish
to abort (YES/NO)? no
Please wait, may take up to a minute...
Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, inactive
Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active
Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active
Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active, background initialization in progress
SAS2IRCU: Volume created successfully.
SAS2IRCU: Command CREATE Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
# Jul 21 13:00:03 t3-1-sin06-a scsi: WARNING:
/pci@400/pci@2/pci@0/pci@4/scsi@0/iport@v0/disk@w3d5c9144a29779de,0 (sd14):
Jul 21 13:00:03 t3-1-sin06-a Corrupt label; wrong magic number
Jul 21 13:00:03 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:00:03 t3-1-sin06-a send_sep act 0: ioc status:8001
Jul 21 13:00:03 t3-1-sin06-a scsi: WARNING:
/pci@400/pci@2/pci@0/pci@4/scsi@0/iport@v0/disk@w3d5c9144a29779de,0 (sd14):
Jul 21 13:00:03 t3-1-sin06-a Corrupt label; wrong magic number
#
# ./sas2ircu-6.250.02.00 0 status
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Background command progress status for controller 0...
IR Volume 1
Volume ID : 904
Current operation : None
Volume status : Enabled
Volume state : Optimal
Physical disk I/Os : Not quiesced
IR Volume 2
Volume ID : 905
Current operation : Background Init
Volume status : Enabled
Volume state : Optimal
Physical disk I/Os
: Not quiesced
Volume size (in sectors) : 583983104
Number of remaining sectors : 563631328
Percentage complete : 3.48%
SAS2IRCU: Command STATUS Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
#
19.) How to delete a RAID volume with sas2ircu
# ./sas2ircu-6.250.02.00 1 status
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Background command progress status for controller 1...
IR Volume 1
Volume ID : 904
Current operation : Background Init
Volume status : Enabled
Volume state : Optimal
Physical disk I/Os : Not quiesced
Volume size (in sectors) : 583983104
Number of remaining sectors : 436668576
Percentage complete : 25.23%
IR Volume 2
Volume ID : 905
Current operation : None
Volume status : Enabled
Volume state : Optimal
Physical disk I/Os : Not quiesced
SAS2IRCU: Command STATUS Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
# ./sas2ircu-6.250.02.00 1 deletevolume 904
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
You are about to delete an existing RAID Volume on a controller. This command will
delete the specified RAID volume and associated HotSpare drive(s).
WARNING: Proceeding with this operation may cause data loss or data
corruption. Are you sure you want to proceed (YES/NO)? yes
WARNING: This is you last chance to abort this operation. Do you wish
to abort (YES/NO)? no
Please wait, may take up to a minute...
Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):
Jul 21 13:52:31 t3-1-sin06-a raid doesn't exist at specified target.
SAS2IRCU: Volume deleted successfully.
SAS2IRCU: Command DELETEVOLUME Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
# Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00abce514 (sd3):
Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number
Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00a8065f0 (sd2):
Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number
Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00abce514 (sd3):
Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number
Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00a8065f0 (sd2):
Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number
# ./sas2ircu-6.250.02.00 1 status
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Background command progress status for controller 1...
IR Volume 1
Volume ID : 905
Current operation : None
Volume status : Enabled
Volume state : Optimal
Physical disk I/Os : Not quiesced
SAS2IRCU: Command STATUS Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
#
- How to replace a failed disk on a degraded volume
If the system has RAID1 or RAID1E disk volumes, the system can afford to have a single disk failure
without data loss. To determine the status of the RAID volume and location of the of the faulty drive
in the server, you can use the following command.
sas2ircu 0 display
sas2ircu 0 display
Once the faulty disk slot is identified, just pullout the failed disk and insert the replacement disk into
the chassis. Depending on the size and activity on the server , a 300GB disk on an idle system will take
approximately 1 hour to rebuild.
The following command will give you the percentage of the rebuilding process
sas2ircu 0 status
sas2ircu 1 status
# ./sas2ircu-6.250.02.00 1 status
LSI Corporation SAS2 IR Configuration Utility.
Version 6.250.02.00 (2010.06.30)
Copyright (c) 2009-2010 LSI Corporation. All rights reserved.
Background command progress status for controller 1...
IR Volume 1
Volume ID : 905
Current operation : Synchronize
Volume status : Enabled
Volume state : Degraded
Physical disk I/Os : Not quiesced
Volume size (in sectors) : 583983104
Number of remaining sectors : 520166912
Percentage complete : 10.93%
SAS2IRCU: Command STATUS Completed Successfully.
SAS2IRCU: Utility Completed Successfully.
- Oracle Hardware Management Pack:
The third utility that can be used to administer hardware raid volumes is raidconfig. Like the fwupdate command, raidconfig is part of the HMP - Oracle Hardware Management Pack which is discussed on an earlier slide. Documentation for the utility can be found:
Errors might be returned if you are attempting to configure the RAID entry to an unsupported parameter. The list of error codes and strings specific to this tool are show in the . Sample commands are as follows:
# raidconfig list all -v - Collect the most possible data for troubleshooting
# raidconfig list disk - Listing disks
# raidconfig list raid -v - lists all the raid volumes for either controller.
# raidconfig create raid -c c0 --level 1 -d c0d2,c0d3 - Create raid
Please see the TOI on the Troubleshooting page for sample output.
Boot Paths for internal disks, both used in HW Raid and not
***********************************
Example
{0} ok probe-scsi-all
/pci@400/pci@2/pci@0/pci@e/scsi@0
FCode Version 1.00.60, MPT Version 2.00,
Firmware Version 9.00.00.00
Target 9
Unit 0 Disk SEAGATE
ST960005SSUN600G 0606 1172123568 Blocks, 600 GB
SASDeviceName
5000c500437b4c8b SASAddress
5000c500437b4c89 PhyNum
0
|
---<-used for non hw raid boot path
Target a
Unit 0 Removable Read Only device TEAC DV-W28SS-R 1.0C
SATA device
PhyNum 7
Target
b
Unit 0 Disk SEAGATE
ST960005SSUN600G 0606 1172123568 Blocks, 600 GB
SASDeviceName 5000c500437e934b SASAddress 5000c500437e9349 PhyNum 1
Target 381
Volume 0
Unit 0 Disk LSI Logical Volume 3000 1169920000 Blocks, 598 GB
VolumeDeviceName 3191d94818ab05da VolumeWWID 0191d94818ab05da |
^-used for hw raid boot path
/pci@400/pci@1/pci@0/pci@b/pci@0/usb@0,2/hub@2/hub@3/storage@2
Unit 0 Removable Read Only device AMI Virtual CDROM 1.00
{0} ok
Assuming you have successfully installed Solaris into a RAID LUN, when booting the device from
OBP , boot the RAID device LUN using the “VolumeDeviceName”.
For example 3191d94818ab05da is the WWN in the OBP boot command below
boot /pci@400/pci@2/pci@0/pci@4/scsi@0/disk@w3191d94818ab05da,0:a
# format
Searching for disks...done
AVAILABLE DISK SELECTIONS:
0. c0t5000C500437B4C8Bd0
/scsi_vhci/disk@g5000c500437b4c8b
1. c0t5000C500437E934Bd0
/scsi_vhci/disk@g5000c500437e934b
2. c3t3191D94818AB05DAd0
How to boot an internal HW Raid volume
To boot the hw raidinternal volume use the following path:
{0} ok boot /pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w0191d94818ab05da,0:a
Note that the path is derived from the
controller path output from probe-scsi-all.
Some systems have two controllers. You
must use the controller path the volume is under.
In this example it
is...../pci@400/pci@2/pci@0/pci@e/scsi@0 And to that added disk@w...and the VolumeDeviceName,
followed by ...... ,0:a
{0} ok boot /pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w3191d94818ab05da,0:a
How to boot an internal disk that is not part a HW Raid volume
To boot the non hw raided internal disk
in slot 0(indicated by PhyNum 0 from probe-scsi-all) use the following path:
{0} ok boot /pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w5000c500437b4c89,0:a
Note that the path is derived from the
controller path output from probe-scsi-all.
Some systems have two controllers. You
must use the controller path the disk is under.
In this example it is...../pci@400/pci@2/pci@0/pci@e/scsi@0 And to that added
disk@w...and the SASAddress, followed by ...... ,0:a
7. The third generation hard raid Chip LSI 2108
Applies to: SPARC T4-1/T4-2
- The HBA is supported on the following SPARC
Hardware Platforms ( as of April 2012 )
SPARC T3-1, SPARC T3-2, SPARC T4-1,SPARC T4-
2,Netra SPARC T3-1, Netra SPARC T4-1 and Netra
SPARC T4-2
- The HBA is NOT supported on SPARC T3-4 and
SPARC T4-4
- The HBA is also supported on x86 platform
running Solaris 10 x86, Oracle Linux, Oracle VM,
Red Hat Linux 4.0, 5.0 (32/64-bit), VMware ESX
server 3.x and MS Windows Standard/Enterprise
Server 2003 (32/64-bit)
- The HBA is an optional component and is factory
configured with the following description in the
yellow Customer Information Sheet (SG-SAS6-RINT-
Z).
- The HBA supports more RAID types vs. Default
On-Board HW Raid
( RAID 0,1,5,6,10,50 and 60 vs RAID 0, 1 and 10)
- The HBA has a battery backed memory
cache(512MB), providing better I/O performance
with a higher level of data protection.
- On SPARC T3-1 and SPARC T4-1 the RAID HBA
must be installed in PCI slot #3. Connect
P2(SAS0) on Disk Backplane to J7(Top) on RAID
HBA. Connect P3(SAS1) on Disk Backplaneto
J8(Bottom) on RAID HBA.
- The SAS6-R-INT-Z RAID HBA card is NOT
supported on SPARC T3-1 with 16 Disk
Backplane. There is no 16 Disk Backplane SPARC
T4-1.
- On a SPARC T3-2 and SPARC T4-2 the RAID HBA
card needs to be installed on PCI slot #0.
Note: The slot#0 isn't necessary for the Niwot for T3/4-2 installation
- The installation of the RAID HBA card will disable
the use of the internal DVD drive
- No Special Cable Kit is required
- Connect SAS 0-3 on Disk Backplane to RAID HBA
port J7 (Top), Connect SAS 4-5 to port J8
(Bottom) on RAID HBA card.
Chip: LSISAS2108 RAID on Chip ( ROC)
- Determine the type of the hard raid HBA
# iostat -En
e.g.
sd2 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0
Vendor: LSI Product: MR9261-8i Revision: 2.50 Device Id: id1,sd@n600605b004ae9620ff005b6472a8fea1
Size: 299.00GB <298999349248 bytes>
Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0
Illegal Request: 3 Predictive Failure Analysis: 0
# cfgadm -alv
e.g.
c3 connected configured unknown
unavailable scsi-bus n /devices/pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0:scsi
c3::dsk/c3t0d0 connected configured unknown LSI MR9261-8i
unavailable disk n /devices/pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0:scsi::dsk/c3t0d0
# prtdiag -v
e.g.
================================ IO Devices ================================
Slot + Bus Name + Model
Status Type Path
----------------------------------------------------------------------------
/SYS/MB/NET0 PCIE network-pciex8086,10c9
----------------------------------------------------------------------------
/SYS/MB/NET0 PCIE network-pciex8086,10c9
/pci@400/pci@1/pci@0/pci@4/network@0
/SYS/MB/NET1 PCIE network-pciex8086,10c9
/pci@400/pci@1/pci@0/pci@4/network@0,1
/SYS/MB/SASHBA PCIE scsi-pciex1000,72 LSI,2008
/pci@400/pci@2/pci@0/pci@e/scsi@0
/SYS/MB/NET2 PCIE network-pciex8086,10c9
/pci@500/pci@1/pci@0/pci@5/network@0
/SYS/MB/NET3 PCIE network-pciex8086,10c9
/pci@500/pci@1/pci@0/pci@5/network@0,1
/SYS/MB/PCIE1 PCIE LSI,mrsas-pciex1000,79 LSI,2108_4
/pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0
# prtconf -vp
e.g.
compatible: 'pciex1000,79'
model: 'LSI,2108_4'
reg: 000a0000.00000000.00000000.00000000.00000000.010a0010.00000000.00000000.00000000.00000100.030a0
014.00000000.00000000.00000000.00004000.030a001c.00000000.00000000.00000000.00040000.020a0030.00000000.00000000.00000000.0004
0000
version: '4.07.04.00'
wide: 00000010
device_type: 'scsi-sas'
name: 'LSI,mrsas'
fcode-rom-offset: 00008000
port-type: 'PCIE-Endpoint'
interrupts: 00000001
cache-line-size: 00000010
class-code: 00010400
subsystem-id: 00009263
subsystem-vendor-id: 00001000
# prtpicl -v
e.g.
:model LSI,2108_4
:compatible pciex1000,79
:assigned-addresses
- FW or patch for LSI 2108 CHIP:
Package Version Firmware Version (Mandatory Update)
12.12.0-0079 2.120.203-1440
Note: Before loading this new firmware, download and install the latest MegaCli for your OS and then use it to upgrade firmware.”
- The latest patch release notes:
************************************
LSI Corp. MegaRAID Firmware Release
************************************
Release date: 05/02/12
======================
Supported Controllers:
======================
MegaRAID SAS 9262-8i
MegaRAID SAS 9281-8i
*******THIS FIRMWARE DOES NOT SUPPORT CACHECADE PRO 2.0 OR CACHEVAULT CONTROLLERS.
*******PLEASE REFER TO DIFFERENT FIRMWARE IMAGES ON LSI.COM FOR CACHECADE 2.0 AND CACHEVAULT SUPPORT
Component:
=========
SAS MegaRAID Firmware Release for MegaRAID Controllers
Release date: 05/02/12
Version Numbers:
===============
Current Firmware Package: 12.12.0-0079
Current Package Details:
APP-2.120.203-1440_BB-2.02.00.00-0000_BIOS-3.24.00_4.12.05.00_0x05090000_HII-_WEBBIOS-6.0-46-e_41-Rel_PCLI-04.04-01900008_BOOT-09.250.01.219_NVDATA-2.09.03-0027_2011_11_08
Previous Firmware Package: 12.12.0-0048
Previous Package Details:
APP-2.120.63-1242_BB-2.02.00.00-0000_BIOS-3.22.00_4.11.05.00_0x05020000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-09.250.01.219_NVDATA-2.09.03-0013_2011_04_06
Bug Fixes and Enhancements:
===========================
Firmware:
=========
LSIP200209948 (CO) Generally available support of selected battery modes for iBBU08.
LSIP200211360 (DFCT) (CL LSIP200209847) Pulling VD 0 from an R60 and rebooting, 2 foreign configs when array 0
LSIP200199863 (DFCT) UnconfigGood PD fails to kick in as EHS (Emergency Hot Spare)
LSIP200212811 (DFCT) Advanced features key have to be re-entered
LSIP200050209 (DFCT) DATA TLB error exception in case of manual removal/insertion of drives.
LSIP200051385 (DFCT) R1 rebuild hangs at 50% mark after CC is run
LSIP200051392 (DFCT) Fw broke into MegaMon while writing a test file to a VD (R6) from a non MR disk.
LSIP200221925 (DFCT) SATA workaround handles LBA incorrectly for 6-byte read CDB
LSIP200212818 (DFCT) Port forward LSIP200211068 to main stream liberator maintenance
LSIP200212644 (DFCT) noticed additional fan speed event messages every 2 minutes
LSIP200198466 (DFCT) wrong slot # reported by megaraid compared to physical slot for EG disk
LSIP200199687 (DFCT) 揘on-Transparent?learns seen with Mode 3
LSIP200210349 (DFCT) BBUmode incorrectly set
LSIP200150196 (DFCT) MR does not pick up replacement drives in PD if drive is added during power cycle
LSIP200150197 (DFCT) NEM removal and re-insertion causes all VD members to change to Foreign Unconfigured Bad state
LSIP200210823 (DFCT) No data is copied to target array after more than 15 secs
LSIP200212264 (DFCT) Multibit ECC error with customer Liberator MR board and vmware ESXi5.0
LSIP200199735 (DFCT) LSI Mega RAID SAS-MFI BIOS Register state 0xF0010002
LSIP200211484 (DFCT) MR bios show "adapter at baseport is not repoonding" intermittence
LSIP200210596 (DFCT) too many messages in fw log "ArDiskTypeMisMatch : SAS_SATA_MIXING_VIOLATION?
LSIP200155948 (DFCT) FW seems to incorrectly mark ibbu's SOH bad (iBBU07, ibbu05, ibbu01)
LSIP200209279 (DFCT) additional reboot required after updating ibbu FCC to design capacity
LSIP200210505 (DFCT) OEM requests qty of prints to be reduced in fw log for ibbu i2c communication retries
LSIP200212471 (DFCT) FW automatically starts learn cycles on ibbu08 every reboot
LSIP200211810 (DFCT) Battery failed message during learn cycle
LSIP200212311 (DFCT) Battery failed message during learn cycle
LSIP200200117 (DFCT) unexpected Power State Change events occur
LSIP200185441 (DFCT) Disk performance disparity between WB controller cache on and off on system disks
LSIP200210751 (DFCT) Patrol Read fails on R1 with around 2000 double medium errors
LSIP200210768 (DFCT) Enabling security causes the config to be missing, and a configured FDE array to be seen as foreign on reboot.
LSIP200149248 (CO) Warning for High Error Counts
LSIP200149890 (CO) SW licensing. Support FoD for PF activations on iMR/2008
LSIP200187910 (CO) Request to use Unconfigured Good Drives for a HotSpare if no spare defined in liberator
LSIP200187724 (CO) MR fw should Enable puncturing for SATA drives
LSIP200187937 (CO) MegaRaid does not pick up replacement drives in PD if replacement is added during power cycle
LSIP200151375 (CO) Need StopOnError functionality when SafeStore passphrase is enabled
LSIP200187926 (CO) Change BBU-08 support Firmware Mode5 WB capacity from 960mAh to 674mAh
LSIP200187931 (CO) disable bad phy checking logic from iMR FW for certain ROMB design
LSIP200158631 (CO) Alert should be sent when one phy connector between Liberator card and SAS expander is degraded
LSIP200199858 (CO) add display of ibbu communication retry counter in 10M09 fw log
LSIP200199484 (CO) Enable copy back on SMART - Liberator
LSIP200186977 (CO) Alert should be sent when one phy connector between Liberator card and SAS expander is degraded
LSIP200149050 (CO) Add SGPIO for Medusa
LSIP200186343 (CO) (LSIP200140211) Add support to pass SMART errors for JBOD drives to MSM
LSIP200185599 (DFCT) LED behavior incorrect, drive 3 blinks when drive 0 is accessed
LSIP200100032 (DFCT) system occasionally hangs when executing an "init 0"
LSIP200209385 (DFCT) Intermittent RAID key failure during power-on
LSIP200209329 (DFCT) FoD features will not activate or deactivate
LSIP200150153 (DFCT) Strange HDD Led behaviour in X4170M2 under raid configuration
LSIP200156822 (DFCT) CC2.0: WT Downgrade FW from CC2.0 to CC1.0(10M09) and import caused MonTask: line 199 in file ../../raid/1078int.c
LSIP200149201 (DFCT) (CL LSIP200146291) Medium error handling fixes for >2TB
LSIP200199812 (DFCT) Alarm sounds with online VD's after flashing latest 10M09 Patch 16 firmware (2.120.163-1343)
LSIP200199961 (DFCT) (CL LSIP200061413) LDBBM entries are incorrect after foreign import
LSIP200186336 (DFCT) Controller does not work with pci set to nomsi in kdump kernel
LSIP200209455 (DFCT) (CL LSIP200188135) Potential data integrity issue running the IO on degraded R5 VDs with 32 drives.
LSIP200158264 (DFCT) Change the capacity level that generates the Predictive failure alert
LSIP200150304 (DFCT) memory test for 2108 in 10M09 FW fails immediately
LSIP200149986 (DFCT) MSM log shows "Battery cannot initiate transparent learn cycles" due to BBU is under mode 5.
LSIP200060949 (DFCT) iBBU07 reported remaining capacity is much too high (64751 mAh) after a learn cycle
LSIP200151502 (DFCT) BBU Absolute Charge is displayed differently
LSIP200149843 (DFCT) VPD (Vital Product Data) needs to show correct Part number
LSIP200149844 (DFCT) VPD (Vital Product Data) needs to identify PCB hardware level
LSIP200199416 (DFCT) Change BBU-08 support Firmware Mode5 WB capacity from 960mAh to 674mAh for low power DDR-2
LSIP200199331 (DFCT) 10M09: Older ibbu batteries display 65K capacity and are marked SOH bad
LSIP200199335 (DFCT) 10M09: Request FW to not mark ibbu battery as bad when very high capacity is detected
LSIP200196885 (DFCT) Batteries Not Charging
LSIP200209376 (DFCT) HWR: 10M09: OEM requests ibbu i2c communication retries to be displayed in fw log
LSIP200147629 (DFCT) (CL LSIP200167360) FW hit montask upon importing foreign config
LSIP200147626 (DFCT) (CL LSIP200139418) System reboots after issuing OCR
LSIP200147907 (DFCT) Disable Dimmer Switch III
LSIP200151221 (DFCT) CLI/EFI CLI/PCLI fail to create a VD on MR
LSIP200148150 (DFCT) WHQL WLK1.6 - SCSI Compliance Test 2.0 and Read Capacity 16 Unit FAILS on Liberator
LSIP200184970 (DFCT) HWR: 10M09: not able to disable BGI with MegaCLI
LSIP200173287 (DFCT) HWR: bbuMode is not in WebBIOS after flashing from 10M05 to 10M09
LSIP200186884 (DFCT) HWR: 10M09: DelaySOD is not working on customer Liberator boards
LSIP200150532 (DFCT) (RW LSIP200145892) (6997086) (RW LSIP200121969) Liberator FW 10M09 cause system (sles11sp1/rhel6) reset while lspci -vvx
LSIP200188226 (DFCT) HWR 10M09 can't sustain at least 950 MBps performance with Cache Flush setting 1 sec
LSIP200200459 (DFCT) MegaCli prematurely returns exit 0x0 status on drive firmware flash
LSIP200209349 (DFCT) Enable PCI capability Bit for VPD exsupport.
LSIP200151328 (DFCT) HWR: EF & 10M09: bluescreen occurs handling R1 with HSP and medium errors
LSIP200186966 (DFCT) HWR: 10M09: configuration becomes foreign after fw fails 3-4 attempts to access raid key during boot
LSIP200147621 (DFCT) (CL LSIP200163913) performance issues within 9285 and 9280 caused by the MegaRAID FW
LSIP200147632 (DFCT) (CL LSIP200173232) Potential data integrity issue found on R1 VDs while injecting the media errors under IO.
LSIP200186406 (DFCT) Code Review defect to deliver rest of changes for PFK preboot support.
LSIP200186594 (DFCT) R5 pfk feature not getting enabled at boot time on iMR
LSIP200186595 (DFCT) After loading FoD key, iMR controller does not perform upgrade of features
LSIP200188345 (DFCT) (CL LSIP200185468) HWR: 10M09: extended retries of write verify command causes driver timeout
LSIP200171125 (DFCT) No event is generated when fan in enclosure fails
LSIP200184943 (DFCT) Need ability to boot with pinned cache with any raid config on Gen 2 MegaRAID/Liberator defaulted to ON
LSIP200185931 (DFCT) Bootwithpinnedcache not default set to '1' for castorlite when suboem is '1'
LSIP200186584 (DFCT) Logical Unit ID (VPD page 0x83) changed after foreign import
LSIP200198521 (DFCT) 10M09 P16 set properties command fails
LSIP200146463 (DFCT) Update OOB (BMC) management
LSIP200171812 (DFCT) Liberator Firmware : Fatal F/W Error* happened when detected drive error** during Read I/O.
LSIP200169751 (DFCT) MegaRaid FW does not wait at post
Webbios:
========
LSIP200210939 (DFCT) (N/A) System showed error message when entering uFEI web-bios
LSIP200210311 (DFCT) Adding OK button instead of cancel in "Advance software Options" page
LSIP200199865 (CO) MegaRAID Missing Stop Locator for Volume
LSIP200209946 (CO) Generally available support of selected battery modes for iBBU08.
LSIP200150553 (CR) MegaRAID Missing Stop Locator for Volume
LSIP200187637 (CR) Generally available support of selected battery modes for iBBU08.
LSIP200199949 (CO) SW licensing. Support FoD for PF activations on MegaRAID
LSIP200150642 (CO) 10M09 Firmware release has Advanced Software Options entry that should be removed
LSIP200188273 (DFCT) uEFI WebBIOS incorrectly displays "UG Spin Down" title
LSIP200139977 (DFCT) Option for populating the VD size field automatically
LSIP200156291 (DFCT) can't create DG with 8 PDs with enableReducedFeatureSet
LSIP200157087 (DFCT) EFI WebBIOS does not show product name/version for second adapter in adapter selection screen
LSIP200121693 (DFCT) Unable to configure Cachecade in webBIOS under 10M09 for 2108 based products
LSIP200166733 (DFCT) WEBBIOS is unable to reconstruct the R11 VD to R5 with removal of one PD.
LSIP200147461 (DFCT) Cannot change the BBU properties in webbios
LSIP200149524 (DFCT) WebBIOS gives garbled message when enabing FDE on an existing Disk Group
BIOS:
========
LSIP200199737 (DFCT) server System posts 'R00' as 'R0' during the reboot
LSIP200199163 (DFCT) Legacy BIOS BBS Boot Issues
Numerous UEFI and HII updates
LSIP200196810 SWR-Macon: After installing the OS, unable to boot from SCU VD
LSIP200199602 FW hangs during POST after flashing new BIOS
LSIP200018807 BIOS sees extra characters "0%" at end of particular product name
LSIP200156493 There is no space between FW version and capacity in the table where Adp/VD/PD info is displayed
Numerous HII fixes
LSIP200209201 (DFCT) UEFI: HII Create Configuration not all virtual drive changed parameters are saved
LSIP200150135 (DFCT) Cannot configure VD on Niwot using UEFI HII menu, problem reported by AMI
LSIP200150702 (DFCT) Attempting to create a virtual drive using UEFI HII resulted in a "hang"
LSIP200187671 (DFCT) UEFI driver 05050000 breaks HII menu in server f2 Bios
NVDATA:
=======
Numerous NVDATA changes
Installation:
=============
Use MegaCLI to flash the SAS controllers. MegaCLI can be downloaded from the support and download
section of .
Command syntax: MegaCli -adpfwflash -f mr2108fw.rom -a0
=================================
Previous Releases
=================================
Component:
=========
SAS MegaRAID Firmware Release for MegaRAID Controllers
Release date: 04/15/11
Version Numbers:
===============
Current Firmware Package: 12.12.0-0048
Current Package Details:
APP-2.120.63-1242_BB-2.02.00.00-0000_BIOS-3.22.00_4.11.05.00_0x05020000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-09.250.01.219_NVDATA-2.09.03-0013_2011_04_06
Previous Firmware Package: 12.12.0-0045
Previous Package Details:
APP-2.120.33-1197_BB-2.02.00.00-0000_BIOS-3.21.00_4.11.05.00_0x05000000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-01.250.04.219_NVDATA-2.09.03-0010_2011_02_25
Bug Fixes and Enhancements:
===========================
Firmware:
=========
LSIP200167677 (CO) FW should issue STANDBY IMMEDIATE command to SATA disks during shutdown from OS
LSIP200164222 (CO) Add mfc paramater that enables prints of topology changes in fw serial port output
LSIP200158868 (DFCT) Power save mode hangs firmware causing driver to send kill adpter command
LSIP200170440 (DFCT) Firmware Incompleted Writes During Power Fail and Recovery
LSIP200153454 (DFCT) Firmware Incompleted Writes During Power Fail and Recovery
LSIP200125226 (DFCT) Red LED is lit on the next empty slot in an specific JBOD if the empy slot is immediatly after an array member
LSIP200153500 (DFCT) 9280-4i4e unable to import a Foreign Configuration
LSIP200166039 (DFCT) Label of Virtual Disk not provided by Drive Identify Command
LSIP200164210 (DFCT) Enable/disable of HDD cache doesn't appear to work
LSIP200167654 (DFCT) Potential data mismatch found when rebuild is progressing on R1 VD with media error under IO.
LSIP200166899 (DFCT) HWR: Fw log should indicate if a created/existing VD is already initialized
LSIP200157150 (DFCT) The new event can't be logged.
LSIP200156699 (DFCT) MegaCli -AdpAllInfo shows BBU is not supported for Castor cards
LSIP200146481 (DFCT) SMART predictive failure detection takes longer in 10M09 than 10M06
LSIP200146484 (DFCT) MR FW reports "Unexpected sense, Invalid field in CDB" for LOG SENSE command sent to Seagate drives
LSIP200169751 (DFCT) MegaRaid FW does not wait at post
BIOS:
========
HII changes only
Installation:
=============
Use MegaCLI to flash the SAS controllers. MegaCLI can be downloaded from the support and download
section of .
Command syntax: MegaCli -adpfwflash -f mr2108fw.rom -a0
- Niwot HBA practice
Lab 1 : 练习Niwot卡的管理及使用
Hardware Resource : T4-2/ 375-3701 [C] 6Gigabit SAS RAID PCI Express HBA, B4 ASIC
Software Resource : Fcode utility/ LSI SAS2 2008 RAID Management Utility/ Oracle Hardware Management Pack
Task 1 : Identify internal disks and create hard raid volumes by Fcode
{0} ok select LSI2008 raid chip physical path
{0} ok show-volumes
{0} ok show-children
create a volume for root disk
Task2 : Install OHMP and use OHMP or sas2ircu to identify volume status; Identify physical disk location
a. Download OHMP from MOS
b. Download sas2ircu from LSI website
c. Show volume status and identify disk members in physical slot by command output
Task3: Dummy replace the 371-4982 6Gigabit SAS RAID PCI Battery Module and connect disk cables to Niwot card
a. Remove and reinstall the BBU
b. Connect two disk cables from motherboard to Niwot card
Task4: Practice the PCLI commands under OBP
a. Find out the Niwot physical path
b. Use PCLI commands get Niwot configuration
1. “cli -PdList” will display all the physical disk
under the HBA's control.
2. “cli -LDInfo -Lall” will display current LUNs
associated with the HBA. LUNs can also be
referred to as a Virtual Drive.
3. “cli -EncInfo” shows more information about the
Enclosure, like enclosure id and # of enclosure
slots.
4. “cli -CfgDsply” displays information about HBA
( Serial No., Backup Battery Unit presence, Cache
Memory Size and LSI Product Name)
“cli -LdPdInfo” maps create LUNs with physical
Disks.
c. Create RAID volume, show volume status and delete volume
1. ok cli -CfgLdAdd -r1[252:0,252:1]
{0} ok cli -LDInfo -L0
{0} ok cli -CfgLdDel -L0
2. ok cli -CfgLdAdd -r0[252:0]
{0} ok cli -LDInfo -L0
3. ok cli -CfgLdAdd -r5[252:2,252:3,252:4]
{0} ok cli -LDInfo -L1
d. Install Solaris in volume
Task5: Install MegaCli and practice to use it
a. Download and install the MegaCli package for Solaris SPARC
Download URL:
b. Verify MegaCli installation and Niwot card status
# ./MegaCli -v
# ./MegaCli -Version -Ctrl -aALL
# ./MegaCli -Version -Driver -aALL
# ./MegaCli -LDInfo -Lall -aALL
# ./MegaCli -CfgDsply -a0
How Many Disks seen by HBA
# ./MegaCli -PDGetNum -a0
Check Battery Status
# ./MegaCli -AdpBbuCmd -GetBbuStatus -aALL
More Backup Battery Info
# ./MegaCli -AdpBbuCmd -GetBbuCapacityInfo -aALL
HBA Cache Policy
# ./MegaCli -LDGetProp -Cache -LALL -aALL
c. Collect HBA event Logs:
# ./MegaCli -AdpEventLog -GetEvents -f eventlogs.txt -a0
# ls -l
Collecting HBA Terminal Log:
# ./MegaCli -fwtermlog -dsply -a0 > terminallog.txt
ls -l
# ./MegaCli -fwtermlog -Clear -a0
d. Locate Disk ( LED Blink ):
./MegaCli -PdLocate -start -physdrv[252:0] -a0
./MegaCli -PdLocate -stop -physdrv[252:0] -a0
The Attention LED (AMBER Light) on the disk will blink until the
disk locate is manually terminated.
e. Adjusting Rebuild Rate:
./MegaCli -AdpGetProp RebuildRate -a0
./MegaCli -AdpSetProp RebuildRate -70
(Default is 30%, higher rebuild rate will adversely impact RAID I/O
Performance)
f. Convert Raid 0 to Raid 1 online:
When system is pre-configured from Factory with the NIWOT card, usually only one Raid 0 LUN is
created by default with Solaris OS pre-installed. If the customer wants to have a RAID 1 disk
configured we can Download the MegaCLI package and run the following command to convert the
RAID 0 to a RAID 1 without re-installing the OS. This activity May take a few hours depending on
disk size, disk performance and system load.
# ./MegaCli -LDRecon -Start -r1 -Add -PhysDrv[252:1] -L0 -a0
g. Adding/Removing Global HotSpare:
# ./MegaCli -PDHSP -Set -PhysDrv [252:5] -a0
# ./MegaCli -PDHSP -Rmv -PhysDrv[252:5] -a0
h. Convert Raid 5 to Raid 6 online:
# ./MegaCli -LDRecon -Start -r6 -Add -PhysDrv[252:5] -L1 -a0
# ./MegaCli -LDRecon -ShowProg -L1 -a0
# ./MegaCli -LDRecon -ProgDsply -L1 -a0
# ./MegaCli -LDInfo -Lall -aALL