Internal Hard Raid evolvement

8705阅读 0评论2012-10-19 billjoy
分类:系统运维

Chip type

Server or HBA card

1. LSI1030

V440

3. LSI1064

 Sun Fire V215 [TM] Server, Sun Fire V245, Sun Fire T1000

4. LSI1064E

Sun Fire T2000, Sun Fire Netra T2000

2. LSI1068

V445

5. LSI1068E

Sun Fire T5120/T5220/T5140/T5240/T5440/NetraT5220/NetraT5440

6. LSI2008

T3-1/T3-2/T3-3/T4-1/T4-2/T4-4

7. LSI2108

Niwot card

 

 

 

 

1. The first generation hard raid Chip LSI1030

 

Applies to: Sun Fire V440 [TM] Server, Netra[TM] 440 Server

 

Chip: LSI1030 SCSI Host Adapter

 

 

# iostat -En

 

e.g.

c1t0d0 LSILOGIC Product: 1030

 

# cfgadm -alv

 

e.g. c1t0d0 LSILOGIC 1030 /devices/pci@1f,700000/scsi@2:scsi::dsk/c1t0d0

 

# prtdiag -v

 

e.g. scsi-pci1000,30 LSI,1030 /pci@1f,700000/scsi@2

 

# prtconf -vp

 

e.g.

LSI1030 FCode Driver: 1.8 03/04/17 scsi: '/pci@1f,700000/scsi@2'firmware-version: '1.03.11.01'

 

# prtpicl -v

 

e.g.

LSI1030 FCode Driver: 1.8 03/04/17 devfs-path /pci@1f,700000/scsi@2

 

 

Remember to install the right and best OBP version, because OBP contains the proper driver to communicate with FCode on HBA.

Current firmware/FCode:

lsi1030-fw-1.03.11.01

 

Above codes can be reached either by factory default implementation, proper OBP-upgrade or by use of following patches

lsi1030-fw-1.03.11.01

patch 115662-01

Hardware/Firmware: LSI1030 U320 SCSI controller

 

 

Keywords: mpt raidctl firmware lsi1030 sun fire v440 v440

Synopsis: Hardware/Firmware: LSI1030 U320 SCSI controller

Date: Sep/24/2003

 

 

Install Requirements: See Special Install Instructions

Additional instructions may be listed below

 

Solaris Release: 8 9

 

SunOS Release: 5.8 5.9

 

Unbundled Product: LSI1030 Firmware

 

Unbundled Release: N/A

NOTE: Sun Fire V440 onboard LSI1030 controller

Xref:

 

Topic:

 

Relevant Architectures: sparc

 

BugId's fixed with this patch: 4834770 4924126

 

Changes incorporated in this version: 4924126 4834770

 

Patches accumulated and obsoleted by this patch:

 

Patches which conflict with this patch:

 

Patches required with this patch:

 

Obsoleted by:

 

Files included with this patch:

 

lsi1030-fw-1.03.11.01

 

Problem Description:

 

4924126 Connected Command timeouts seen on Seagate disks running stress tests
4834770 1030 firmware does not clean up IOs to a RAID after a bus reset

 

Patch Installation Instructions:

1) Copy the patch into any directory, say .
   (Note, if the patch ends in a "tar.Z" suffix, you will need to first
   uncompress and untar it.)  So, you would type:
        % su root
        # cd
        # uncompress
        # tar xf

2) The LSI1030 uses the 'mpt' driver.  Run the following 'grep' command
   to find all device paths which this driver is associated with.

#  grep "mpt" /etc/path_to_inst
"/pci@1f,700000/scsi@2" 0 "mpt"
"/pci@1f,700000/scsi@2,1" 1 "mpt"

3) Run the 'format' command from the prompt to see what disks you have
   in the system that are associated with these device paths:

# format
Searching for disks...done

AVAILABLE DISK SELECTIONS:
       0. c1t0d0
          /pci@1f,700000/scsi@2/sd@0,0
       1. c1t1d0
          /pci@1f,700000/scsi@2/sd@1,0
       2. c1t2d0
          /pci@1f,700000/scsi@2/sd@2,0
       3. c1t3d0
          /pci@1f,700000/scsi@2/sd@3,0
Specify disk (enter its number):

Determine all disks that are associated with all the device paths
identified in step 2.  In this case the disks on the
"/pci@1f,700000/scsi@2/" LSI1030 mpt driver path are assigned controller
number 'c1'.  This will be controller we will update using just the
number '1'.

Type ^D to quit out of 'format'

4) Update the firmware on your controller.

Use the '/usr/sbin/raidctl' utility and type:

# /usr/sbin/raidctl -F lsi1030-fw-1.03.11.01 1

'1' in this case is your controller id.

This will ask you if you want to continue updating the firmware.

ROM image contains MPT firmware version 1.03.11 (w/Integrated Mirroring)
Update flash image on controller 1 (yes/no)?

Type 'yes' and hit enter.

Once you are returned to the prompt the update is complete.

5) Note that the "/pci@1f,700000/scsi@2,1" instance is a 2nd channel on
the same LSI1030 controller identified by the ",1" portion of the device
path being appended to the 1st channel path "scsi@2".  This does not
need to be explicitly upgraded if disks attached to the 1st channel of
the LSI1030 are already upgraded in step 4.

Repeat steps 3 and 4 for each unique LSI1030 controller identified in
step 2.

6) Reboot the system after all the controllers have been updated.

 

Special Install Instructions:

-----------------------------
 
It is preferable that this update be done when disk activity is a minimum.

The system must be rebooted for the new firmware version to appear in prtconf.
After reboot the firmware version can be verified as follows:

# /usr/sbin/prtconf -vp | grep 'firmware-version:'
            firmware-version:  '1.03.11.01'

 

 

 

Description

This document provides the procedure to create or delete a hardware raid

 

mirror in a Sun Fire[TM] V440. Support engineers should check the procedures

 

in this document when assisting customers who may be asking for

 

guidance in this process.

 

 

 

Steps to Follow

How to Create a Hardware Disk Mirror

Note:

To create a raid  volume, the disk being mirrored must have a t#

lower than the t# of the mirror disk.  This requirement is due to the

interaction between the way Solaris[TM] handles the boot disk and the

built-in feature of the hardware raid controller that cannot be

disabled.

Caution   Creating or restoring a disk mirror destroys all data previously stored on the

disk drive.

Perform this procedure to create an internal hardware disk mirror configuration on

your system.

Verify which disk drive corresponds with which logical device name and physical

device name.   Example: Type

# raidctl

No RAID volumes found.

The example above indicates that no RAID volume exists. In another case:

# raidctl

          RAID            RAID             RAID       Disk

Volume Status                 Disk       Status

 ----------------------------------------------------------------------------------

c1t1d0 DEGRADED c1t1d0                   OK

                       c1t2d0 DEGRADED

The example above indicates a hardware mirror has degraded at disk c1t2d0.

-------------------------------------------------------------------------------------------------------------

Note -  The logical device names might appear differently on your system, depending

on the number and type of add-on disk controllers installed.

What to Do

1. Type the following command: # raidctl -c master slave

For example: # raidctl -c c1t0d0 c1t1d0

When you create a RAID mirror, the slave drive (in this case, c1t1d0) disappears from

the Solaris device tree.

2. To check the status of a RAID mirror, type the following command:    # raidctl

 RAID RAID RAID Disk

Volume Status Disk Status -----------------------------------------------------------------------

 c1t0d0 RESYNCING c1t0d0 OK

c1t1d0 OK

The example above indicates that the RAID mirror is still resynchronizing with the

backup drive.

The example below shows that the RAID mirror is completely restored and online.

# raidctl

RAID RAID    RAID  Disk

Volume Status     Disk Status

c1t0d0              OK           c1t0d0          OK

                               c1t1d0          OK

How to Delete a Hardware Disk Mirror

Perform this procedure to remove a hardware disk mirror configuration from your

system.

Verify which disk drive corresponds with which logical device name and physical

device name.

Disk Slot Numbers, Logical Device Names, and Physical Device Names

Disk Slot Number Logical Device Name* Physical Device Name

Slot 0 c1t0d0 /devices/pci@1f,700000/scsi@2/sd@0,0

Slot 1 c1t1d0 /devices/pci@1f,700000/scsi@2/sd@1,0

Slot 2 c1t2d0 /devices/pci@1f,700000/scsi@2/sd@2,0

Slot 3 c1t3d0 / /devices/pci@1f,700000/scsi@2/sd@3,0

______________________________________________________________________

*The logical device names might appear differently on your system, depending on the

number and type of add-on disk controllers installed.

  1. Determine the name of the mirrored volume. Type the following command:

# raidctl

RAID RAID RAID Disk

Volume Status Disk Status

c1t0d0 OK c1t0d0 OK

c1t1d0 OK

In this example, the mirrored volume is c1t0d0.

Note   The logical device names might appear differently on your system,

depending on the number and type of add-on disk controllers installed.

  1. To delete the volume, type the following command:   # raidctl -d mirrored-volume

For example: # raidctl -d c1t0d0

       RAID Volume  c1t0d0  deleted

  1. To confirm that you have deleted the RAID array, type the following command:

# raidctl

For example: # raidctl

No RAID volumes found

 

Note:

LSI1030 HBA supports only 1 RAID volume (IM, Integrated Mirror) on V440

H/W mirroring can be initiated after installation of data onto physical disk.

Backwards, breaking the mirror will give two disks with valid data on each.

 

 

2.  The first generation hard raid Chip LSI1068

 

Applies to: Sun Fire V445 [TM] Server

 

Chip: LSI SAS1068 controller is mounted on PCI-X slot on V445 for up to 8 disks

 

 

 

# iostat -En

 

e.g.

sd1       Soft Errors: 2 Hard Errors: 0 Transport Errors: 0

Vendor: LSILOGIC Product: Logical Volume   Revision: 3000 Serial No: 

Size: 73.01GB <73012215808 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

 

# cfgadm -alv

 

e.g.

c1                             connected    configured   unknown

unavailable  scsi-bus     n        /devices/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1:scsi

c1::dsk/c1t0d0                 connected    configured   unknown    LSILOGIC Logical Volume

unavailable  disk         n        /devices/pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1:scsi::dsk/c1t0d0

 

# prtdiag -v

 

e.g.

 

pci     199   MB/PCI4     LSILogic,sas-pci1000,54 (scs+ LSI,1068          

              okay        /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1

 

 

# prtconf -vp

 

e.g.

                            firmware-version:  '1.08.04.00'

                            mpt-version:  '1.05'

                            assigned-addresses:  81080810.00000000.00002000.00000000.00000100.82080814.00000000.04600000.00000000.00010000.8208081c.00000000.04610000.00000000.00010000.82080830.00000000.04800000.00000000.00200000

                            compatible: 'pci13e9,54' + 'pci1000,54' + 'AAPL,MPT'

                            model:  'LSI,1068'

                            reg:  00080800.00000000.00000000.00000000.00000000.01080810.00000000.00000000.00000000.00000100.03080814.00000000.00000000.00000000.00010000.0308081c.00000000.00000000.00000000.00010000.02080830.00000000.00000000.00000000.00200000

                            version: '1.00.39'

                            wide:  00000010

                            device_type:  'scsi-2'

                            name:  'LSILogic,sas'

 

# prtpicl -v

 

e.g.

                              :model         LSI,1068

                              :compatible        (a30000093fTBL)

                               | pci13e9,54 |

                               | pci1000,54 |

                               | AAPL,MPT |

                              :assigned-addresses        

 81  08  08  10  00  00  00  00  00  00  20  00  00  00  00  00  00  00  01  00

 82  08  08  14  00  00  00  00  04  60  00  00  00  00  00  00  00  01  00  00

 82  08  08  1c  00  00  00  00  04  61  00  00  00  00  00  00  00  01  00  00

 82  08  08  30  00  00  00  00  04  80  00  00  00  00  00  00  00  20  00  00

                              :mpt-version         1.05

                              :pcix-capid-pointer         0x68

                              :ddi-intr-weight         0xa

                              :mpxio-disable         yes

                              :ddi-vhci-class         scsi_vhci

                              :fm-ereport-capable

                              :fm-errcb-capable

                              :fm-dmachk-capable

                              :fm-accchk-capable

                              :pm-components        (a300000930TBL)

                               | NAME=mpt0 |

                               | 0=Off (PCI D3 State) |

                               | 3=On (PCI D0 State) |

                              :firmware-version         1.8.4.0

 

 

 

sas1068 on V445

123485-01

 

Hardware/PROM Sun Fire V445 1068X update 1.0.39

In general HBA fcodes are part of obp code on machines with on-board HBAs.

Every OBP- or FCode upgrade requires a power-cyle, not just a reboot.

 

 

 

 

Patch-ID# 123485-01

 

NOTE:

***********************************************************************

READ THE TERMS OF THE AGREEMENT ("AGREEMENT") IN THE LEGAL_LICENSE.TXT

FILE CAREFULLY BEFORE USING THIS SOFTWARE. BY USING THE SOFTWARE, YOU

AGREE TO THE TERMS OF THIS AGREEMENT. IF YOU DO NOT AGREE TO ALL OF THE

TERMS, PROMPTLY DESTROY THE UNUSED SOFTWARE.

***********************************************************************

 

Keywords: sun fire v445 1068x_firmware

Synopsis: Hardware/PROM Sun Fire V445 1068X update 1.0.39

Date: Feb/02/2007

 

 

Install Requirements: Additional instructions may be listed below                     

                     

Solaris Release: 9 10

 

SunOS Release: 5.9 5.10

 

Unbundled Product: Sun Fire Hardware 1068X PROM

 

Unbundled Release: 1.0.39

 

Xref:

 

Topic: Sun Fire V445 1068X FCode update 1.0.39

 

Relevant Architectures: sparc

NOTE:

Adds RAID Migration Support to 1068X FCode by enabling the activate-volume function.

NOTE:

The 1068X SAS Controller is used as the embedded disk controller on the V445. A small number of systems were shipped with 1068X FCode at revisions below 1.0.39. If you are using the embedded 1068X to control hardware RAID volumes  and the FCode is below revision 1.0.39 the activate-volume function is not available. The activate-volume function is needed if the RAID Set is moved or the 1068X controller replaced.  Please refer to your V445 Server Administration Guide for information on using the 1068X to create and control hardware RAID volumes.

 

BugId's fixed with this patch: 6518779

 

Changes incorporated in this version: 6518779

 

Patches accumulated and obsoleted by this patch:

 

Patches which conflict with this patch:

 

Patches required with this patch:

 

Obsoleted by:

 

Files included with this patch:

 

Install.info

LSI1068X.ROM

 

Problem Description:

 

6518779 LSI1068X Does not support Activate Volume

 

Patch Installation Instructions:

==============================

 

Please refer to the Install.info file for instructions on updating

the 1068X FCode using the files included in this patch.

 

Special Install Instructions:

-----------------------------

 

Unzip the patch into a directory and then follow the instructions provided

in the file Install.info.

 

README -- Last modified date:  Friday, January 26, 2007

 

README -- Last modified date:  Friday, February 2, 2007

 

 

 

Patch Installation Instructions:

===============================

 

ABOUT THIS PATCH

-------------------------

 

This patch updates the PCI-X 1068 FCode:

 

 To version   1.0.39

 

 

TO CHECK THE VERSION OF THE Firmware

--------------------------------------------------

 

There are two ways to check the  PCI-X Firmware version:

 

 o A. From the Solaris Operating System (OS)

 o B. From the OpenBoot PROM (OBP)

 

A. From the Solaris OS (2 steps):

 

The system prompt is shown as bash-3.00#, your prompt will vary depending on

your individual setup and shell in use.

 

Step 1. As root, run the prtconf command.

 

bash-3.00# prtconf -vp | egrep '(model|version)' |\

> egrep -vi '(obp|firmwareversion| mpt-version)'

    model:  'SUNW,501-7066'

        model:  'SUNW,4.22.11'

        sparc-version:  00000009

        sparc-version:  00000009

        sparc-version:  00000009

        sparc-version:  00000009

                            model:  'm5823'

                            firmware-version:  '1.08.04.00'

                            model:  'LSI,1068'

                            version: '1.00.39'

            model:  'SUNW,525-2274'

bash-3.00#

 

Step 2: Check the FCode Version

 

Locate the lines model: followed by version:, the version data is the FCode

version. In the example shown the version is 1.00.39 and no update is required.

 

 

B. From OBP (4 steps):

 

Check the version as follows, the ok prompt is shown as {1} ok in this

example, your prompt may vary depending on the number of processors installed:

 

 

Step 1: Select the LSI1068X Controller

 

{1} ok select /pci@1e,600000/pci@0/pci@1/pci@0/pci@8/LSILogic,sas@1

 

Step 2: Display information from the LSI1068X controller

 

{1} ok .properties

firmware-version         1.08.04.00

mpt-version              1.05

assigned-addresses       81060810 00000000 00000000 00000000 00000100

                         82060814 00000000 00800000 00000000 00010000

                         8206081c 00000000 00810000 00000000 00010000

                         82060830 00000000 00a00000 00000000 00200000

compatible               pci13e9,54

                         pci1000,54

                         AAPL,MPT

model                    LSI,1068

reg                      00060800 00000000 00000000 00000000 00000000

                         01060810 00000000 00000000 00000000 00000100

                         03060814 00000000 00000000 00000000 00010000

                         0306081c 00000000 00000000 00000000 00010000

                         02060830 00000000 00000000 00000000 00200000

version                  1.00.39

wide                     00000010

device_type              scsi-2

name                     LSILogic,sas

fcode-rom-offset         00000000

66mhz-capable

devsel-speed             00000001

latency-timer            00000040

cache-line-size          00000010

max-latency              0000000a

min-grant                00000040

interrupts               00000001

class-code               00010000

subsystem-id             00003031

subsystem-vendor-id      00001000

revision-id              00000001

device-id                00000054

vendor-id                00001000

{1} ok

 

Step 3: Check the FCode Version

 

Locate the line starting with version, this is the FCode version. In the

example shown the version is 1.00.39 and no update is required.

 

Step 4: Un-select the node before booting the server.

 

{1} ok unselect-dev

{1} ok

 

TO UPDATE THE 1068X Firmware

------------------------------

 

Step 1: Boot the Solaris OS.

 

Step 2: Login as root.

 

Step 3: Copy the new 1064X Firmware image onto the system.

 

The new 1064 Firmware image is included in this patch and is named

LSI1068XROM. Copy the image into a directory on the system which you

are going to update or to a network accessible directory. Use this

full path along with the  image name in step 5) below.

 

Step 4. Identify the controller.

 

        Determine the controller number via the format command. In the example

        shown the controller is number 0. Exit format with ^C

 

bash-3.00# format

Searching for disks...done

 

 

AVAILABLE DISK SELECTIONS:

       0. c0t0d0

          /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@0,0

       1. c0t2d0

          /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@2,0

       2. c0t5d0

          /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@5,0

       3. c0t6d0

          /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@6,0

       4. c0t7d0

          /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1/sd@7,0

Specify disk (enter its number): ^C

bash-3.00#

 

Step 5: Update the Firmware using the raidctl command.

 

        Syntax:

         # raidctl -F  [full_path_of_image] [controller]

        Example:

 bash-3.00# raidctl -F LSI1068X.ROM 0

 

 bash-3.00#

 

Step 6: Reboot the system to use the new FCode

 

       At this point the New FCode is loaded into the 1068X flash device but

       the system is still running off of the old image.

 

       Reboot the server to load the new 1064 Firmware.

       Syntax:

         bash-3.00# reboot

 

 

LSI 1068 HBA Supported RAID Operations - Support for RAID0 (striping) and RAID1 (mirroring)

 

RAID1 - mirroring

RAID0 - striping

Management of RAID volumes by use of the Solaris raidctl command

Examples

1.) Creating a mirror

# raidctl -c c0t0d0 c0t1d0

2.) Creating a stripe

# raidctl -c -r 0 c0t1d0 c0t2d0 c0t3d0

After RAID volume creation

Labeling Volumes After Creation

After creating a RAID volume by using 'raidctl' or by using OBP level commands the volume must be labeled correctly before it can be used. This can be done using the 'format' command in Solaris:

As the root user run 'format' and choose the newly created volume from the list of drives.

(you may see some messages about the volume having a corrupt label or the block count being incorrect when booting or running format, this is due to the need for a correct label)

After the volume is selected chose "type" from the format menu, this will give you a list of available drive types to chose from.

Chose "0. Auto configure" from the list. This will define the volume with the correct type.

Then write the new label to the volume using the "label" command from the format> prompt. The system will ask you if you are ready to label the disk. type in "yes"

After you perform these steps the drive is ready for use. You can now partition the drive as needed and use the volume within Solaris normally.

Note also that after a RAID volume is deleted the drives that were previously part of the RAID volume should be re-labeled at that time for proper use also.

RAID Volumes Possible Status

OK - volume is optimal

DEGRADED - volume has a failed disk (mirror) or is otherwise functionally minimized

RESYNCING - volume is synchronizing (mirror)

FAILED - volume is offline

Disk Possible Status

OK - disk is operating normally

MISSING - disk is removed or otherwise offline

FAILED - disk is in a failed state and needs service

 

 

3.  The first generation hard raid Chip LSI1064

 

Applies to: Sun Fire V215 [TM] Server, Sun Fire V245, Sun Fire T1000

Chip: LSI1064 SCSI Host Adapter

 

 

# iostat -En

 

e.g.

sd1       Soft Errors: 2 Hard Errors: 0 Transport Errors: 1

Vendor: LSILOGIC Product: Logical Volume   Revision: 3000 Serial No: 

Size: 146.56GB <146561286144 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

 

 

# cfgadm -alv

 

e.g.

c1::dsk/c1t0d0                 connected    configured   unknown    LSILOGIC Logical Volume

unavailable  disk         n        /devices/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1:scsi::dsk/c1t0d0

 

# prtdiag -v

 

e.g.

pci     188   MB          scsi-pci1000,50 (scsi-2)      LSI,1064          

              okay        /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

 

# prtconf -vp

 

e.g.

                            firmware-version:  '1.08.04.00'

                            mpt-version:  '1.05'

                            assigned-addresses:  810d0810.00000000.00002000.00000000.00000100.820d0814.00000000.04c00000.00000000.00010000.820d081c.00000000.04c10000.00000000.00010000.820d0830.00000000.04e00000.00000000.00200000

                            compatible: 'pci1000,50.1000.3020.2' + 'pci1000,50.1000.3020' + 'pci1000,3020' + 'pci1000,50.2' + 'pci1000,50' + 'pciclass,010000' + 'pciclass,0100'

                            model:  'LSI,1064'

 

 

# prtpicl -v

 

e.g.

:model         LSI,1064

 

 

 

SAS1064 V1.18.05.00IR on V215 /V245

136762-01

 

LSI SAS1064/1068 HBA supports up to 2 RAID volumes of same RAID-level for machines with more than two harddisks.

Just one RAID is supported on two-disk machines T1000, V215, Ultra25.

 

sas1064 on T1000

123456-01

Hardware/PROM Sun Fire T1000 1064 Firmware update V1.10.00.00IR

 

 

 

Keywords: sun fire v245 1064_firmware

Synopsis: Hardware/PROM Sun Fire V245 1064 Firmware update V1.18.05.00IR

Date: Dec/13/2007

 

 

Install Requirements: Additional instructions may be listed below                     

                     

Solaris Release: 10

 

SunOS Release: 5.10

 

Unbundled Product: Hardware/PROM

 

Unbundled Release: 1.18.05.00IR

 

Xref:

 

Topic: Sun Fire V245 1064 Firmware update V1.18.05.00IR

 

Relevant Architectures: sparc

Note:

Patch 119850-22 is required for S10U1. This patch is included in S10U2 and greater.

 

BugId's fixed with this patch: 6531216

 

Changes incorporated in this version: 6531216

 

Patches accumulated and obsoleted by this patch:

 

Patches which conflict with this patch:

 

Patches required with this patch: 119850-22 (or greater)

 

Obsoleted by:

 

Files included with this patch:

 

1.18.05-1064-Seattle.fw

 

Problem Description:

 

6531216 When replaced the mother board of the mirror system V215/V245, resyncing time were about 30hours.

 

 

Patch Installation Instructions:

===============================

 

ABOUT THIS PATCH

-------------------------

 

This patch updates the PCI-X 1064 Firmware on the Sun Fire v215/v245

to version:

 

1.18.05.00

 

 

TO CHECK THE VERSION OF THE FIRMWARE

--------------------------------------------------

 

There are two ways to check the  PCI-X Firmware version:

 

 o From the Solaris Operating System (OS)

 o From the OpenBoot PROM (OBP)

 

From the Solaris OS:

 

 1. As root, run the prtconf command.

 

    Example:

 

# prtconf -vp | egrep '(model|version)' | \

> egrep -vi '(obp|sparc-version|mpt-version)'

    model:  'SUNW,375-3463'

model:  'SUNW,4.25.10'

    model:  'm5823'

    firmware-version:  '1.08.04.00'

    model:  'LSI,1064'

    version: '1.00.40'

    model:  'SUNW,525-2320'

    version: '5.10'

#

 

    In this example, the 1064 PCI-X Firmware is version 1.08.04.00, and needs

    to be updated.

 

From the OBP:

 

  1. Identify the node for the PCI-X 1064:

 

     Example:

{0} ok  show-disks

a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk  <----

b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom

c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk

q) NO SELECTION

Enter Selection, q to quit: q

{0} ok

 

     In this example on the Sun Fire V245, "a" represents

     the 1064 node and the disk underneath the node.

 

     Once you identify the node in this list, type q to return

     to the ok prompt.

 

  2. Select the node.

 

     Example:

{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

{0} ok

 

 

     In this example, only the full path to the node is specified

     (the path does not include the "disk").

 

  3. Display the node properties.

 

     Example:

{0} ok .properties

firmware-version         1.08.04.00        <----

mpt-version              1.05

assigned-addresses       810d0810 00000000 00002000 00000000 00000100

 820d0814 00000000 04c00000 00000000 00010000

 820d081c 00000000 04c10000 00000000 00010000

 820d0830 00000000 04e00000 00000000 00200000

compatible               pci1000,50.1000.3020.2

 pci1000,50.1000.3020

 pci1000,3020

 pci1000,50.2

 pci1000,50

 pciclass,010000

 pciclass,0100

model                    LSI,1064

reg                      000d0800 00000000 00000000 00000000 00000000

 010d0810 00000000 00000000 00000000 00000100

 030d0814 00000000 00000000 00000000 00010000

 030d081c 00000000 00000000 00000000 00010000

 020d0830 00000000 00000000 00000000 00200000

version                  1.00.40

wide                     00000010

device_type              scsi-2

name                     scsi

local-wwid               50 80 02 00 00 00 29 99

devsel-speed             00000001

latency-timer            000000f8

max-latency              0000000a

min-grant                00000040

interrupts               00000001

cache-line-size          00000080

class-code               00010000

subsystem-id             00003020

subsystem-vendor-id      00001000

revision-id              00000002

device-id                00000050

vendor-id                00001000

{0} ok

 

 

     In this example, the 1064 Firmware is version 1.08.04.00, and needs

     to be updated.

 

  4. UN-select the node before booting the server.

 

     Example:

        {0}ok unselect-dev

        {0}ok

 

TO UPDATE THE 1064 Firmware

------------------------------

 

  1. Boot the Solaris OS.

 

  2. Login as root.

 

  3. Copy the new 1064 Firmware image onto the system.

          The new 1064 Firmware image is included in this patch. Copy the image

into a dir on the system which you are going to update or to

a network accessible dir. Use this full path along with the

image name in step 5) below.

 

  4. Identify the controller.

          In the following example, the controller contains the same path as shown

above in the OBP example:

 

# ls -al /dev/cfg

total 30

drwxr-xr-x   5 root     root         512 Sep 25 13:03 .

drwxr-xr-x  17 root     sys         7168 Dec  4 16:33 ..

lrwxrwxrwx   1 root     root          63 Sep 21 09:51 c0 -> ../../devices/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1:scsi

lrwxrwxrwx   1 root     root          57 Sep 25 13:03 pxb_plx5.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@1:pcie0

lrwxrwxrwx   1 root     root          57 Sep 25 13:03 pxb_plx6.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@2:pcie0

lrwxrwxrwx   1 root     root          57 Sep 25 13:03 pxb_plx7.pcie0 -> ../../devices/pci@1e,600000/pci@0/pci@2/pci@0/pci@3:pcie0

drwxr-xr-x   2 root     root         512 Sep 21 09:51 usb0

drwxr-xr-x   2 root     root         512 Sep 21 09:51 usb1

drwxr-xr-x   2 root     root         512 Sep 21 09:51 usb2

#

 

The controller in the above example is 0.

 

  5. Update the firmware using the raidctl command.

       Syntax:

         # raidctl -F  [full_path_of_image] [controller]

       Example:

# raidctl -F 1.18.05-1064-Seattle.fw 0

ROM image contains MPT firmware version 1.18.05.00

Update flash image on controller 0 (yes/no)? yes

Flash updated successfully.

 

#

 

     After the flash update is completed, the Solaris driver resets the 1064 to

     put it back into an operational state. The following error(s) are expected

     as a result of this reset:

 

       Example:

Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.warning] WARNING: /pci@7c0/pci@0/

pci@8/scsi@2/sd@0,0 (sd0):

Jun 14 08:03:16 wgs57-68        Error for Command: write                   Error

         Level: Retryable

Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice]  Requested Block: 11375

                           Error Block: 11375

Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice]  Vendor: SEAGATE

                           Serial Number: 0517104C7L

Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice]  Sense Key: Unit_Attention

Jun 14 08:03:16 wgs57-68 scsi: [ID 107833 kern.notice]  ASC: 0x29 (scsi bus reset

 occurred), ASCQ: 0x2, FRU: 0x2

 

  6. At this point the New Firmware image is loaded into the 1064's flash device but

     the system is still running off of the old image.

 

     Reboot the server to load the new 1064 Firmware.

       Syntax:

         # reboot

 

 

*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*-*

 

Special Install Instructions:

-----------------------------

 

Note:

Patch 119850-14 (or Greater) is required for S10U1. Patch 119850-14 is included in S10U2.

 

 

 

Hardware - LSI SAS1064 4 Phy-wide SAS HBA built into the motherboard.

The kernel module is "mpt".

Each LSI1064 HBA supports up to 2 RAID volumes at any one time.

Supported RAID Operations - Support for RAID0 (striping) and RAID1 (mirroring)

RAID1 - mirroring

RAID0 - striping

Management of RAID volumes by use of the Solaris raidctl command

Examples

1.) Creating a mirror

# raidctl -c c0t0d0 c0t1d0

2.) Creating a stripe

# raidctl -c -r 0 c0t1d0 c0t2d0 c0t3d0

After RAID volume creation

Labeling Volumes After Creation

After creating a RAID volume by using 'raidctl' or by using OBP level commands the volume must be labeled correctly before it can be used. This can be done using the 'format' command in Solaris:

As the root user run 'format' and choose the newly created volume from the list of drives.

(you may see some messages about the volume having a corrupt label or the block count being incorrect when booting or running format, this is due to the need for a correct label)

After the volume is selected chose "type" from the format menu, this will give you a list of available drive types to chose from.

Chose "0. Auto configure" from the list. This will define the volume with the correct type.

Then write the new label to the volume using the "label" command from the format> prompt. The system will ask you if you are ready to label the disk. type in "yes"

After you perform these steps the drive is ready for use. You can now partition the drive as needed and use the volume within Solaris normally.

Note also that after a RAID volume is deleted the drives that were previously part of the RAID volume should be re-labeled at that time for proper use also.

RAID Volumes Possible Status

OK - volume is optimal

DEGRADED - volume has a failed disk (mirror) or is otherwise functionally minimized

RESYNCING - volume is synchronizing (mirror)

FAILED - volume is offline

Disk Possible Status

OK - disk is operating normally

MISSING - disk is removed or otherwise offline

FAILED - disk is in a failed state and needs service

OBP Level Commands

PLEASE NOTE: OBP Level RAID commands are NOT SUPPORTED and are not intended for normal customer use. Future firmware updates may change command behavior and/or remove access to these commands altogether.

Before running the RAID OBP commands the auto-boot? variable needs to be set to false and the fcode-debug? variable needs to be set to true, then a reset-all is performed. After the reset finishes the controller must be selected. In the example below the "show-disks" command is used to determine the controller path. Note that the "/disk" was left off the path since it is the controller you want to select. (You can also use the probe-scsi-all command to find the path.)

{1} ok setenv auto-boot? false

auto-boot? =            false

{1} ok setenv fcode-debug? true

fcode-debug? =          true

{1} ok reset-all

:

:

:

{1} ok show-disks

a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk

b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom

c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk

q) NO SELECTION

Enter Selection, q to quit: a

/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk has been selected.

Type ^Y ( Control-Y ) to insert it in the command line.

e.g. ok nvalias mydev ^Y

for creating devalias mydev for    /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk

{1} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

The "show-volumes" command will show if there are any existing volumes and their state.

{1} ok show-volumes

Volume 0 Target 0  Type IM (Integrated Mirroring)

Optimal  Enabled

2 Members                                         143112591 Blocks, 73 GB

Disk 1

Primary  Online

Target 4        FUJITSU MAY2073RCSUN72G 0401

Disk 0

Secondary  Online

Target 1        FUJITSU MAY2073RCSUN72G 0401

{1} ok

If you wanted to create a volume you would select the controller as above and use the "create" command. The following example creates a mirrored volume with disks 0 and 1.

{0} ok 0 1 create-im-volume

  Target 0 size is 143243264 Blocks, 73 GB

  Target 1 size is 143243264 Blocks, 73 GB

  The volume can be any size from 1 MB to 69943 MB

  What size do you want?  [69943]

  Volume size will be 143243264 Blocks, 73 GB

  PhysDisk 0 has been created for target 0

  PhysDisk 1 has been created for target 1

  Volume has been created

{0} ok

This command creates a striped volume using disks 0 and 1

{0} ok 0 1 create-is-volume

  Target 0 size is 143243264 Blocks, 73 GB

  Target 1 size is 143243264 Blocks, 73 GB

  The volume can be any size from 1 MB to 139886 MB

  What size do you want?  [139886]

  Volume size will be 286486528 Blocks, 146 GB

  PhysDisk 0 has been created for target 0

  PhysDisk 1 has been created for target 1

  Volume has been created

{0} ok

To unselect

{0} ok unselect-dev

when you have completed managing the RAID volumes you can set the auto-boot? and fcode-debug? variables back and reset the system:

{1} ok setenv auto-boot? true

auto-boot? =            true

{1} ok setenv fcode-debug? false

fcode-debug? =          false

{1} ok reset-all

Remember that even if you create a RAID volume at the OBP level you will still need to label the volume using format in Solaris before they can be properly used.

If you have existing RAID volumes and have to replace the system board the

following commands will need to be run to see the volume with the new board

{1} ok show-disks

a) /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk

b) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/cdrom

c) /pci@1e,600000/pci@0/pci@1/pci@0/ide@1f/disk

q) NO SELECTION

Enter Selection, q to quit: q

/pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk has been selected.

Type ^Y ( Control-Y ) to insert it in the command line.

e.g. ok nvalias mydev ^Y

for creating devalias mydev for  /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1/disk

{1} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

{0} ok show-volumes ( Look for the inactive volume)

{0} ok X activate-volume (X is the volume number)

{0} ok unselect-dev

 

If activating multiple volumes ensure they are activated in descending order:

 

{0} ok show-volumes

{0} ok 1 activate-volume

{0} ok 0 activate-volume

{0} ok unselect-dev

{0} ok probe-scsi-all

 

Failure to do so will result in the following error and will require a reset-all prior to activating the second volume:

 

{0} ok 0 activate-volume

Volume 0 is now activated

{0} ok 1 activate-volume

RAID volume request failed, ioc-status = 0007

Volume 1 could not be activate

 

The first volume activated will remain 'Degraded' until both volumes have completed syncing.

 

NOTE: IF the EEPROM variable "auto-boot?" was set to true you may get the following message.

{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

Base SAS World Wide ID is 0!

This must be fixed immediately using set-sas-wwid

{0} ok

If you do see this message set "auto-boot?" to false and issue a "reset-all"

 

 

4.  The first generation hard raid Chip LSI1064E

 

Applies to: Sun Fire T2000, Sun Fire Netra T2000

Chip: LSI1064E SCSI Host Adapter

 

 

# iostat -En

 

e.g.

sd0       Soft Errors: 0 Hard Errors: 0 Transport Errors: 0

Vendor: LSILOGIC Product: Logical Volume   Revision: 3000 Serial No: 

Size: 73.34GB <73340551168 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

 

 

# cfgadm -alv

 

e.g.

Ap_Id                          Receptacle   Occupant     Condition  Information

When         Type         Busy     Phys_Id

c0                             connected    configured   unknown

unavailable  scsi-bus     n        /devices/pci@780/pci@0/pci@9/scsi@0:scsi

c0::dsk/c0t0d0                 connected    configured   unknown    LSILOGIC Logical Volume

unavailable  disk         n        /devices/pci@780/pci@0/pci@9/scsi@0:scsi::dsk/c0t0d0

 

# prtdiag -v

 

e.g.

========================= IO Configuration =========================

 

            IO

Location    Type  Slot Path                                          Name                      Model

----------- ----- ---- --------------------------------------------- ------------------------- ---------

IOBD/NET0    PCIE IOBD                /pci@780/pci@0/pci@1/network@0    network-pciex8086,105e

IOBD/NET1    PCIE IOBD              /pci@780/pci@0/pci@1/network@0,1    network-pciex8086,105e

IOBD/PCIE    PCIE IOBD                   /pci@780/pci@0/pci@9/scsi@0         scsi-pciex1000,56 LSI,1064E

IOBD/NET2    PCIE IOBD                /pci@7c0/pci@0/pci@2/network@0    network-pciex8086,105e

IOBD/NET3    PCIE IOBD              /pci@7c0/pci@0/pci@2/network@0,1    network-pciex8086,105e

IOBD/PCIE1   PCIE    1                   /pci@7c0/pci@0/pci@8/emlx@0         emlx-pci10df,fe00LPe11000-E

IOBD/PCIE2   PCIE    2                   /pci@7c0/pci@0/pci@9/emlx@0         emlx-pci10df,fe00LPe11000-E

IOBD/PCIX    PCIX IOBD              /pci@7c0/pci@0/pci@1/pci@0/isa@2                       isa

IOBD/PCIX    PCIX IOBD              /pci@7c0/pci@0/pci@1/pci@0/usb@5       usb-pciclass,0c0310

IOBD/PCIX    PCIX IOBD              /pci@7c0/pci@0/pci@1/pci@0/usb@6       usb-pciclass,0c0310

IOBD/PCIX    PCIX IOBD              /pci@7c0/pci@0/pci@1/pci@0/ide@8          ide-pci10b9,5229

 

# prtconf -vp

 

e.g.

                    firmware-version:  '1.09.00.00'

                    mpt-version:  '1.05'

                    assigned-addresses:  81070010.00000000.00001000.00000000.00000100.82070014.00000000.00300000.00000000.00010000.8207001c.00000000.00310000.00000000.00010000.82070030.00000000.00400000.00000000.00400000

                    compatible: 'pciex1000,56.1000.1000.2' + 'pciex1000,56.1000.1000' + 'pciex1000,56.2' + 'pciex1000,56' + 'pciexclass,010000' + 'pciexclass,0100' + 'pci1000,56'

                    model:  'LSI,1064E'

 

 

# prtpicl -v

 

e.g.

                      :model         LSI,1064E

                      :compatible        (1b6000003e6TBL)

                       | pciex1000,56.1000.1000.2 |

                       | pciex1000,56.1000.1000 |

                       | pciex1000,56.2 |

                       | pciex1000,56 |

                       | pciexclass,010000 |

                       | pciexclass,0100 |

                       | pci1000,56 |

                      :assigned-addresses        

 

 

 

sas1064 on T2000 with PCI-X Plugin cards FCode_1_00_40.ROM

122165-02

Hardware/1064_FCode Sun Fire T2000 1064 FCode update

 

 

Keywords: sun fire t2000 1064_fcode 370-7696-04 lsi3041xl

Synopsis: Hardware/1064_FCode Sun Fire T2000 1064 FCode update

Date: Mar/15/2006

 

 

Install Requirements: Additional instructions may be listed below

 

Solaris Release: 10

 

SunOS Release: 5.10

 

Unbundled Product: Hardware/Fcode LSI1064 PCI-X

 

Unbundled Release: N/A

 

Xref:

 

Topic: 1064_FCode 370-7696-04 LSI3041XL update for the Sun Fire T2000

 

Relevant Architectures: sparc

 

BugId's fixed with this patch: 6378605 6394464

 

Changes incorporated in this version: 6394464

 

Patches accumulated and obsoleted by this patch:

 

Patches which conflict with this patch:

 

Patches required with this patch:

 

Obsoleted by:

 

Files included with this patch:

 

370-7696-04_FCode_1_00_40.ROM

 

Problem Description:

 

6394464 Volume type is being referenced instead of volume id in 1.00.39 1064 fcode on PCI-X 1064 plugin card
 
(from )
 
6378605 activate_volume FCode command is needed for PCI-X 1064 plugin card

Patch Installation Instructions:

===============================

 

ABOUT THIS PATCH

-------------------------

 

This patch updates the PCI-X 1064 FCode:

 

 To version   1.00.40

 

Before using the RAID feature on your boot drive, ensure that the PCI-X

1064 plug-in card's Fcode has been upgraded to version 1.00.39 or later.

 

 

TO CHECK THE VERSION OF THE FCODE

--------------------------------------------------

 

There are two ways to check the  PCI-X FCODE version:

 

 o From the Solaris Operating System (OS)

 o From the OpenBoot PROM (OBP)

 

From the Solaris OS:

 

 1. As root, run the prtconf command.

 

    Example:

 

                    # prtconf -vp | egrep '(model|version)' | \

                      egrep -vi '(obp|firmwareversion| mpt-version)'

                           model: 'LSI,1064'

                      version: '1.00.39'

                    #

    In this example, the PCI-X FCODE is version 1.00.39, and needs

    to be updated.

 

From the OBP:

 

  1. Identify the node for the PCI-X 1064 card:

 

     Example:

           {0} ok show-disks

                  a) /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2/disk

                  b) /pci@7c0/pci@0/pci@1/pci@0/ide@8/cdrom

                  c) /pci@7c0/pci@0/pci@1/pci@0/ide@8/disk

                  q) NO SELECTION

               Enter Selection, q to quit: q

           {0} ok

 

     In this example on the Sun-Fire-T2000, "a" represents

     the plugin card. The plugin card has LSILogic in the name

     because it is treated as a third party add in card.

     Once you identify the node in this list, type q to return

     to the ok prompt.

 

  2. Select the node.

 

     Example:

           {0} ok select /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2

           {0} ok

 

     In this example, only the full path to the node is specified

     (the path does not include the "disk").

 

  3. Display the node properties.

 

     Example:

         {0} ok .properties

         firmware-version       1.06.00.00

         mpt-version            1.05

         assigned-addresses     81061010 00000000 00001000 00000000 00000100

                                82061014 00000000 03000000 00000000 00010000

                                8206101c 00000000 03010000 00000000 00010000

                                82061030 00000000 03200000 00000000 00200000

         compatible             pci13e9,50

                                pci1000,50

                                AAPL,MPT

         model                  LSI,1064

         reg                    00061000 00000000 00000000 00000000 00000000

                                01061010 00000000 00000000 00000000 00000100

                                03061014 00000000 00000000 00000000 00010000

                                0306101c 00000000 00000000 00000000 00010000

                                02061030 00000000 00000000 00000000 00200000

    ---> version                1.00.39

         wide             00000010

         device_type         scsi-2

         name             LSILogic,sas

         fcode-rom-offset     00009a00

         66mhz-capable

         devsel-speed         00000001

         latency-timer     00000040

         cache-line-size     00000010

         max-latency         0000000a

         min-grant         00000040

         interrupts         00000001

         class-code         00010000

         subsystem-id         000030f0

         subsystem-vendor-id     00001000

         revision-id         00000002

         device-id         00000050

         vendor-id         00001000

 

      In this example, the PCI-X FCODE is version 1.00.39, and needs

      to be updated.

 

  4. UN-select the node before booting the server.

 

          Example:

 

             {0}ok unselect-dev

             {0}ok

 

TO UPDATE THE FCODE

------------------------------

 

  1. Boot the Solaris OS.

 

  2. Login as root.

 

  3. Copy the new FCODE image onto the system.

  The new FCode image is included in this patch. Copy the image

  into a dir on the system which you are going to update or to

  a network accessible dir. Use this full path along with the

  image name in step 5) below.

 

  4. Identify the controller.

          In the following example, the controller is identified with

          the "LSI" string:

             # ls -al /dev/cfg

               total 22

         drwxr-xr-x 4 root root 512 Jan 19 17:28 .

         drwxr-xr-x 16 root sys 4608 Jan 20 10:47 ..

         lrwxrwxrwx 1 root root 50 Jan 19 17:28 c0 ->

            ../../ devices/pci@7c0/pci@0/pci@1/pci@0/ide@8:scsi

   --->  lrwxrwxrwx 1 root root 61 Jan 19 17:28 c1 ->

            ../../ devices/pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2:scsi

         lrwxrwxrwx 1 root root 45 Jan 19 17:28 c2 ->

            ../../ devices/pci@780/pci@0/pci@9/scsi@0:scsi

 

  5. Update the FCODE using the raidctl command.

       Syntax:

         # raidctl -F  [full_path_of_image] [controller]

       Example:

         # raidctl -F 370-7696-04_FCode_1_00_40.ROM 1

         Image file contains:

         1.00.40

         Update flash image on controller 1 (yes/no)? y

         Flash updated successfully.

         #

 

  6. Reboot the server to load the new FCode.

 

 

Brief summary of the RAID operations and command usage:

 

Create a mirror:

# raidctl -c [-r 1] primary secondary

 

Create a stripe:

# raidctl -c -r 0 disk1 disk2 [disk3] . . .

 

Delete a RAID volume:

# raidctl -d volume

 

Display RAID volumes:

# raidctl

 

Display the disk controller firmware/fcode version:

 

# prtconf -vp | egrep '(model|version)' | \

egrep -vi '(obp|firmwareversion| mpt-version)'

 

Update the disk controller firmware:

 

# raidctl -F [path_of_image_file] [controller#]

 

 

5.  The first generation hard raid Chip LSI1068E

 

Applies to: Sun Fire T5120/T5220/T5140/T5240/T5440/NetraT5220/NetraT5440

 

Chip: LSI1068E SCSI Host Adapter

 

 

# iostat -En

e.g.

sd0       Soft Errors: 2 Hard Errors: 0 Transport Errors: 0

Vendor: LSILOGIC Product: Logical Volume   Revision: 3000 Serial No: 

Size: 146.56GB <146561286144 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

 

# cfgadm -alv

e.g.

c1                             connected    configured   unknown

unavailable  scsi-bus     n        /devices/pci@400/pci@0/pci@8/scsi@0:scsi

c1::dsk/c1t0d0                 connected    configured   unknown    LSILOGIC Logical Volume

unavailable  disk         n        /devices/pci@400/pci@0/pci@8/scsi@0:scsi::dsk/c1t0d0

 

# prtdiag -v

 

e.g.

================================ IO Devices ================================

Slot +            Bus   Name +                            Model  

Status            Type  Path                                     

----------------------------------------------------------------------------

MB/SASHBA         PCIE  scsi-pciex1000,58                 LSI,1068E

                        /pci@400/pci@0/pci@8/scsi@0                 

 

# prtconf -vp

 

e.g.

                    firmware-version:  '1.27.00.00'

                    mpt-version:  '1.05'

                    assigned-addresses:  81060010.00000000.00000000.00000000.00000100.82060014.00000000.00300000.00000000.00010000.8206001c.00000000.00310000.00000000.00010000.82060030.00000000.00400000.00000000.00400000

                    compatible: 'pciex1000,58.1000.1000.4' + 'pciex1000,58.1000.1000' + 'pciex1000,58.4' + 'pciex1000,58' + 'pciexclass,010000' + 'pciexclass,0100' + 'pci1000,56'

                    model:  'LSI,1068E'

                    reg:  00060000.00000000.00000000.00000000.00000000.01060010.00000000.00000000.00000000.00000100.03060014.00000000.00000000.00000000.00010000.0306001c.00000000.00000000.00000000.00010000.02060030.00000000.00000000.00000000.00400000

                    version: '1.00.41'

 

# prtpicl -v

 

e.g.

                      :model         LSI,1068E

                      :compatible        (b800000b18TBL)

                       | pciex1000,58.1000.1000.4 |

                       | pciex1000,58.1000.1000 |

                       | pciex1000,58.4 |

                       | pciex1000,58 |

                       | pciexclass,010000 |

                       | pciexclass,0100 |

                       | pci1000,56 |

                      :assigned-addresses

 

 

 

Patches affecting all T1000 T2000 T5120/T5220 T5140/T5240 V445 V215 V245

 

 

Patch ID

Description

Min level

Notes & Dependencies

Sun OS 5.10 raidctl

-05

This resolves a number of the issues affecting behaviour changed in s10 8/07 (update 4)from prior S10

 

 

T5120/T5220 Patches:

 

 

Patch ID

Description

Min level

Notes & Dependencies

Sun OS 5.10 librcm.so.1 patch

-01

For Solaris 10 1/06 OS only,not required for the Solaris 10 6/06 or later OS.

Sun OS 5.10: mpt radioictl libpiclsbl libpiclfrudr patch

-14

118822-20

Hardware/PROM Sun Fire T1000 1064 Firmware update V1.10.00.00IR

-01

119850-14

 

 

Required Solaris patches:

Solaris 8:

 

SunOS 5.8: mpt driver patch

Solaris 9:

 

SunOS 5.9: mpt Patch

 

Solaris 10:

(or later) SunOS 5.10: MPT driver patch

 

Please note the Required Patches for 143128-xx listed in the Patch Description

In general mpt driver is part of kernel patches of Sol9 and Sol10; kernel patching is required.

 

 

Hardware Raid HBA firmware

 

SPARC Enterprise T5120 / T5220

Huron System Board 501-7781 / 501-7812 / 501-7813 / 501-7814

 

Netra T5220

Turgo System Board

540-7755 / 501-7813 + 541-2814 / 501-7781 + 541-2867 / 501-7814

 

 

SAS1068E-B1 1.27.00.00 (not ECO released)

 

Huron+ System Board 511-1087 / 511-1198 / 511-1200

 

SAS1068E-B2 1.27.00.00 (not ECO released) 2

 

SPARC Enterprise T5140 / T5240

Maramba System Board 501-7846 / 501-7847 / 501-7848 / 511-1035

 

SAS1068E-B2 Version 1.27.00.00 2

 

SPARC Enterprise T5440

Batoka System Board 541-2749 / 501-7783

 

SAS1068E-B2 1.27.00.00 (not ECO released) 1

 

 

Refer the patch readme

 

 

There is a new version of the "raidctl" utility introduced with Solaris 10 8/07.The usage of the utility has changed dramatically.Two potential areas of concern are:

1.) Previous version of "raidctl" required only entering "raidctl" to get volume and disk status.It now requires "raidctl -l volumename".Concern here that a disk or volume in a non optimal or failed state may go unnoticed.

2.) There are option and features listed that are not supported in many platforms(example:hot spares) that may cause customers to think they are available.

 

The new version of the 'raidctl' included original function

 

How to obtain RAID status:

 

raidctl -l 1 -- Controller 1

raidctl -l c1t2d0 -- Volume c1t2d0

raidctl -l c1t0d0

raidctl -l -g 0.0.0 1 -- Disk 0.0.0

raidctl -l -g 0.1.0 1

raidctl -l -g 0.2.0 1

raidctl -l -g 0.3.0 1

 

Facts:

 

How to activate a RAID volume on T1000/T2000, V215/V245/V445, T5120/T5220, T5140/T5240/T5440 Servers

 

This document provides a reference to the procedure for re-activating an existing RAID volume, on Sun Fire T1000/T2000, V215/V245/V445, SRARC Enterprise T5120/T5220, T5140/T5240 and T5440 Servers. If you have an existing RAID volume and have to replace the raid controller on one of these servers, you'll need to run some OBP commands in order to see the volume with the new board. The outlined procedure includes generic instructions, as well as platform specific instructions.

 

When the RAID controller for internal disks has been replaced, customers may need to re-activate the RAID volume by using OBP commands.

For T2000 systems, if using CPU/mem board part number 501-6843, the controller is a plug-in PCI-X card, part number 370-7696. For all other instances on the T2000, T1000, V215/V245/V445, T5120/T5220, T5140/T5240 and T5440, the controller is on the system board (motherboard).

Generic instructions for T1000, T2000, V215/V245/V445, T5x20 and T5x40:

{0} ok setenv fcode-debug? true

{0} ok setenv auto-boot? false

{0} ok reset-all

 

Path select instructions for T1000/Erie (dual-disk configuration is required for RAID operations):

{0} ok select /pci@7c0/pci@0/pci@8/scsi@2

 

Path select instructions for T2000/Ontario (pn 501-6843):

{0} ok select /pci@7c0/pci@0/pci@1/pci@0,2/LSILogic,sas@2

 

Path select instructions for T2000+/Ontario+ (pn 501-7501):

{0} ok select /pci@780/pci@0/pci@9/scsi@0

 

Path select instructions for V215/V245/Seattle:

{0} ok select /pci@1e,600000/pci@0/pci@a/pci@0/pci@8/scsi@1

 

Path select instructions for V445/Boston:

  {0} ok select /pci@1f,700000/pci@0/pci@2/pci@0/pci@8/LSILogic,sas@1

 

Path select instructions for V445 with 501-7244 SAS controller

  {0} ok select /pci@1e,600000/pci@0/pci@2/scsi@0

 

Path select instructions for  T5120/T5220/Huron:

{0} ok select /pci@0/pci@0/pci@2/scsi@0

 

Path select instructions for  T5140/T5240/Maramba:

 

{0} ok select /pci@400/pci@0/pci@8/scsi@0/

 

Path select instructions for  T5440/Batoka:

 

{0} ok select /pci@400/pci@0/pci@1/scsi@0/

 

 

Generic instructions for T1000, T2000, V215, V245, V445, T5x20, T5x40:

{0} ok show-volumes ( Look for the inactive volume)

{0} ok X activate-volume (X is the volume number)

{0} ok unselect-dev

{0} ok probe-scsi-all (to check for disks)

 

When dealing with multiple volumes make sure they are activated in descending order:

 

Generic instructions for T1000, T2000, V215, V245, V445, T5x20, T5x40:

{0} ok show-volumes

{0} ok 1 activate-volume

{0} ok 0 activate-volume

{0} ok unselect-dev

{0} ok probe-scsi-all (to check for disks)

 

Failure to do so will result in the following error and will require a reset-all prior to

activating the second volume:

 

{0} ok 0 activate-volume

Volume 0 is now activated

{0} ok 1 activate-volume

RAID volume request failed, ioc-status = 0007

Volume 1 could not be activated

 

Following activation the first volume will remain 'Degraded' until both volumes have completed syncing.

 

6. The second generation hard raid Chip LSI2008

 

Applies to: SPARC T3-1/T3-2/T3-4; T4-1/T4-2/T4-4

 

Chip: LSI 2008 mpt sas HBA

 

 

# iostat -En

 

e.g.

sd4       Soft Errors: 0 Hard Errors: 0 Transport Errors: 0

Vendor: LSI      Product: Logical Volume   Revision: 3000 Device Id: id1,sd@n600

508e0000000004e09f9a472cafd04

Size: 299.00GB <298999349248 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

 

sd7       Soft Errors: 0 Hard Errors: 0 Transport Errors: 0

Vendor: LSI      Product: Logical Volume   Revision: 3000 Device Id: id1,sd@n600508e000000000ae093bd052e7bc08

Size: 599.00GB <598999040000 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 2 Predictive Failure Analysis: 0

sd8       Soft Errors: 0 Hard Errors: 0 Transport Errors: 0

 

 

# cfgadm -alv

 

e.g.

c7                             connected    configured   unknown

unavailable  scsi-sas     n        /devices/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/

iport@v0:scsi

c7::dsk/c7t34FDCA72A4F9094Ed0  connected    configured   unknown    LSI Logical

Volume

unavailable  disk         n        /devices/pci@400/pci@1/pci@0/pci@0/LSI,sas@0/

iport@v0:scsi::dsk/c7t34FDCA72A4F9094Ed0

 

# prtdiag -v

 

e.g.

================================ IO Devices ================================

Slot +            Bus   Name +                            Model

Status            Type  Path

----------------------------------------------------------------------------

/SYS/MB/REM0/SASHBA0PCIE  LSI,sas-pciex1000,72              LSI,2008

                        /pci@400/pci@1/pci@0/pci@0/LSI,sas@0

 

# prtconf -vp

 

e.g.

                        firmware-version:  '11.05.02.00'

                        mpt-version:  '2.00'

                        local-wwid:  500605b0.02925120

                        vf-assigned-addresses:  83030000.00000041.00080000.00000

000.00004000.83030002.00000041.000c0000.00000000.00040000

                        assigned-addresses:  81030010.00000000.00000000.00000000

.00000100.83030014.00000041.00000000.00000000.00004000.8303001c.00000041.0004000

0.00000000.00040000.82030030.00000000.00100000.00000000.00080000

                        vf-reg:  03030000.00000000.00000000.00000000.00004000.03

030002.00000000.00000000.00000000.00040000

                        compatible: 'pciex1000,72' + 'pci1000,72'

                        model:  'LSI,2008'

 

 

# prtpicl -v

 

e.g.

                          :device_type   scsi-sas

                          :wide  0x10

                          :version       1.00.60

                          :reg

 00  03  00  00  00  00  00  00  00  00  00  00  00  00  00  00  00  00  00  00

 01  03  00  10  00  00  00  00  00  00  00  00  00  00  00  00  00  00  01  00

 03  03  00  14  00  00  00  00  00  00  00  00  00  00  00  00  00  00  40  00

 03  03  00  1c  00  00  00  00  00  00  00  00  00  00  00  00  00  04  00  00

 02  03  00  30  00  00  00  00  00  00  00  00  00  00  00  00  00  08  00  00

                          :model         LSI,2008

 

 

SPARC T3-1 LSI 2008 Firmware to support TSST DVD drive

This version is for SPARC T3-1 LSI 2008 Firmware to support TSST DVD drive. Refer to the README file for instructions on updating the firmware using the files included in this patch.

SPARC T3-1 LSI SAS2 Expander firmware to force zoning on the 16 disk backplane

This version is for SPARC T3-1 LSI SAS2 Expander firmware to force zoning on the 16 disk backplane. Refer to the README file for instructions on updating the firmware using the files included in this patch.

 

 

 

RAID technology allows for the construction of a logical volume, made up of several physical disks, to provide data redundancy, increased performance, or both. The system's onboard disk controller supports the following RAID configurations:

The onboard disk controller(s) of the servers provide for the creation of  two hardware RAID volumes per controller. The T3-2 and T4-2 have one onboard controller so there is a max of two raid volumes. The other systems which have two onboard controllers or REM's would support four.

 

NOTE: It is extremely important to be aware with all of these platforms that when a system board, or REM  is replaced,certain procedures have to be followed to have the existing RAID volumes be seen after replacement. Please reference document: 1387771.1.

 

=============================

 

Similarities from T5xx0 systems:

  1. Creating and deleting volumes will both destroy the data on all member disks
  2. Once a raid volume is created all members disks will no longer be seen by Solaris or OBP, only the raid volume will be visible
  3. Procedures must be followed to see any configured volumes after a system board or REM  is replaced

 

Differences from T5xx0 systems:

  1. Unlike T5xxx systems, OBP raid commands are supported for customer use
  2. The solaris utility raidctl is not supported, in it's place is sas2ircu but it must be downloaded from an LSI website, another is raidconfig which is part of HMP
  3. Raid level 1E is supported (not T3-4 and T4-4)

 

Beware of disk Labels:

 

The volume must be labeled with one of type SMI for boot disks which are limited to 1TB, but can be extended to 2TB.  An EFI label is required for all volumes larger than 2TB.  More information can be found ALWAYS ensure that the newly created volume has been relabeled!!!

 

Three environments for creating and managing Hardware RAID:

  1. Fcode utility -- This utility consists of a set of special commands to display targets and manage logical volumes on your server. You access these commands through the OBP environment.
  2. LSI SAS2 2008 RAID Management Utility for SPARC T4 servers -- You can use the sas2ircu commands contained in the LSI SAS2 Integrated Configuration Utility to configure and manage RAID volumes on your server. To use the sas2ircu command set, download and install the SAS2IRCU software from the following location: for T3-1, T3-2, T4-1, & T4-2; for T3-4, T4-4, T3-1B, & T4-1B.
  3. Oracle Hardware Management Pack 2.2 -- You can use the raidconfig commands contained in this software's Oracle Server CLI Tools component to create and manage RAID volumes on your server. To use these commands, download and install the latest version of the Hardware Management Pack from this location:

 

 

The following OBP level commands are supported:

 

show-children                Lists all connected physical drives and logical volumes.

show-volumes               Lists all connected logical volumes in detail.

create-raid0-volume      Creates a RAID 0 volume (minimum two targets).

create-raid1-volume      Creates a RAID 1 volume (exactly two targets).

create-raid1e-volume    Creates a RAID 1e volume (minimum three targets).

Delete-volume               Deletes a RAID volume.

activate-volume             Re-activate a RAID volume after controller has been replaced.

 

You can configure two global hot spare drives to protect data on mirrored RAID volumes. If one of the drives in a RAID 1 or RAID 1E mirrored volume fails, the onboard RAID controller will replace the failed drive automatically with a hot spare drive and then resynchronize the mirrored data. You must use the the sas2ircu LSI utility or raidconfig to add global hot spare drives.

 

Example of Console Output

 

SPARC T3-1, No Keyboard

Copyright (c) 1998, 2011, Oracle and/or its affiliates. All rights reserved.

OpenBoot 4.32.4, 2048 MB memory available, Serial #94789078.

Ethernet address 0:21:28:a6:5d:d6, Host ID: 85a65dd6.

{0} ok

{0} ok select /pci@400/pci@2/pci@0/pci@4/scsi@0

{0} ok show-volumes

No volumes to show

 

{0} ok show-children

FCode Version 1.00.54, MPT Version 2.00, Firmware Version 5.00.17.00

Target a

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abb2358 SASAddress 5000cca00abb2359 PhyNum 0

Target b

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abba630 SASAddress 5000cca00abba631 PhyNum 1

Target c

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abc5270 SASAddress 5000cca00abc5271 PhyNum 2

Target d

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00aa80228 SASAddress 5000cca00aa80229 PhyNum 3

Target e

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00a82fbfc SASAddress 5000cca00a82fbfd PhyNum 4

Target f

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abcd2b4 SASAddress 5000cca00abcd2b5 PhyNum 5

Target 10

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00aac16e4 SASAddress 5000cca00aac16e5 PhyNum 6

Target 12

Unit 0 Encl Serv device SUN SAS2 X16DBP 0305

SASAddress 50800200000272bd PhyNum 18

 

{0} ok a b create-raid0-volume

Target a size is 583983104 Blocks, 298 GB

Target b size is 583983104 Blocks, 298 GB

The volume can be any size from 1 MB to 570296 MB

What size do you want? [570296]

Volume size will be 1167966208 Blocks, 597 GB

Enter a volume name: [0 to 15 characters] RAID0

Volume has been created

 

{0} ok c d e create-raid1e-volume

Target c size is 583983104 Blocks, 298 GB

Target d size is 583983104 Blocks, 298 GB

Target e size is 583983104 Blocks, 298 GB

The volume can be any size from 1 MB to 427722 MB

What size do you want? [427722]

Volume size will be 875974656 Blocks, 448 GB

Enter a volume name: [0 to 15 characters] RAID1E

Volume has been created

 

{0} ok f 10 create-raid1-volume

Maximum number of volumes already exist

 

{0} ok show-volumes

Volume 0 Target 388 Type RAID1E (Mirroring Extended)

Name RAID1E WWID 0673b84764732723

Optimal Enabled Consistent Background Init In Progress

3 Members 875974656 Blocks, 448 GB

Disk 3

Member 0 Optimal

Target c HITACHI H103030SCSUN300G A2A8

Disk 4

Member 1 Optimal

Target d HITACHI H103030SCSUN300G A2A8

Disk 5

Member 2 Optimal

Target e HITACHI H103030SCSUN300G A2A8

Volume 1 Target 389 Type RAID0 (Striping)

Name RAID0 WWID 00c3b81bae2324f7

Optimal Enabled

2 Members 1167966208 Blocks, 597 GB

Disk 1

Member 0 Optimal

Target a HITACHI H103030SCSUN300G A2A8

Disk 2

Member 1 Optimal

Target b HITACHI H103030SCSUN300G A2A8

 

{0} ok reset-all

CONSERVER: XOFF timed out on t3-1-sin06-a.sg (/dev/term/e025)

SPARC T3-1, No Keyboard

Copyright (c) 1998, 2011, Oracle and/or its affiliates. All rights reserved.

OpenBoot 4.32.4, 2048 MB memory available, Serial #94789078.

Ethernet address 0:21:28:a6:5d:d6, Host ID: 85a65dd6.

 

{0} ok ok select scsi1

 

{0} ok show-children

FCode Version 1.00.54, MPT Version 2.00, Firmware Version 5.00.17.00

Target a

Unit 0 Removable Read Only device TEAC DV-W28SS-R 1.0C

SATA device PhyNum 6

Target b

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abce7f8 SASAddress 5000cca00abce7f9 PhyNum 8

Target c

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00aaec29c SASAddress 5000cca00aaec29d PhyNum 9

Target d

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00abce514 SASAddress 5000cca00abce515 PhyNum a

Target e

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00a8065f0 SASAddress 5000cca00a8065f1 PhyNum b

Target f

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00a7f8ea8 SASAddress 5000cca00a7f8ea9 PhyNum c

Target 10

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00aaec608 SASAddress 5000cca00aaec609 PhyNum d

Target 11

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00aaf4ab0 SASAddress 5000cca00aaf4ab1 PhyNum e

Target 12

Unit 0 Disk HITACHI H103030SCSUN300G A2A8 585937500 Blocks, 300 GB

SASDeviceName 5000cca00a81f620 SASAddress 5000cca00a81f621 PhyNum f

Target 13

Unit 0 Encl Serv device SUN SAS2 X16DBP 0305

SASAddress 50800200000272bd PhyNum 18

 

{0} ok show-volumes

No volumes to show

 

{0} ok a b create-raid1-volume

Couldn't get RAID PhysDisk information for target .a

 

{0} ok b c create-raid1-volume

Target b size is 583983104 Blocks, 298 GB

Target c size is 583983104 Blocks, 298 GB

The volume can be any size from 1 MB to 285148 MB

What size do you want? [285148]

Volume size will be 583983104 Blocks, 298 GB

Enter a volume name: [0 to 15 characters] RAID1

Volume has been created

 

{0} ok show-volumes

Volume 0 Target 389 Type RAID1 (Mirroring)

Name RAID1 WWID 0d8bbe28ad6b3ae2

Optimal Enabled Background Init In Progress

2 Members 583983104 Blocks, 298 GB

Disk 0

Primary Optimal

Target b HITACHI H103030SCSUN300G A2A8

Disk 1

Secondary Optimal

Target c HITACHI H103030SCSUN300G A2A8

{0} ok

 

The utility is sas2ircu and does not come as part of Solaris. It must be downloaded from the LSI website The raidctl utility is not supported for the T3/T4 platforms.  It is recommended to rename this utility to sas2ircu(SAS2 Integrated RAID Configuration Utility)

  (removing the version indication) & place this utility in /usr/sbin, as well as giving it root execute permissions.  Command syntax & sample commands are:

 

Note:

SAS2IRCU is a command-line utility that you use to configure the Integrated RAID

functions of LSI SAS2 controllers. SAS2IRCU runs in the DOS, Linux, EFI, and Windows

Pre-Installation (WinPE) environments. It is a minimally interactive program that you

can execute from a command-line prompt or a shell script. When you use SAS2IRCU

commands, the program sends results to the environment through the program status

value that it returns when it exits.

OEMs can use SAS2IRCU to configure Integrated RAID devices on the LSI family of

MPT2.0-based SAS2 controllers. Typical uses for this utility include, but are not limited

to, system configuration on the manufacturing floor, post-delivery configuration by

end users, and ongoing configuration and monitoring of Integrated RAID devices.

 

 

 

sas2ircu ...

    where is: Number between 0 and 255

    where is:

      DISPLAY    - display controller, volume and physical device info

      LIST       - Lists all available LSI adapters (does not need ctlr #>

      CREATE     - create an IR volume

      DELETE     - Delete entire RAID configuration on the selected controller

      DELETEVOLUME  - Delete a specific RAID Volume on the selected controller

      HOTSPARE   - make drive a hot spare

      STATUS     - display current volume status info

      CONSTCHK   - Start Consistency Check operation on the specified IR Volume

      ACTIVATE   - Activate an Inactive IR volume

      LOCATE     - Locate a disk drive on an enclosure

      LOGIR      - Upload or Clear IR Log data

      BOOTIR     - Select an IR Boot Volume as primary boot device

      BOOTENCL   - Select an Enclosure/Bay as primary boot device

    where are: Command specific values; enter "sas2ircu " to get command specific help

 

Please see the TOI on the Troubleshooting page for sample output.  Obtain the following command output from a system for diagnosis:

 

# sas2ircu 0 display all

# sas2ircu 0 status

# sas2ircu 1 display all

# sas2ircu 1 status

 

A sample mirror creation is as follows:

 

# sas2ircu 1 create raid1 MAX 2:9 2:15

 

Example of sas2ircu output:

 

# ./sas2ircu-6.250.02.00

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

SAS2IRCU: No command specified.

sas2ircu ...

where is:

Number between 0 and 255

where is:

DISPLAY - display controller, volume and physical device info

LIST - Lists all available LSI adapters (does not need ctlr #>

CREATE - create an IR volume

DELETE - Delete entire RAID configuration on the selected controller

DELETEVOLUME - Delete a specific RAID Volume on the selected controller

HOTSPARE - make drive a hot spare

STATUS - display current volume status info

CONSTCHK - Start Consistency Check operation on the specified IR Volume

ACTIVATE - Activate an Inactive IR volume

LOCATE - Locate a disk drive on an enclosure

LOGIR - Upload or Clear IR Log data

BOOTIR - Select an IR Boot Volume as primary boot device

BOOTENCL - Select an Enclosure/Bay as primary boot device

where are:

Command specific values; enter "sas2ircu "

to get command specific help

 

# ./sas2ircu-6.250.02.00 LIST

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Adapter Vendor Device SubSys SubSys

Index Type ID ID Pci Address Ven ID Dev ID

----- ------------ ------ ------ ----------------- ------ ------

0 SAS2008 1000h 72h 00h:400h:00h:00h 1000h 0072h

Adapter Vendor Device SubSys SubSys

Index Type ID ID Pci Address Ven ID Dev ID

----- ------------ ------ ------ ----------------- ------ ------

1 SAS2008 1000h 72h 00h:b00h:00h:00h 1000h 0072h

SAS2IRCU: Utility Completed Successfully.

 

# ./sas2ircu-6.250.02.00 0 DISPLAY

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Read configuration has been initiated for controller 0

------------------------------------------------------------------------

Controller information

------------------------------------------------------------------------

Controller type : SAS2008

BIOS version : 0.00.00.00

Firmware version : 5.00.17.00

Channel description : 1 Serial Attached SCSI

Initiator ID : 0

Maximum physical devices : 831

Concurrent commands supported : 1871

Slot : Unknown

Segment : 0

Bus : 1024

Device : 0

Function : 0

RAID Support : Yes

------------------------------------------------------------------------

IR Volume information

------------------------------------------------------------------------

Physical device information

------------------------------------------------------------------------

Initiator at ID #0

Device is a Hard disk

Enclosure # : 2

Slot # : 0

SAS Address : 5000cca-0-0abb-2359

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1031G9XHHE

Protocol : SAS

Drive Type : SAS_HDD

Device is a Sequential access device

Enclosure # : 2

Slot # : 0

SAS Address : 5080020-0-0002-72bd

State : Standby (SBY)

Manufacturer : SUN

Model Number : SAS2 X16DBP

Firmware Revision : 0305

Serial No : x365351

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 1

SAS Address : 5000cca-0-0abb-a631

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1031GA66ME

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 2

SAS Address : 5000cca-0-0abc-5271

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1031GAKP7E

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 3

SAS Address : 5000cca-0-0aa8-0229

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1027GZDB5E

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 4

SAS Address : 5000cca-0-0a82-fbfd

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1017GA14AE

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 5

SAS Address : 5000cca-0-0abc-d2b5

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1031GAV71E

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 6

SAS Address : 5000cca-0-0aac-16e5

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1027G1MY1E

Protocol : SAS

Drive Type : SAS_HDD

Device is a Hard disk

Enclosure # : 2

Slot # : 7

SAS Address : 5000cca-0-0aab-331d

State : Ready (RDY)

Size (in MB)/(in sectors) : 286102/585937499

Manufacturer : HITACHI

Model Number : H103030SCSUN300G

Firmware Revision : A2A8

Serial No : 1027G14STE

Protocol : SAS

Drive Type : SAS_HDD

------------------------------------------------------------------------

Enclosure information

------------------------------------------------------------------------

Enclosure# : 1

Logical ID : 50800200:00d69868

Numslots : 8

StartSlot : 0

Enclosure# : 2

Logical ID : 50800200:000272bf

Numslots : 17

StartSlot : 0

------------------------------------------------------------------------

SAS2IRCU: Command DISPLAY Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

#

 

# ./sas2ircu-6.250.02.00 1 create

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

SAS2IRCU: Incorrect number of command line parameters for 'CREATE' command!

The format of the CREATE command is

sas2ircu create

[Volume Name] [noprompt]

where is:

A controller number between 0 and 255.

where is:

The type of the volume to create and is either RAID1 (or)

RAID1E (or) RAID0 (or) RAID10.

where is:

The size of the volume to create. It should be given in Mbytes

e.g. 2048 or 'MAX' to use the maximum size possible.

where is:

A list of Encl:Bay pairs identifying the disk drives you

wish to include in the volume being created. If the volume type is

'RAID1', the first drive will be selected as the primary and the

second as the secondary drive.

For a type 'RAID1' volume exactly 2 disks must be specified.

For a type 'RAID1E' volume min of 3 disks must be specified.

For a type 'RAID0' volume min of 2 disks must be specified.

For a type 'RAID10' volume min of 4 disks must be specified.

where [Volume Name] is an optional argument that can be used

to identify a Volume with a user specified Alpha-numeric string

where noprompt is an optional argument that eliminates

warnings and prompts

 

 

# ./sas2ircu-6.250.02.00 1 create RAID1 MAX 2:10 2:11

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

You are about to create an IR volume.

WARNING: Proceeding with this operation may cause data loss or data

corruption. Are you sure you want to proceed (YES/NO)? yes

WARNING: This is your last chance to abort this operation. Do you wish

to abort (YES/NO)? no

Please wait, may take up to a minute...

Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, inactive

Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active

Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active

Jul 21 13:00:02 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:00:02 t3-1-sin06-a Volume 0 is now , enabled, active, background initialization in progress

SAS2IRCU: Volume created successfully.

SAS2IRCU: Command CREATE Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

# Jul 21 13:00:03 t3-1-sin06-a scsi: WARNING:

/pci@400/pci@2/pci@0/pci@4/scsi@0/iport@v0/disk@w3d5c9144a29779de,0 (sd14):

Jul 21 13:00:03 t3-1-sin06-a Corrupt label; wrong magic number

Jul 21 13:00:03 t3-1-sin06-a scsi: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:00:03 t3-1-sin06-a send_sep act 0: ioc status:8001

Jul 21 13:00:03 t3-1-sin06-a scsi: WARNING:

/pci@400/pci@2/pci@0/pci@4/scsi@0/iport@v0/disk@w3d5c9144a29779de,0 (sd14):

Jul 21 13:00:03 t3-1-sin06-a Corrupt label; wrong magic number

#

 

# ./sas2ircu-6.250.02.00 0 status

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Background command progress status for controller 0...

IR Volume 1

Volume ID : 904

Current operation : None

Volume status : Enabled

Volume state : Optimal

Physical disk I/Os : Not quiesced

IR Volume 2

Volume ID : 905

Current operation : Background Init

Volume status : Enabled

Volume state : Optimal

Physical disk I/Os

: Not quiesced

Volume size (in sectors) : 583983104

Number of remaining sectors : 563631328

Percentage complete : 3.48%

SAS2IRCU: Command STATUS Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

#

19.) How to delete a RAID volume with sas2ircu

# ./sas2ircu-6.250.02.00 1 status

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Background command progress status for controller 1...

IR Volume 1

Volume ID : 904

Current operation : Background Init

Volume status : Enabled

Volume state : Optimal

Physical disk I/Os : Not quiesced

Volume size (in sectors) : 583983104

Number of remaining sectors : 436668576

Percentage complete : 25.23%

IR Volume 2

Volume ID : 905

Current operation : None

Volume status : Enabled

Volume state : Optimal

Physical disk I/Os : Not quiesced

SAS2IRCU: Command STATUS Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

 

# ./sas2ircu-6.250.02.00 1 deletevolume 904

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

You are about to delete an existing RAID Volume on a controller. This command will

delete the specified RAID volume and associated HotSpare drive(s).

WARNING: Proceeding with this operation may cause data loss or data

corruption. Are you sure you want to proceed (YES/NO)? yes

WARNING: This is you last chance to abort this operation. Do you wish

to abort (YES/NO)? no

Please wait, may take up to a minute...

Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /pci@400/pci@2/pci@0/pci@4/scsi@0 (mpt_sas1):

Jul 21 13:52:31 t3-1-sin06-a raid doesn't exist at specified target.

SAS2IRCU: Volume deleted successfully.

SAS2IRCU: Command DELETEVOLUME Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

# Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00abce514 (sd3):

Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number

Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00a8065f0 (sd2):

Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number

Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00abce514 (sd3):

Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number

Jul 21 13:52:31 t3-1-sin06-a scsi: WARNING: /scsi_vhci/disk@g5000cca00a8065f0 (sd2):

Jul 21 13:52:31 t3-1-sin06-a Corrupt label; wrong magic number

 

# ./sas2ircu-6.250.02.00 1 status

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Background command progress status for controller 1...

IR Volume 1

Volume ID : 905

Current operation : None

Volume status : Enabled

Volume state : Optimal

Physical disk I/Os : Not quiesced

SAS2IRCU: Command STATUS Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

#

 

If the system has RAID1 or RAID1E disk volumes, the system can afford to have a single disk failure

without data loss. To determine the status of the RAID volume and location of the of the faulty drive

in the server, you can use the following command.

sas2ircu 0 display

sas2ircu 0 display

 

Once the faulty disk slot is identified, just pullout the failed disk and insert the replacement disk into

the chassis. Depending on the size and activity on the server , a 300GB disk on an idle system will take

approximately 1 hour to rebuild.

The following command will give you the percentage of the rebuilding process

sas2ircu 0 status

sas2ircu 1 status

 

# ./sas2ircu-6.250.02.00 1 status

LSI Corporation SAS2 IR Configuration Utility.

Version 6.250.02.00 (2010.06.30)

Copyright (c) 2009-2010 LSI Corporation. All rights reserved.

Background command progress status for controller 1...

IR Volume 1

Volume ID : 905

Current operation : Synchronize

Volume status : Enabled

Volume state : Degraded

Physical disk I/Os : Not quiesced

Volume size (in sectors) : 583983104

Number of remaining sectors : 520166912

Percentage complete : 10.93%

SAS2IRCU: Command STATUS Completed Successfully.

SAS2IRCU: Utility Completed Successfully.

 

 

 

The third utility that can be used to administer hardware raid volumes is raidconfig.  Like the fwupdate command, raidconfig is part of the HMP - Oracle Hardware Management Pack which is discussed on an earlier slide. Documentation for the utility can be found:

  

  

 

Errors might be returned if you are attempting to configure the RAID entry to an unsupported parameter. The list of error codes and strings specific to this tool are show in the Sample commands are as follows:

 

# raidconfig list all -v   - Collect the most possible data for troubleshooting

# raidconfig list disk     - Listing disks

# raidconfig list raid -v  - lists all the raid volumes for either controller.

# raidconfig create raid -c c0 --level 1 -d c0d2,c0d3    - Create raid

 

Please see the TOI on the Troubleshooting page for sample output.

Boot Paths for internal disks, both used in HW Raid and not

 

 ***********************************
 
Example
 
 {0} ok probe-scsi-all
 /pci@400/pci@2/pci@0/pci@e/scsi@0
 
 FCode Version 1.00.60, MPT Version 2.00, Firmware Version 9.00.00.00
 
 Target 9
  
Unit 0   Disk   SEAGATE  ST960005SSUN600G 0606    1172123568 Blocks, 600 GB
   SASDeviceName 5000c500437b4c8b  SASAddress 5000c500437b4c89  PhyNum 0
                                                          |
                                                        ---<-used for non hw raid boot path
 Target a
  
Unit 0   Removable Read Only device   TEAC    DV-W28SS-R      1.0C
   SATA device  PhyNum 7
 Target b
  
Unit 0   Disk   SEAGATE  ST960005SSUN600G 0606    1172123568 Blocks, 600 GB
   SASDeviceName 5000c500437e934b  SASAddress 5000c500437e9349  PhyNum 1
 Target 381 Volume 0
  
Unit 0   Disk   LSI      Logical Volume   3000    1169920000 Blocks, 598 GB
   VolumeDeviceName 3191d94818ab05da  VolumeWWID 0191d94818ab05da |
              ^-used for hw raid boot path

 /pci@400/pci@1/pci@0/pci@b/pci@0/usb@0,2/hub@2/hub@3/storage@2
  
Unit 0   Removable Read Only device    AMI     Virtual CDROM   1.00
 
 {0} ok
 

Assuming you have successfully installed Solaris into a RAID LUN, when booting the device from

OBP , boot the RAID device LUN using the “VolumeDeviceName”.

For example 3191d94818ab05da is the WWN in the OBP boot command below

boot /pci@400/pci@2/pci@0/pci@4/scsi@0/disk@w3191d94818ab05da,0:a

 
 # format
 Searching for disks...done
 
 
 AVAILABLE DISK SELECTIONS:
       
0. c0t5000C500437B4C8Bd0   solaris
           /scsi_vhci/disk@g5000c500437b4c8b
        1. c0t5000C500437E934Bd0   solaris
           /scsi_vhci/disk@g5000c500437e934b
        2. c3t3191D94818AB05DAd0
           /pci@400/pci@2/pci@0/pci@e/scsi@0/iport@v0/disk@w3191d94818ab05da,0
 
 
How to boot an internal  HW Raid volume
 
 To boot the hw raidinternal volume 
use the following path:
 
 {0} ok boot
/pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w0191d94818ab05da,0:a
 
 Note that the path is derived from the controller path output from probe-scsi-all.
 Some systems have two controllers. You must use the controller path the volume is under.
 In this example it is...../pci@400/pci@2/pci@0/pci@e/scsi@0 And to that added
disk@w...and the VolumeDeviceName, followed by ......  ,0:a

{0} ok boot /pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w3191d94818ab05da,0:a

 
 
 
How to boot an internal disk that is not part  a HW Raid volume
 
 To boot the non hw raided internal disk in slot 0(indicated by PhyNum 0 from probe-scsi-all) 
use the following path:
 
 {0} ok boot
/pci@400/pci@2/pci@0/pci@e/scsi@0/disk@w5000c500437b4c89,0:a
 
 Note that the path is derived from the controller path output from probe-scsi-all.
 Some systems have two controllers. You must use the controller path the disk is under.
 In this example it is.....
/pci@400/pci@2/pci@0/pci@e/scsi@0 And to that added disk@w...and the SASAddress, followed by ......  ,0:a

 

 

7. The third generation hard raid Chip LSI 2108

Applies to: SPARC T4-1/T4-2

 

Hardware Platforms ( as of April 2012 )

SPARC T3-1, SPARC T3-2, SPARC T4-1,SPARC T4-

2,Netra SPARC T3-1, Netra SPARC T4-1 and Netra

SPARC T4-2

SPARC T4-4

running Solaris 10 x86, Oracle Linux, Oracle VM,

Red Hat Linux 4.0, 5.0 (32/64-bit), VMware ESX

server 3.x and MS Windows Standard/Enterprise

Server 2003 (32/64-bit)

 

configured with the following description in the

yellow Customer Information Sheet (SG-SAS6-RINT-

Z).

On-Board HW Raid

( RAID 0,1,5,6,10,50 and 60 vs RAID 0, 1 and 10)

cache(512MB), providing better I/O performance

with a higher level of data protection.

 

must be installed in PCI slot #3. Connect

P2(SAS0) on Disk Backplane to J7(Top) on RAID

HBA. Connect P3(SAS1) on Disk Backplaneto

J8(Bottom) on RAID HBA.

 

supported on SPARC T3-1 with 16 Disk

Backplane. There is no 16 Disk Backplane SPARC

T4-1.

 

card needs to be installed on PCI slot #0.

 

Note: The slot#0 isn't necessary for the Niwot for T3/4-2 installation

 

the use of the internal DVD drive

port J7 (Top), Connect SAS 4-5 to port J8

(Bottom) on RAID HBA card.

 

 

Chip: LSISAS2108 RAID on Chip ( ROC)

 

 

# iostat -En

 

e.g.

sd2       Soft Errors: 0 Hard Errors: 0 Transport Errors: 0

Vendor: LSI Product: MR9261-8i Revision: 2.50 Device Id: id1,sd@n600605b004ae9620ff005b6472a8fea1

Size: 299.00GB <298999349248 bytes>

Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0

Illegal Request: 3 Predictive Failure Analysis: 0

 

# cfgadm -alv

 

e.g.

c3                             connected    configured   unknown

unavailable  scsi-bus     n        /devices/pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0:scsi

c3::dsk/c3t0d0                 connected    configured   unknown    LSI MR9261-8i

unavailable  disk         n        /devices/pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0:scsi::dsk/c3t0d0

 

# prtdiag -v

 

e.g.

================================ IO Devices ================================

Slot +            Bus   Name +                            Model

Status            Type  Path

----------------------------------------------------------------------------

/SYS/MB/NET0      PCIE  network-pciex8086,10c9

 

----------------------------------------------------------------------------

/SYS/MB/NET0      PCIE  network-pciex8086,10c9

                        /pci@400/pci@1/pci@0/pci@4/network@0

/SYS/MB/NET1      PCIE  network-pciex8086,10c9

                        /pci@400/pci@1/pci@0/pci@4/network@0,1

/SYS/MB/SASHBA    PCIE  scsi-pciex1000,72                 LSI,2008

                        /pci@400/pci@2/pci@0/pci@e/scsi@0

/SYS/MB/NET2      PCIE  network-pciex8086,10c9

                        /pci@500/pci@1/pci@0/pci@5/network@0

/SYS/MB/NET3      PCIE  network-pciex8086,10c9

                        /pci@500/pci@1/pci@0/pci@5/network@0,1

/SYS/MB/PCIE1     PCIE  LSI,mrsas-pciex1000,79            LSI,2108_4

                        /pci@500/pci@2/pci@0/pci@a/LSI,mrsas@0

 

# prtconf -vp

 

e.g.

                      

                        compatible: 'pciex1000,79'

                        model:  'LSI,2108_4'

                        reg:  000a0000.00000000.00000000.00000000.00000000.010a0010.00000000.00000000.00000000.00000100.030a0

014.00000000.00000000.00000000.00004000.030a001c.00000000.00000000.00000000.00040000.020a0030.00000000.00000000.00000000.0004

0000

                        version: '4.07.04.00'

                        wide:  00000010

                        device_type:  'scsi-sas'

                        name:  'LSI,mrsas'

                        fcode-rom-offset:  00008000

                        port-type:  'PCIE-Endpoint'

                        interrupts:  00000001

                        cache-line-size:  00000010

                        class-code:  00010400

                        subsystem-id:  00009263

                        subsystem-vendor-id:  00001000

 

# prtpicl -v

 

e.g.

:model         LSI,2108_4

                          :compatible    pciex1000,79

                          :assigned-addresses

 

Package Version         Firmware Version (Mandatory Update)

12.12.0-0079             2.120.203-1440

Note: Before loading this new firmware, download and install the latest MegaCli for your OS and then use it to upgrade firmware.”

 

 

************************************

 

LSI Corp. MegaRAID Firmware Release

 

************************************

 

Release date: 05/02/12

 

======================

Supported Controllers:

======================

 

MegaRAID SAS 9262-8i

MegaRAID SAS 9281-8i

 

*******THIS FIRMWARE DOES NOT SUPPORT CACHECADE PRO 2.0 OR CACHEVAULT CONTROLLERS. 

*******PLEASE REFER TO DIFFERENT FIRMWARE IMAGES ON LSI.COM FOR CACHECADE 2.0 AND CACHEVAULT SUPPORT

 

Component:

=========

SAS MegaRAID Firmware Release for MegaRAID Controllers

Release date: 05/02/12

 

 

Version Numbers:        

===============

Current Firmware Package:    12.12.0-0079

Current Package Details:

APP-2.120.203-1440_BB-2.02.00.00-0000_BIOS-3.24.00_4.12.05.00_0x05090000_HII-_WEBBIOS-6.0-46-e_41-Rel_PCLI-04.04-01900008_BOOT-09.250.01.219_NVDATA-2.09.03-0027_2011_11_08

  

Previous Firmware Package:    12.12.0-0048

Previous Package Details:

APP-2.120.63-1242_BB-2.02.00.00-0000_BIOS-3.22.00_4.11.05.00_0x05020000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-09.250.01.219_NVDATA-2.09.03-0013_2011_04_06

     

 

Bug Fixes and Enhancements:

===========================

Firmware:

=========

LSIP200209948 (CO)   Generally available support of selected battery modes for iBBU08.

LSIP200211360 (DFCT) (CL LSIP200209847) Pulling VD 0 from an R60 and rebooting, 2 foreign configs when array 0

LSIP200199863 (DFCT) UnconfigGood PD fails to kick in as EHS (Emergency Hot Spare)

LSIP200212811 (DFCT) Advanced features key have to be re-entered

LSIP200050209 (DFCT) DATA TLB error exception in case of manual removal/insertion of drives.

LSIP200051385 (DFCT) R1 rebuild hangs at 50% mark after CC is run

LSIP200051392 (DFCT) Fw broke into MegaMon while writing a test file to a VD (R6) from a non MR disk.

LSIP200221925 (DFCT) SATA workaround handles LBA incorrectly for 6-byte read CDB

LSIP200212818 (DFCT) Port forward LSIP200211068 to main stream liberator maintenance

LSIP200212644 (DFCT) noticed additional fan speed event messages every 2 minutes

LSIP200198466 (DFCT) wrong slot # reported by megaraid compared to physical slot for EG disk

LSIP200199687 (DFCT) 揘on-Transparent?learns seen with Mode 3

LSIP200210349 (DFCT) BBUmode incorrectly set

LSIP200150196 (DFCT) MR does not pick up replacement drives in PD if drive is added during power cycle

LSIP200150197 (DFCT) NEM removal and re-insertion causes all VD members to change to Foreign Unconfigured Bad state

LSIP200210823 (DFCT) No data is copied to target array after more than 15 secs

LSIP200212264 (DFCT) Multibit ECC error with customer Liberator MR board and vmware ESXi5.0

LSIP200199735 (DFCT) LSI Mega RAID SAS-MFI BIOS Register state 0xF0010002

LSIP200211484 (DFCT) MR bios show "adapter at baseport is not repoonding" intermittence

LSIP200210596 (DFCT) too many messages in fw log "ArDiskTypeMisMatch : SAS_SATA_MIXING_VIOLATION?

LSIP200155948 (DFCT) FW seems to incorrectly mark ibbu's SOH bad (iBBU07, ibbu05, ibbu01)

LSIP200209279 (DFCT) additional reboot required after updating ibbu FCC to design capacity

LSIP200210505 (DFCT) OEM requests qty of prints to be reduced in fw log for ibbu i2c communication retries

LSIP200212471 (DFCT) FW automatically starts learn cycles on ibbu08 every reboot

LSIP200211810 (DFCT) Battery failed message during learn cycle

LSIP200212311 (DFCT) Battery failed message during learn cycle

LSIP200200117 (DFCT) unexpected Power State Change events occur

LSIP200185441 (DFCT) Disk performance disparity between WB controller cache on and off on system disks

LSIP200210751 (DFCT) Patrol Read fails on R1 with around 2000 double medium errors

LSIP200210768 (DFCT) Enabling security causes the config to be missing, and a configured FDE array to be seen as foreign on reboot.

LSIP200149248 (CO) Warning for High Error Counts

LSIP200149890 (CO) SW licensing. Support FoD for PF activations on iMR/2008

LSIP200187910 (CO) Request to use Unconfigured Good Drives for a HotSpare if no spare defined in liberator

LSIP200187724 (CO) MR fw should Enable puncturing for SATA drives

LSIP200187937 (CO) MegaRaid does not pick up replacement drives in PD if replacement is added during power cycle

LSIP200151375 (CO) Need StopOnError functionality when SafeStore passphrase is enabled

LSIP200187926 (CO) Change BBU-08 support Firmware Mode5 WB capacity from 960mAh to 674mAh

LSIP200187931 (CO) disable bad phy checking logic from iMR FW for certain ROMB design

LSIP200158631 (CO) Alert should be sent when one phy connector between Liberator card and SAS expander is degraded

LSIP200199858 (CO) add display of ibbu communication retry counter in 10M09 fw log

LSIP200199484 (CO) Enable copy back on SMART - Liberator

LSIP200186977 (CO) Alert should be sent when one phy connector between Liberator card and SAS expander is degraded

LSIP200149050 (CO) Add SGPIO for Medusa

LSIP200186343 (CO) (LSIP200140211) Add support to pass SMART errors for JBOD drives to MSM

LSIP200185599 (DFCT) LED behavior incorrect, drive 3 blinks when drive 0 is accessed

LSIP200100032 (DFCT) system occasionally hangs when executing an "init 0"

LSIP200209385 (DFCT) Intermittent RAID key failure during power-on

LSIP200209329 (DFCT) FoD features will not activate or deactivate

LSIP200150153 (DFCT) Strange HDD Led behaviour in X4170M2 under raid configuration

LSIP200156822 (DFCT) CC2.0: WT Downgrade FW from CC2.0 to CC1.0(10M09) and import caused MonTask: line 199 in file ../../raid/1078int.c

LSIP200149201 (DFCT) (CL LSIP200146291) Medium error handling fixes for >2TB

LSIP200199812 (DFCT) Alarm sounds with online VD's after flashing latest 10M09 Patch 16 firmware (2.120.163-1343)

LSIP200199961 (DFCT) (CL LSIP200061413) LDBBM entries are incorrect after foreign import

LSIP200186336 (DFCT) Controller does not work with pci set to nomsi in kdump kernel

LSIP200209455 (DFCT) (CL LSIP200188135) Potential data integrity issue running the IO on degraded R5 VDs with 32 drives.

LSIP200158264 (DFCT) Change the capacity level that generates the Predictive failure alert

LSIP200150304 (DFCT) memory test for 2108 in 10M09 FW fails immediately

LSIP200149986 (DFCT) MSM log shows "Battery cannot initiate transparent learn cycles" due to BBU is under mode 5.

LSIP200060949 (DFCT) iBBU07 reported remaining capacity is much too high (64751 mAh) after a learn cycle

LSIP200151502 (DFCT) BBU Absolute Charge is displayed differently

LSIP200149843 (DFCT) VPD (Vital Product Data) needs to show correct Part number

LSIP200149844 (DFCT) VPD (Vital Product Data) needs to identify PCB hardware level

LSIP200199416 (DFCT) Change BBU-08 support Firmware Mode5 WB capacity from 960mAh to 674mAh for low power DDR-2

LSIP200199331 (DFCT) 10M09: Older ibbu batteries display 65K capacity and are marked SOH bad

LSIP200199335 (DFCT) 10M09: Request FW to not mark ibbu battery as bad when very high capacity is detected

LSIP200196885 (DFCT) Batteries Not Charging

LSIP200209376 (DFCT) HWR: 10M09: OEM requests ibbu i2c communication retries to be displayed in fw log

LSIP200147629 (DFCT) (CL LSIP200167360) FW hit montask upon importing foreign config

LSIP200147626 (DFCT) (CL LSIP200139418) System reboots after issuing OCR

LSIP200147907 (DFCT) Disable Dimmer Switch III

LSIP200151221 (DFCT) CLI/EFI CLI/PCLI fail to create a VD on MR

LSIP200148150 (DFCT) WHQL WLK1.6 - SCSI Compliance Test 2.0 and Read Capacity 16 Unit FAILS on Liberator

LSIP200184970 (DFCT) HWR: 10M09: not able to disable BGI with MegaCLI

LSIP200173287 (DFCT) HWR: bbuMode is not in WebBIOS after flashing from 10M05 to 10M09

LSIP200186884 (DFCT) HWR: 10M09: DelaySOD is not working on customer Liberator boards

LSIP200150532 (DFCT) (RW LSIP200145892) (6997086) (RW LSIP200121969) Liberator FW 10M09 cause system (sles11sp1/rhel6) reset while lspci -vvx

LSIP200188226 (DFCT) HWR 10M09 can't sustain at least 950 MBps performance with Cache Flush setting 1 sec

LSIP200200459 (DFCT) MegaCli prematurely returns exit 0x0 status on drive firmware flash

LSIP200209349 (DFCT) Enable PCI capability Bit for VPD exsupport.

LSIP200151328 (DFCT) HWR: EF & 10M09: bluescreen occurs handling R1 with HSP and medium errors

LSIP200186966 (DFCT) HWR: 10M09: configuration becomes foreign after fw fails 3-4 attempts to access raid key during boot

LSIP200147621 (DFCT) (CL LSIP200163913) performance issues within 9285 and 9280 caused by the MegaRAID FW

LSIP200147632 (DFCT) (CL LSIP200173232) Potential data integrity issue found on R1 VDs while injecting the media errors under IO.

LSIP200186406 (DFCT) Code Review defect to deliver rest of changes for PFK preboot support.

LSIP200186594 (DFCT) R5 pfk feature not getting enabled at boot time on iMR

LSIP200186595 (DFCT) After loading FoD key, iMR controller does not perform upgrade of features

LSIP200188345 (DFCT) (CL LSIP200185468) HWR: 10M09: extended retries of write verify command causes driver timeout

LSIP200171125 (DFCT) No event is generated when fan in enclosure fails

LSIP200184943 (DFCT) Need ability to boot with pinned cache with any raid config on Gen 2 MegaRAID/Liberator defaulted to ON

LSIP200185931 (DFCT) Bootwithpinnedcache not default set to '1' for castorlite when suboem is '1'

LSIP200186584 (DFCT) Logical Unit ID (VPD page 0x83) changed after foreign import

LSIP200198521 (DFCT) 10M09 P16 set properties command fails

LSIP200146463 (DFCT) Update OOB (BMC) management

LSIP200171812 (DFCT) Liberator Firmware : Fatal F/W Error* happened when detected drive error** during Read I/O.

LSIP200169751 (DFCT) MegaRaid FW does not wait at post

 

 

Webbios:

========

LSIP200210939 (DFCT) (N/A) System showed error message when entering uFEI web-bios

LSIP200210311 (DFCT) Adding OK button instead of cancel in "Advance software Options" page

LSIP200199865 (CO) MegaRAID Missing Stop Locator for Volume

LSIP200209946 (CO) Generally available support of selected battery modes for iBBU08.

LSIP200150553 (CR) MegaRAID Missing Stop Locator for Volume

LSIP200187637 (CR) Generally available support of selected battery modes for iBBU08.

LSIP200199949 (CO) SW licensing. Support FoD for PF activations on MegaRAID

LSIP200150642 (CO) 10M09 Firmware release has Advanced Software Options entry that should be removed

LSIP200188273 (DFCT) uEFI WebBIOS incorrectly displays "UG Spin Down" title

LSIP200139977 (DFCT) Option for populating the VD size field automatically

LSIP200156291 (DFCT) can't create DG with 8 PDs with enableReducedFeatureSet

LSIP200157087 (DFCT) EFI WebBIOS does not show product name/version for second adapter in adapter selection screen

LSIP200121693 (DFCT) Unable to configure Cachecade in webBIOS under 10M09 for 2108 based products

LSIP200166733 (DFCT) WEBBIOS is unable to reconstruct the R11 VD to R5 with removal of one PD.

LSIP200147461 (DFCT) Cannot change the BBU properties in webbios

LSIP200149524 (DFCT) WebBIOS gives garbled message when enabing FDE on an existing Disk Group

 

 

BIOS:

========

LSIP200199737 (DFCT) server System posts 'R00' as 'R0' during the reboot

LSIP200199163 (DFCT) Legacy BIOS BBS Boot Issues

Numerous UEFI and HII updates

LSIP200196810  SWR-Macon: After installing the OS, unable to boot from SCU VD

LSIP200199602  FW hangs during POST after flashing new BIOS

LSIP200018807  BIOS sees extra characters "0%" at end of particular product name

LSIP200156493  There is no space between FW version and capacity in the table where Adp/VD/PD info is displayed

Numerous HII fixes

LSIP200209201 (DFCT) UEFI: HII Create Configuration not all virtual drive changed parameters are saved

LSIP200150135 (DFCT) Cannot configure VD on Niwot using UEFI HII menu, problem reported by AMI

LSIP200150702 (DFCT) Attempting to create a virtual drive using UEFI HII resulted in a "hang"

LSIP200187671 (DFCT) UEFI driver 05050000 breaks HII menu in server f2 Bios

 

NVDATA:

=======

Numerous NVDATA changes

 

 

Installation:

=============

Use MegaCLI to flash the SAS controllers.  MegaCLI can be downloaded from the support and download

section of .

 

Command syntax:  MegaCli -adpfwflash -f mr2108fw.rom -a0

 

 

 

=================================

Previous Releases

=================================

     

Component:

=========

SAS MegaRAID Firmware Release for MegaRAID Controllers

Release date: 04/15/11

 

 

Version Numbers:        

===============

Current Firmware Package:    12.12.0-0048

Current Package Details:

APP-2.120.63-1242_BB-2.02.00.00-0000_BIOS-3.22.00_4.11.05.00_0x05020000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-09.250.01.219_NVDATA-2.09.03-0013_2011_04_06

    

Previous Firmware Package:   12.12.0-0045

Previous Package Details:

APP-2.120.33-1197_BB-2.02.00.00-0000_BIOS-3.21.00_4.11.05.00_0x05000000_HII-_WEBBIOS-6.0-34-e_29-Rel_PCLI-04.04-01700008_BOOT-01.250.04.219_NVDATA-2.09.03-0010_2011_02_25                

     

 

Bug Fixes and Enhancements:

===========================

Firmware:

=========

LSIP200167677 (CO) FW should issue STANDBY IMMEDIATE command to SATA disks during shutdown from OS

LSIP200164222 (CO) Add mfc paramater that enables prints of topology changes in fw serial port output

LSIP200158868 (DFCT) Power save mode hangs firmware causing driver to send kill adpter command

LSIP200170440 (DFCT) Firmware Incompleted Writes During Power Fail and Recovery

LSIP200153454 (DFCT) Firmware Incompleted Writes During Power Fail and Recovery

LSIP200125226 (DFCT) Red LED is lit on the next empty slot in an specific JBOD if the empy slot is immediatly after an array member

LSIP200153500 (DFCT) 9280-4i4e unable to import a Foreign Configuration

LSIP200166039 (DFCT) Label of Virtual Disk not provided by Drive Identify Command

LSIP200164210 (DFCT) Enable/disable of HDD cache doesn't appear to work

LSIP200167654 (DFCT) Potential data mismatch found when rebuild is progressing on R1 VD with media error under IO.

LSIP200166899 (DFCT) HWR: Fw log should indicate if a created/existing VD is already initialized

LSIP200157150 (DFCT) The new event can't be logged.

LSIP200156699 (DFCT) MegaCli -AdpAllInfo shows BBU is not supported for Castor cards

LSIP200146481 (DFCT) SMART predictive failure detection takes longer in 10M09 than 10M06

LSIP200146484 (DFCT) MR FW reports "Unexpected sense, Invalid field in CDB" for LOG SENSE command sent to Seagate drives

LSIP200169751 (DFCT) MegaRaid FW does not wait at post

 

BIOS:

========

HII changes only

 

 

Installation:

=============

Use MegaCLI to flash the SAS controllers.  MegaCLI can be downloaded from the support and download

section of .

 

Command syntax:  MegaCli -adpfwflash -f mr2108fw.rom -a0

 

 

Lab 1 : 练习Niwot卡的管理及使用

 

Hardware Resource : T4-2/ 375-3701 [C]         6Gigabit SAS RAID PCI Express HBA, B4 ASIC

Software Resource : Fcode utility/ LSI SAS2 2008 RAID Management Utility/ Oracle Hardware Management Pack

 

Task 1 : Identify internal disks and create hard raid volumes by Fcode

 

{0} ok select LSI2008 raid chip physical path

 

{0} ok show-volumes

 

{0} ok show-children

 

create a volume for root disk

 

Task2 : Install OHMP and use OHMP or sas2ircu to identify volume status; Identify physical disk location

 

a. Download OHMP from MOS

 

b. Download sas2ircu from LSI website

 

c. Show volume status and identify disk members in physical slot by command output

 

Task3: Dummy replace the 371-4982 6Gigabit SAS RAID PCI Battery Module and connect disk cables to Niwot card

 

a. Remove and reinstall the BBU

 

b. Connect two disk cables from motherboard to Niwot card

 

Task4: Practice the PCLI commands under OBP

 

a. Find out the Niwot physical path

 

b. Use PCLI commands get Niwot configuration

 

1. “cli -PdList” will display all the physical disk

under the HBA's control.

 

2. “cli -LDInfo -Lall” will display current LUNs

associated with the HBA. LUNs can also be

referred to as a Virtual Drive.

 

3. “cli -EncInfo” shows more information about the

Enclosure, like enclosure id and # of enclosure

slots.

 

4. “cli -CfgDsply” displays information about HBA

( Serial No., Backup Battery Unit presence, Cache

Memory Size and LSI Product Name)

“cli -LdPdInfo” maps create LUNs with physical

Disks.

 

c. Create RAID volume, show volume status and delete volume

 

1. ok cli -CfgLdAdd -r1[252:0,252:1]

{0} ok cli -LDInfo -L0

{0} ok cli -CfgLdDel -L0

 

2. ok cli -CfgLdAdd -r0[252:0]

{0} ok cli -LDInfo -L0

 

3. ok cli -CfgLdAdd -r5[252:2,252:3,252:4]

{0} ok cli -LDInfo -L1

 

d. Install Solaris in volume

 

Task5: Install MegaCli and practice to use it

 

a. Download and install the MegaCli package for Solaris SPARC

Download URL:

 

b. Verify MegaCli installation and Niwot card status

 

# ./MegaCli -v

# ./MegaCli -Version -Ctrl -aALL

# ./MegaCli -Version -Driver -aALL

# ./MegaCli -LDInfo -Lall -aALL

 

# ./MegaCli -CfgDsply -a0

 

How Many Disks seen by HBA

# ./MegaCli -PDGetNum -a0

 

Check Battery Status

# ./MegaCli -AdpBbuCmd -GetBbuStatus -aALL

 

More Backup Battery Info

# ./MegaCli -AdpBbuCmd -GetBbuCapacityInfo -aALL

 

HBA Cache Policy

# ./MegaCli -LDGetProp -Cache -LALL -aALL

 

c. Collect HBA event Logs:

 

# ./MegaCli -AdpEventLog -GetEvents -f eventlogs.txt -a0

 

# ls -l

 

Collecting HBA Terminal Log:

 

# ./MegaCli -fwtermlog -dsply -a0 > terminallog.txt

ls -l

 

# ./MegaCli -fwtermlog -Clear -a0

 

d. Locate Disk ( LED Blink ):

 

./MegaCli -PdLocate -start -physdrv[252:0] -a0

 

./MegaCli -PdLocate -stop -physdrv[252:0] -a0

 

The Attention LED (AMBER Light) on the disk will blink until the

disk locate is manually terminated.

 

e. Adjusting Rebuild Rate:

 

./MegaCli -AdpGetProp RebuildRate -a0

 

./MegaCli -AdpSetProp RebuildRate -70

 

(Default is 30%, higher rebuild rate will adversely impact RAID I/O

Performance)

 

f.  Convert Raid 0 to Raid 1 online:

 

When system is pre-configured from Factory with the NIWOT card, usually only one Raid 0 LUN is

created by default with Solaris OS pre-installed. If the customer wants to have a RAID 1 disk

configured we can Download the MegaCLI package and run the following command to convert the

RAID 0 to a RAID 1 without re-installing the OS. This activity May take a few hours depending on

disk size, disk performance and system load.

# ./MegaCli -LDRecon -Start -r1 -Add -PhysDrv[252:1] -L0 -a0

 

g. Adding/Removing Global HotSpare:

 

# ./MegaCli -PDHSP -Set -PhysDrv [252:5] -a0

 

# ./MegaCli -PDHSP -Rmv -PhysDrv[252:5] -a0

 

 

h. Convert Raid 5 to Raid 6 online:

 

# ./MegaCli -LDRecon -Start -r6 -Add -PhysDrv[252:5] -L1 -a0

# ./MegaCli -LDRecon -ShowProg -L1 -a0

 

# ./MegaCli -LDRecon -ProgDsply -L1 -a0

 

# ./MegaCli -LDInfo -Lall -aALL

上一篇:查看Oracle server的物理及虚拟CPU
下一篇:没有了