Iss 007 5818 003

User Manual: 007-5818-003

Open the PDF directly: View PDF PDF.
Page Count: 178 [warning: Documents this large are best viewed by clicking the View PDF Link!]

SGI® Modular InfiniteStorage™ (MIS) Platform
User Guide
007-5818-003
COPYRIGHT
© 2012 Silicon Graphics International Corp. All rights reserved; provided portions may be copyright in third parties, as indicated elsewhere herein. No
permission is granted to copy, distribute, or create derivative works from the contents of this electronic documentation in any manner, in whole or in part,
without the prior written permission of SGI.
LIMITED RIGHTS LEGEND
The software described in this document is “commercial computer software” provided with restricted rights (except as to included open/free source) as specified
in the FAR 52.227-19 and/or the DFAR 227.7202, or successive sections. Use beyond license provisions is a violation of worldwide intellectual property laws,
treaties and conventions. This document is provided with limited rights as defined in 52.227-14.
The electronic (software) version of this document was developed at private expense; if acquired under an agreement with the USA government or any
contractor thereto, it is acquired as “commercial computer software” subject to the provisions of its applicable license agreement, as specified in (a) 48 CFR
12.212 of the FAR; or, if acquired for Department of Defense units, (b) 48 CFR 227-7202 of the DoD FAR Supplement; or sections succeeding thereto.
Contractor/manufacturer is SGI, 46600 Landing Parkway, Fremont, CA 94538.
TRADEMARKS AND ATTRIBUTIONS
Silicon Graphics, SGI, the SGI logo, InfiniteStorage, and Supportfolio are trademarks or registered trademarks of Silicon Graphics International Corp. or its
subsidiaries in the United States and/or other countries worldwide.
Fusion-MPT, Integrated RAID, MegaRAID, and LSI Logic are trademarks or registered trademarks of LSI Logic Corporation.InfiniBand is a registered
trademark of the InfiniBand Trade Association. Intel and Xeon are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United
States and other countries. Internet Explorer and Windows are registered trademarks of Microsoft Corporation. Java and Java Virtual Machine are trademarks
or registered trademarks of Sun Microsystems, Inc. Linux is a registered trademark of Linus Torvalds, used with permission by SGI. Novell and Novell Netware
are registered trademarks of Novell Inc. PCIe and PCI-X are registered trademarks of PCI SIG. Red Hat and all Red Hat-based trademarks are trademarks or
registered trademarks of Red Hat, Inc. in the United States and other countries. Sharp is a registered trademark of Sharp corporation. SUSE LINUX and the
SUSE logo are registered trademarks of Novell, Inc. VMware is a registered trademark or trademark of VMware, Inc. in the United States and/or other
jurisdictions.
All other trademarks mentioned herein are the property of their respective owners.
Adaptec, HostRAID, and the Adaptec logo are registered trademarks of Adaptec Inc.
007-5818-003 iii
Record of Revision
Version Description
001 June 2012
Original printing.
002 August 2012
Edited and updated for technical and editorial changes. Images updated to reflect
changes, added Zones and CLI Zoning Tool software information, available RAID
configurations updated.
003 October 2012
Electromagnetic Compatibility (EMC) compliance and safety information included.
Section on CPU/Riser/HBA configuration restrictions and options added. Weight
safety maximums included. Additional information on RAID options and conditions
provided. Includes updates from the stand-alone version of Chapter 3 issued in
September 2012. Information on grouping control added per customer feedback
(Appendix C).
iv 007-5818-003
:
007-5818-003 v
Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . xix
Audience. . . . . . . . . . . . . . . . . . . . . . . . . . xix
Important Information . . . . . . . . . . . . . . . . . . . . . . xix
Safety Precautions . . . . . . . . . . . . . . . . . . . . . . xix
ESD Precautions . . . . . . . . . . . . . . . . . . . . . . xxi
Safety & Emissions . . . . . . . . . . . . . . . . . . . . . . . xxi
Electromagnetic Compatibility . . . . . . . . . . . . . . . . . . .xxii
Safety Certification . . . . . . . . . . . . . . . . . . . . . .xxii
Chapter Descriptions . . . . . . . . . . . . . . . . . . . . . xxiii
Related Publications. . . . . . . . . . . . . . . . . . . . . . xxiv
Conventions . . . . . . . . . . . . . . . . . . . . . . . . .xxv
Product Support . . . . . . . . . . . . . . . . . . . . . . . .xxv
CRU/FRU . . . . . . . . . . . . . . . . . . . . . . . .xxv
Purchasable Support & Maintenance Programs . . . . . . . . . . . . . xxvi
Reader Comments . . . . . . . . . . . . . . . . . . . . . . xxvi
1 System Overview . . . . . . . . . . . . . . . . . . . . . . . 1
MIS Enclosure . . . . . . . . . . . . . . . . . . . . . . . . 5
Front Grille and Control Panels . . . . . . . . . . . . . . . . . . 6
Rear Panel Components . . . . . . . . . . . . . . . . . . . . 7
MIS Common Modules . . . . . . . . . . . . . . . . . . . . . . 10
Power Supply Module . . . . . . . . . . . . . . . . . . . . . 10
Fan Assembly Module . . . . . . . . . . . . . . . . . . . . . 11
StorBrick Module . . . . . . . . . . . . . . . . . . . . . . 12
MIS Server Platform or JBOD Unit. . . . . . . . . . . . . . . . . . . 14
Server Module . . . . . . . . . . . . . . . . . . . . . . . 14
Layout of Server CPUs, and PCIe Risers HBAs . . . . . . . . . . . . . . 20
Boot Drives Module . . . . . . . . . . . . . . . . . . . . . 21
vi 007-5818-003
Contents
MIS JBOD I/O Module . . . . . . . . . . . . . . . . . . . . . 22
System Block Diagram . . . . . . . . . . . . . . . . . . . . . . 23
2 System Interfaces . . . . . . . . . . . . . . . . . . . . . . . 25
Control Panel . . . . . . . . . . . . . . . . . . . . . . . . . 25
MIS Server Control Panel . . . . . . . . . . . . . . . . . . . . 25
MIS JBOD Control Panel . . . . . . . . . . . . . . . . . . . . 27
Disk Drive LEDs . . . . . . . . . . . . . . . . . . . . . . . . 28
Power Supply LEDs . . . . . . . . . . . . . . . . . . . . . . . 28
BMC Integrated Web Console . . . . . . . . . . . . . . . . . . . . 29
System Information . . . . . . . . . . . . . . . . . . . . . . 30
FRU Information . . . . . . . . . . . . . . . . . . . . . 31
System Debug Log . . . . . . . . . . . . . . . . . . . . . 32
CPU Information . . . . . . . . . . . . . . . . . . . . . 34
DIMM Information . . . . . . . . . . . . . . . . . . . . . 34
Server Health . . . . . . . . . . . . . . . . . . . . . . . 35
Sensor Readings . . . . . . . . . . . . . . . . . . . . . 35
Event Log . . . . . . . . . . . . . . . . . . . . . . . 36
Power Statistics . . . . . . . . . . . . . . . . . . . . . . 37
Configuration Tab . . . . . . . . . . . . . . . . . . . . . . 38
IPv4 Network . . . . . . . . . . . . . . . . . . . . . . 39
IPv6 Network . . . . . . . . . . . . . . . . . . . . . . 40
Users . . . . . . . . . . . . . . . . . . . . . . . . 42
Login . . . . . . . . . . . . . . . . . . . . . . . . 43
LDAP . . . . . . . . . . . . . . . . . . . . . . . . 45
VLAN . . . . . . . . . . . . . . . . . . . . . . . . 46
SSL . . . . . . . . . . . . . . . . . . . . . . . . . 48
Remote Session . . . . . . . . . . . . . . . . . . . . . . 48
Mouse Mode . . . . . . . . . . . . . . . . . . . . . . 50
Keyboard Macros . . . . . . . . . . . . . . . . . . . . . 52
Alerts . . . . . . . . . . . . . . . . . . . . . . . . 54
Alert Email . . . . . . . . . . . . . . . . . . . . . . . 55
Node Manager . . . . . . . . . . . . . . . . . . . . . . 56
Remote Control Tab . . . . . . . . . . . . . . . . . . . . . . 59
Contents
007-5818-003 vii
Console Redirection . . . . . . . . . . . . . . . . . . . . 59
Server Power Control . . . . . . . . . . . . . . . . . . . . . 61
3 System Software . . . . . . . . . . . . . . . . . . . . . . . 65
Overview . . . . . . . . . . . . . . . . . . . . . . . . . 65
Section Guide . . . . . . . . . . . . . . . . . . . . . . . . 66
Linux Zoning Tools . . . . . . . . . . . . . . . . . . . . . . . 68
Installing Linux Software . . . . . . . . . . . . . . . . . . . . 68
MegaRAID Storage Manager for Linux . . . . . . . . . . . . . . . 68
MegaCli64 for Linux . . . . . . . . . . . . . . . . . . . . 69
Zones for Linux . . . . . . . . . . . . . . . . . . . . . 69
CLI Zoning Tool for Linux . . . . . . . . . . . . . . . . . . 70
Verify Drives Seen. . . . . . . . . . . . . . . . . . . . . 71
Linux Zones Tool . . . . . . . . . . . . . . . . . . . . . . 71
Installing a Drive through Zones for Linux . . . . . . . . . . . . . . 74
Creating the Drive Groups in MegaRAID for Linux . . . . . . . . . . . . 77
Formatting the Drives using YaST2 in Linux . . . . . . . . . . . . . . 79
Removing a Drive in Zones for Linux. . . . . . . . . . . . . . . . 83
Additional Features in Zones for Linux . . . . . . . . . . . . . . . . 85
Loading .csv Configuration Files in Linux . . . . . . . . . . . . . . 86
Save a Configuration to .csv file in Linux . . . . . . . . . . . . . . . 86
Windows Zoning Tools . . . . . . . . . . . . . . . . . . . . . . 88
Installing Windows Software . . . . . . . . . . . . . . . . . . . 88
MegaRAID Storage Manager for Windows . . . . . . . . . . . . . . 88
MegaCli64 for Windows . . . . . . . . . . . . . . . . . . . 89
Zones for Windows . . . . . . . . . . . . . . . . . . . . 89
Python for Windows . . . . . . . . . . . . . . . . . . . . 91
CLI Zoning Tool for Windows . . . . . . . . . . . . . . . . . 91
Verify Drives Seen in Windows . . . . . . . . . . . . . . . . . 92
Zones for Windows. . . . . . . . . . . . . . . . . . . . . . 92
Installing a Drive in Zones for Windows . . . . . . . . . . . . . . . 95
Creating the Drive Groups in MegaRAID for Windows . . . . . . . . . . . 98
Formatting the Drives in Windows Server Manager . . . . . . . . . . . .100
Removing a Drive in Zones for Windows . . . . . . . . . . . . . . .106
viii 007-5818-003
Contents
Additional Features in Zones for Windows . . . . . . . . . . . . . . . 107
Loading .csv Configuration Files in Zones for Windows . . . . . . . . . . 107
Adapter Assignment Synchronization in Zones for Windows . . . . . . . . . 108
Save a Configuration to .csv file in Zones for Windows . . . . . . . . . . 109
CLI Zoning Tool . . . . . . . . . . . . . . . . . . . . . . . 109
Preparing to Zone using the CLI Zoning Tool . . . . . . . . . . . . . . 110
Editing the ShackCLI.ini file for Linux . . . . . . . . . . . . . . 111
Editing the ShackCLI.ini file for Windows . . . . . . . . . . . . . 112
CLI Zoning Tool Main Menu . . . . . . . . . . . . . . . . . 113
Zoning Using CLI Zoning Tool . . . . . . . . . . . . . . . . . . 117
Editing the .csv File for the CLI Zoning Tool . . . . . . . . . . . . . . 119
Disk RAID Support . . . . . . . . . . . . . . . . . . . . . . 122
RAID 0 . . . . . . . . . . . . . . . . . . . . . . . . 123
RAID 1 . . . . . . . . . . . . . . . . . . . . . . . . 124
RAID 5 . . . . . . . . . . . . . . . . . . . . . . . . 125
RAID 6 . . . . . . . . . . . . . . . . . . . . . . . . 125
RAID 00 . . . . . . . . . . . . . . . . . . . . . . . . 126
RAID 10 . . . . . . . . . . . . . . . . . . . . . . . . 127
RAID 50 . . . . . . . . . . . . . . . . . . . . . . . . 128
RAID 60 . . . . . . . . . . . . . . . . . . . . . . . . 128
RAID Configuration Notes . . . . . . . . . . . . . . . . . . . 129
4 System Maintenance . . . . . . . . . . . . . . . . . . . . . 133
Detecting Component Failures . . . . . . . . . . . . . . . . . . . 134
Sliding the Chassis Forward/Backwards . . . . . . . . . . . . . . . . . 134
Removing the Front or Rear Chassis Cover . . . . . . . . . . . . . . . . 134
Replacing a Power Supply . . . . . . . . . . . . . . . . . . . . 135
Replacing a Fan Module . . . . . . . . . . . . . . . . . . . . . 136
Replacing a Disk Drive . . . . . . . . . . . . . . . . . . . . . 137
Removing the Drive . . . . . . . . . . . . . . . . . . . . . 138
Re-installing the Drive . . . . . . . . . . . . . . . . . . . . 139
Checking the System Air Flow . . . . . . . . . . . . . . . . . . . 140
Contents
007-5818-003 ix
5 Troubleshooting . . . . . . . . . . . . . . . . . . . . . . .141
No Video . . . . . . . . . . . . . . . . . . . . . . . . .141
Losing the System’s Setup Configuration . . . . . . . . . . . . . . . . .141
I/O Time-outs and MegaRAID Drivers . . . . . . . . . . . . . . . . . .142
Safe Power-Off . . . . . . . . . . . . . . . . . . . . . . . .142
A Technical Specifications . . . . . . . . . . . . . . . . . . . . .143
B BIOS Error Codes . . . . . . . . . . . . . . . . . . . . . . .147
C Zone Permission Groups Rules . . . . . . . . . . . . . . . . . . .149
x007-5818-003
Contents
007-5818-003 xi
Figures
Figure 1-1 SGI Destination Rack (D-Rack) . . . . . . . . . . . . 4
Figure 1-2 MIS Chassis and Case . . . . . . . . . . . . . . 5
Figure 1-3 Bi-directional rail mount . . . . . . . . . . . . . . 6
Figure 1-4 Single Control Panel . . . . . . . . . . . . . . . 7
Figure 1-5 Dual Control Panel . . . . . . . . . . . . . . . 7
Figure 1-6 Rear View – MIS Server Platform (single server) . . . . . . . 8
Figure 1-7 Rear View – MIS Server Platform (single server, four power supplies). . 8
Figure 1-8 Rear View – MIS Server Platform (dual server). . . . . . . . 8
Figure 1-9 Rear View – MIS Server Platform (single dual-server module) . . . . 9
Figure 1-10 Rear View – MIS JBOD Unit (single I/O module) . . . . . . . 9
Figure 1-11 Rear View – MIS JBOD unit (dual I/O modules) . . . . . . . 9
Figure 1-12 Power Supply Module (rated at 1100 Watts) . . . . . . . . 11
Figure 1-13 Power Supply Numbering . . . . . . . . . . . . . 11
Figure 1-14 Fan Assembly Module (each contains two impellers) . . . . . . 11
Figure 1-15 StorBrick Modules for 3.5" or 2.5" 15mm Drives (left)
and 2.5" 9.5 mm Drives (right) . . . . . . . . . . . . 12
Figure 1-16 3.5" 15mm Drive and Carrier . . . . . . . . . . . . 13
Figure 1-17 3.5" 15mm Drive Carrier (top view, with thumb latch) . . . . . . 13
Figure 1-18 Two 2.5" 9.5mm Drives and Carrier . . . . . . . . . . . 13
Figure 1-19 2.5" 9.5mm Drive Carrier (isometric view with dual thumb latches) . . 14
Figure 1-20 MIS Server Platform (single server) . . . . . . . . . . . 15
Figure 1-21 MIS Server Platform (dual server) . . . . . . . . . . . 16
Figure 1-22 MIS JBOD Unit . . . . . . . . . . . . . . . . 17
Figure 1-23 MIS Server Module – single server . . . . . . . . . . . 18
Figure 1-24 Server Module (single server) – component view . . . . . . . 18
Figure 1-25 MIS Server Module – dual server (half height) . . . . . . . . 19
Figure 1-26 Dual Server Module – component view . . . . . . . . . . 19
Figure 1-27 CPU and PCIe Riser layout . . . . . . . . . . . . . 20
xii 007-5818-003
Figure 1-28 HBA population layout . . . . . . . . . . . . . . 21
Figure 1-29 Boot Drive Module . . . . . . . . . . . . . . . 22
Figure 1-30 I/O Module for MIS JBOD Unit . . . . . . . . . . . . 22
Figure 1-31 MIS JBOD Midplane I/O Connector (right & left views) . . . . . 23
Figure 1-32 System-Level Block Diagram. . . . . . . . . . . . . 24
Figure 2-1 MIS Control Panel . . . . . . . . . . . . . . . 26
Figure 2-2 Disk Drive LEDs . . . . . . . . . . . . . . . . 28
Figure 2-3 Power Supply LEDs
. . . . . . . . . . . . . . . . . . . . 28
Figure 2-4 BMC Web Console – System Information Page . . . . . . . . 31
Figure 2-5 BMC Web Console – FRU Information . . . . . . . . . . 32
Figure 2-6 BMC Web Console – System Debug Log . . . . . . . . . 33
Figure 2-7 BMC Web Console – CPU Information . . . . . . . . . . 34
Figure 2-8 BMC Web Console – DIMM Information . . . . . . . . . 35
Figure 2-9 BMC Web Console – Server Health . . . . . . . . . . . 36
Figure 2-10 BMC Web Console – Event Log . . . . . . . . . . . . 37
Figure 2-11 BMC Web Console – Power Statistics . . . . . . . . . . 38
Figure 2-12 BMC Web Console – IPv4 Network Settings . . . . . . . . 40
Figure 2-13 BMC Web Console – IPv6 Network Settings . . . . . . . . 42
Figure 2-14 BMC Web Console – Login Security Settings . . . . . . . . 44
Figure 2-15 BMC Web Console – LDAP Settings. . . . . . . . . . . 46
Figure 2-16 BMC Web Console – VLAN Settings . . . . . . . . . . 47
Figure 2-17 BMC Web Console – SSL Upload . . . . . . . . . . . 48
Figure 2-18 BMC Web Console – Remote Session . . . . . . . . . . 50
Figure 2-19 BMC Web Console – Mouse Mode Setting . . . . . . . . . 51
Figure 2-20 BMC Web Console – Logout, Refresh, and Help buttons . . . . . 52
Figure 2-21 BMC Web Console – Keyboard Macros . . . . . . . . . . 52
Figure 2-22 BMC Web Console – Alerts . . . . . . . . . . . . . 55
Figure 2-23 BMC Web Console – Alert Email Settings . . . . . . . . . 56
Figure 2-24 BMC Web Console – Node Manager Power Policies . . . . . . 57
Figure 2-25 BMC Web Console – Console Redirection (greyed-out). . . . . . 60
Figure 2-26 BMC Web Console – Launch Console button (available) . . . . . 60
Figure 2-27 Remote Console – Java redirection window . . . . . . . . . 61
Figures
007-5818-003 xiii
Figure 2-28 BMC Web Console – Power Control and Status. . . . . . . . 62
Figure 3-1 Zones – First-use path configuration. . . . . . . . . . . 70
Figure 3-2 Zones – Error message: Improper path . . . . . . . . . . 70
Figure 3-3 Linux Zones Welcome . . . . . . . . . . . . . . 71
Figure 3-4 Zones User Interface . . . . . . . . . . . . . . . 72
Figure 3-5 Zones – Show All, Open Session, Save Session, Download Session, and Exit
buttons . . . . . . . . . . . . . . . . . . 72
Figure 3-6 Zones – Opening a Session . . . . . . . . . . . . . 73
Figure 3-7 Zones – Enter Session Alias . . . . . . . . . . . . . 73
Figure 3-8 Zones – Alias Help Warning Message . . . . . . . . . . 74
Figure 3-9 Zones – Adapter tabs . . . . . . . . . . . . . . . 74
Figure 3-10 Zones – Show All . . . . . . . . . . . . . . . 75
Figure 3-11 Zones – Adapter Assignment Warning Message . . . . . . . 75
Figure 3-12 Zones – Select StorBricks for download . . . . . . . . . . 76
Figure 3-13 MegaRAID – Create a Virtual Drive. . . . . . . . . . . 77
Figure 3-14 MegaRAID – Create Virtual Drive mode . . . . . . . . . 77
Figure 3-15 Create Virtual Drive – Drive Group Settings . . . . . . . . 78
Figure 3-16 Create Virtual Drive – Summary. . . . . . . . . . . . 78
Figure 3-17 Yast2 Server Manager GUI . . . . . . . . . . . . . 79
Figure 3-18 YaST2 – Warning Message . . . . . . . . . . . . . 80
Figure 3-19 YaST2 – Drives have appeared . . . . . . . . . . . . 80
Figure 3-20 YaST2 – Add button . . . . . . . . . . . . . . . 81
Figure 3-21 YaST2 – Select Partition Size . . . . . . . . . . . . 81
Figure 3-22 YaST2 – Format & Mount the Drive. . . . . . . . . . . 81
Figure 3-23 YaST2 – Check for Partition . . . . . . . . . . . . . 82
Figure 3-24 YaST – Click Finish . . . . . . . . . . . . . . . 82
Figure 3-25 YaST2 – Disk Mounting (in process) . . . . . . . . . . 83
Figure 3-26 Zones – Downloading setting to StorBrick(s) . . . . . . . . 84
Figure 3-27 Zones – Are you sure to download files to expander? . . . . . . 85
Figure 3-28 Zones – Select CSV File (Linux) . . . . . . . . . . . 86
Figure 3-29 Zones – Save Configuration to .csv . . . . . . . . . . . 86
Figure 3-30 Zones – Save CSV pop-up . . . . . . . . . . . . . 87
Figure 3-31 Zones – Select Directory navigation pane . . . . . . . . . 87
xiv 007-5818-003
Figure 3-32 Zones – Error: Improper Directory Selection. . . . . . . . . 88
Figure 3-33 Zones – First-use path configuration (Windows) . . . . . . . . 90
Figure 3-34 Zones – Error messages from improper path configuration . . . . . 91
Figure 3-35 Windows Server Manager – Disk Management . . . . . . . . 92
Figure 3-36 Zones for Windows Welcome . . . . . . . . . . . . 93
Figure 3-37 Zones Windows User Interface . . . . . . . . . . . . 93
Figure 3-38 Zones – Open Session, Save Session, Download Session, and Exit buttons . 93
Figure 3-39 Zones – Open Session. . . . . . . . . . . . . . . 94
Figure 3-40 Zones – Enter Session Alias . . . . . . . . . . . . . 94
Figure 3-41 Zones – Alias Help Warning Message . . . . . . . . . . 95
Figure 3-42 Zones – Adapter tabs (Windows) . . . . . . . . . . . . 95
Figure 3-43 Zones – Show All . . . . . . . . . . . . . . . . 96
Figure 3-44 Zones – Adapter Assignment Warning Message . . . . . . . . 96
Figure 3-45 Zones – Select StorBricks for download . . . . . . . . . . 97
Figure 3-46 Zones tool – Verify download . . . . . . . . . . . . 98
Figure 3-47 MegaRAID – Create a Virtual Drive . . . . . . . . . . . 98
Figure 3-48 MegaRAID – Create Virtual Drive mode. . . . . . . . . . 99
Figure 3-49 Create Virtual Drive – Simple Settings . . . . . . . . . . 99
Figure 3-50 Create Virtual Drive – Summary . . . . . . . . . . . 100
Figure 3-51 Server Manager – Disk Management . . . . . . . . . . 101
Figure 3-52 Server Manager – Initialize Disks . . . . . . . . . . 101
Figure 3-53 Server Manager – Select GPT (GUID Partition Table) . . . . . 102
Figure 3-54 Server Manager – Disks Initialized and Online . . . . . . . 102
Figure 3-55 Server Manager – New Simple Volume . . . . . . . . . 103
Figure 3-56 Server Manager – New Simple Volume Wizard . . . . . . . 103
Figure 3-57 New Simple Volume Wizard – Volume Size. . . . . . . . 104
Figure 3-58 New Simple Volume Wizard – Assign Drive Letter or Path . . . . 104
Figure 3-59 New Simple Volume Wizard – Format Partition . . . . . . . 105
Figure 3-60 New Simple Volume Wizard – Settings Confirmation . . . . . 105
Figure 3-61 New Simple Volume in Server Manager . . . . . . . . . 106
Figure 3-62 Zones – Verify Download. . . . . . . . . . . . . 106
Figure 3-63 Zones – Select CSV File . . . . . . . . . . . . . 107
Figure 3-64 Zones – Error message: Canceling csv file selection (Windows) . . 108
Figures
007-5818-003 xv
Figure 3-65 Zones – Select All, Unselect All buttons. . . . . . . . . .108
Figure 3-66 MIS-S9D proprietary network interface . . . . . . . . . .111
Figure 3-67 Block diagram of MIS-Server StorBrick SB0 . . . . . . . .121
Figure 3-68 RAID 0 . . . . . . . . . . . . . . . . . .123
Figure 3-69 RAID 1 . . . . . . . . . . . . . . . . . .124
Figure 3-70 RAID 5 . . . . . . . . . . . . . . . . . .125
Figure 3-71 RAID 6 . . . . . . . . . . . . . . . . . .125
Figure 3-72 RAID 00 . . . . . . . . . . . . . . . . . .126
Figure 3-73 RAID 10 . . . . . . . . . . . . . . . . . .127
Figure 3-74 RAID 50 . . . . . . . . . . . . . . . . . .128
Figure 3-75 RAID 60 . . . . . . . . . . . . . . . . . .128
Figure 3-76 RAID 1 with one drive per StorBrick . . . . . . . . . .129
Figure 3-77 RAID 1 with two drives spanning a StorBrick . . . . . . . .130
Figure 3-78 RAID 5 or 6 with one drive per StorBrick . . . . . . . . .131
Figure 3-79 Loss of a drive with multiple drives on a StorBrick does not affect RAID 6, but
will impact RAID 5 . . . . . . . . . . . . . . .131
Figure 3-80 Three drive loss in RAID 6 require StorBrick replacement . . . . .131
Figure 4-1 Front & Rear Chassis Covers . . . . . . . . . . . .135
Figure 4-2 Replacing a Power Supply . . . . . . . . . . . . .136
Figure 4-3 Replacing a Fan Module . . . . . . . . . . . . . .137
Figure 4-4 Hard Drive Carrier . . . . . . . . . . . . . . .139
Figure 4-5 MIS Chassis Midspan Support Brace . . . . . . . . . .140
Figure C-1 Zone Permission Groups – Example 1 . . . . . . . . . .150
Figure C-2 Zone Permission Groups – Example 2 . . . . . . . . . .151
xvi 007-5818-003
007-5818-003 xvii
Tables
Table -1 MIS Server Platform Region and EMC Compliance References . . .xxii
Table -2 MIS Server Platform Region and EMC Compliance References . . xxiii
Table 2-1 MIS Server Platform Control Panel Buttons and LEDs . . . . . . 26
Table 2-2 Disk Drive LEDs . . . . . . . . . . . . . . . 28
Table 2-3 Power Supply LEDs . . . . . . . . . . . . . . . 29
Table 2-4 System Information Details . . . . . . . . . . . . . 30
Table 2-5 Supported Key Names . . . . . . . . . . . . . . 53
Table 2-6 Server Power Control Actions . . . . . . . . . . . . 61
Table 2-7 System Information Details . . . . . . . . . . . . . 63
Table 3-1 Zone Group Implementation . . . . . . . . . . . . .110
Table 3-2 CLI Zoning Tool Menu Options and Descriptions . . . . . . .114
Table 3-3 Zone Group Implementation . . . . . . . . . . . . .120
Table A-1 Technical Specifications . . . . . . . . . . . . . .143
Table B-1 BMC Beep Codes . . . . . . . . . . . . . . .147
xviii 007-5818-003
007-5818-003 xix
.Introduction
This guide describes the features and components of the SGI® Modular InfiniteStorage™ (MIS)
platform. With two main configurations possible for the enclosure (server and storage, or JBOD—
Just Bunch Of Disks) this guide covers the different configurations, their respective components,
interface panels, indicator lights and meanings, software, maintenance, and troubleshooting.
Audience
This guide is written for owners/users of the MIS platform. It is written with the assumption that
the reader has a good working knowledge of computers, servers, networking, hardware, software
and RAID arrays.
Important Information
The following section details several safety precautions that should be observed at all times. First,
a fully loaded MIS Platform can weigh up to 220lbs. Second, electricity is a major concern,
especially Electrostatic Discharge (ESD), detailed later in this section. Please read these sections
carefully prior to using the MIS Platform.
Safety Precautions
Do NOT wear loose clothing, such as neckties or unbuttoned shirt sleeves, while working on the
unit which can be pulled into a cooling fan or tangled in cabling.
Remove any jewelry any metal objects from your body, which are excellent electrical conductors,
and can harm you and/or cause short circuits if they come into contact with printed circuit boards
or powered areas.
xx 007-5818-003
Introduction
Be aware of the locations of the power on/off switch on the chassis as well as the room's
emergency power-off switch, disconnection switch or electrical outlet. If an electrical accident
occurs, you can then quickly remove power from the system.
Do NOT work alone when working with high voltage components.
When working around exposed electrical circuits, another person should be nearby, who is
familiar with the power-off controls, to switch off the power if necessary.
Use only one hand when working with powered-on electrical equipment. This is to avoid making
a complete circuit, which will cause electrical shock. Use extreme caution when using metal tools,
which can easily damage any electrical components or circuit boards with which they come into
contact.
Do NOT use mats designed to decrease static electrical discharge as protection from electrical
shock. Instead, use rubber mats that have been specifically designed as electrical insulators.
The power supply power cords must include a grounding plug and must be plugged into grounded
electrical outlets.
Do NOT attempt to transport/move a fully loaded MIS system. An MIS system can weigh up to
220lbs. when fully loaded. If the system must be moved, first remove the drives from the chassis.
When lifting the system, two people (one at each end) should lift slowly with feet spread apart to
distribute the weight. Always follow safe lifting practices when moving heavy objects. More
information on moving large objects, requiring a two-person team, is available in the Centers for
Disease Control’s, “Ergonomic Guidelines for Manual Material Handling”
(http://www.cdc.gov/niosh/docs/2007-131/pdfs/2007-131.pdf)
Power should always be disconnected from the system when removing or installing system
components that are not hot-swappable, such as server boards and memory modules. When
disconnecting power, you should first do a clean shut down of the operating system, then power
down the system, and then unplug all power cords (the unit has more than one power supply cord).
More information on powering off the MIS Platform is available in Chapter 4, “System
Maintenance.”
Introduction
007-5818-003 xxi
ESD Precautions
Caution: Electrostatic Discharge (ESD) is generated by two objects with different electrical
charges coming into contact with each other. An electrical discharge is created to neutralize this
difference, which can damage electronic components and printed circuit boards.
The following measures are generally sufficient to neutralize this difference before contact is
made to protect your equipment from ESD:
Use a grounded wrist strap designed to prevent static discharge.
Keep all components and printed circuit boards (PCBs) in their antistatic bags until ready for
use.
Touch a grounded metal object before removing the board from the antistatic bag.
Do not let components or PCBs come into contact with your clothing, which may retain a
charge even if you are wearing a wrist strap.
Handle a board by its edges only; do not touch its components, peripheral chips, memory
modules or contacts.
When handling chips or modules, avoid touching their pins.
Put the server board and peripherals back into their antistatic bags when not in use.
For grounding purposes, make sure your computer chassis provides excellent conductivity
between the power supply, the case, the mounting fasteners and the server board.
Safety & Emissions
The following is a list of agency approvals for MIS on safety and emissions.
!
xxii 007-5818-003
Introduction
Electromagnetic Compatibility
Table -1 lists the region and compliance reference for EMC (Electromagnetic Compatibility)
compliance.
Safety Certification
Underwriters Laboratories (UL) provides safety certification for electronic devices. UL offers a
Functional Safety Listing Mark that can be added for those qualifying companies in the process
of getting a traditional Listing from UL. In essence, the Functional Safety Listing Mark replaces
the traditional UL listing mark on products certified for functional safety. Functional safety
examines the efficacy of the safety-related system by considering the input variables to a device
and confirming that the activating quantities of the output are within its designed
parameters/ratings. So it goes beyond the traditional fire and electric shock safety associated with
Table -1 MIS Server Platform Region and EMC Compliance References
Region Compliance Reference
Australia/
New Zealand
AS/NZS 3548 (Emissions)
Canada/USA CSA 60950 / UL60950/
60950-1 cert to CAN/CSA STD C22.2 No. 60950-1
Industry Canada ICES-003
FCC CFR47, Part 15
CENELEC Europe EN55022 Emissions
EN55024 Immunity
International CISPR 22/ CISPR 24
Japan VCCI Certification
Korea KCC Certification
Taiwan BSMI CNS 13438
China CCC
Russia GOST
Introduction
007-5818-003 xxiii
the traditional UL Listing Mark. Table -2 lists the region and compliance reference for EMC
(Electromagnetic Compatibility) compliance.
Chapter Descriptions
Chapter 1, “System Overview‚” describes the hardware components of the MIS enclosures, the
common modules in unit, and the major differences between the MIS Server Platform and MIS
JBOD Unit. Additional information includes the operating systems supported, and RAID
configurations possible with the MIS enclosures.
Chapter 2, “System Interfaces‚” describes the hardware and software interfaces used to operate
the MIS Server and MIS JBOD. This includes the front control panel, disk drive LEDs, power
supply LEDs, and the BMC Web Console.
Chapter 3, “System Software‚” covers the software used on the MIS Platforms, including
installation information for the tools, using the MegaRAID tool, and the available zoning software
(Zones and CLI Zoning Tool). Depending on the operating system, there are certain prerequisite
programs and this chapter gives instructions for download and installation of these programs.In
this chapter are step-by-step instructions for Zones tool, its features and their function, plus
warnings and error codes. Screen shots are given for a walk-through of the tool. detailed in this
chapter next are step-by-step instructions for zoning using the CLI Zoning tool (the only tool that
can zone JBODs at this time).
Chapter 4, “System Maintenance‚” describes how to use Sensor Data Records for detecting
component failures, and service instructions for modules that are customer replaceable units
(CRUs). The service instructions include how to move the chassis forward or backwards in the
rack, how to remove the case front and read covers, how to remove the midspan support bar for
ease of access to cabling, how to replace a power supply, how to replace a storage drive in a
Table -2 MIS Server Platform Region and EMC Compliance References
Region Compliance Reference
Canada/USA CSA 60950 / UL60950/
Compliance Document UL report
60950-1 cert to CAN/CSA STD C22.2 No. 60950-1
IEC (Europe) IEC60950-1 – CB Certification, CE Mark
Compliance Document UL report
Russia GOST
xxiv 007-5818-003
Introduction
StorBrick, how to replace a boot drive, how to replace a fan module, and additional air-flow
precautions.
Chapter 5, “Troubleshooting‚” describes some problem-solving techniques, plus when and how
to contact customer support.
Appendix A, “Technical Specifications,” gives the technical specifications for the MIS
enclosures.
Appendix B, “BIOS Error Codes,” details the beep codes used when a problem is detected by the
BMC environmental controls.
Related Publications
The following documents are relevant to the MIS Platform:
MegaRAID SAS Software User Guide, publication number 51530-00, Rev E.
MegaRAID 6Gb/s SAS RAID Controllers User Guide, publication number 41450-02, Rev E.
Intel Server Boards and Server Platforms Server Management Guide,
publication number 37830-002
SGI Foundation Software, publication number 007-5641-00x
SGI Performance Suite, publication number 007-5680-00x
SGI InfiniteStorage series documentation (http://techpubs.sgi.com)
Man pages (http://www.linuxmanpages.com/)
Various formats of SGI documentation, release notes, and man pages are available. The SGI
Technical Publications Library (http://docs.sgi.com/) contains the most recent and most
comprehensive set of online books, release notes, man pages, and other information. Refer to the
SGI Supportfolio™ web page for documents which access requires a support contract (as do the
MegaRAID books cited above). See “Product Support” on page xxv. You can also view man
pages by typing man <title> on a command line in Linux.
Introduction
007-5818-003 xxv
Conventions
The following conventions are used throughout this document:
Product Support
SGI provides a comprehensive product support and maintenance program for its products, as
follows:
If you are in North America, contact the Technical Assistance Center at +1 800 800
4SGI (4744) or contact your authorized service provider.
If you are outside North America, contact the SGI subsidiary or authorized distributor in
your country. International customers can visit http://www.sgi.com/support/ Click on
the “Support Centers” link under the “Online Support” heading for information on how
to contact your nearest SGI customer support center.
CRU/FRU
Some of the components on the MIS Platform are customer-replaceable units (CRUs), meaning
that these modules were designed to be repaired/replaced by you, the customer. These include fan
assemblies, power supplies, storage drives, and boot drives, all of which are hot-swappable.
However, many of the other components on the MIS Platform should be serviced by SGI field
technicians ONLY, so as not to violate the warranty agreement. The components are
field-technician replaceable units, or FRUs. It is important to note that our CRUs can be easily
installed and replaced by customers, which enables a speedy recovery of proper system operation.
Convention Meaning
Command This fixed-space font denotes literal items such as commands, files,
routines, path names, signals, messages, and programming language
structures.
variable The italic typeface denotes variable entries and words or concepts being
defined. Italic typeface is also used for book titles.
[ ] Brackets enclose optional portions of a command or directive line.
GUI element This font denotes the names of graphical user interface (GUI) elements such
as windows, screens, dialog boxes, menus, toolbars, icons, buttons, boxes,
fields, and lists.
xxvi 007-5818-003
Introduction
For additional information about CRUs, please see:
Customer Replaceable Units (CRUs) Installation Policy
Customer Replaceable Units (CRU) and Customer Obligations\
Purchasable Support & Maintenance Programs
SGI provides several comprehensive product support and maintenance programs for its products.
SGI also offers services to implement and integrate Linux applications in your environment.
Refer to http://www.sgi.com/services/
If you are in North America, contact the Technical Assistance Center at
+1-800-800-4SGI (4744), or contact your authorized service provider.
If you are outside North America, contact the SGI subsidiary or authorized distributor in
your country. See http://www.sgi.com/global/index.html for more information.
Reader Comments
If you have comments about the technical accuracy, content, or organization of this document,
please contact SGI. Be sure to include the title and document number of the manual with your
comments. (Online, the document number is located in the front matter of the manual. In printed
manuals, the document number is located at the bottom of each page.)
You can contact SGI in any of the following ways:
Send e-mail to the following address: techpubs@sgi.com
Contact your customer service representative, and ask that an incident be filed in the SGI
incident tracking system.
Send mail to the following address:
SGI
Technical Publications
46600 Landing Parkway
Fremont, CA 94538
SGI values your comments, and will respond to them promptly.
007-5818-003 1
Chapter 1
1. System Overview
The SGI Modular InfiniteStorage Platform is a high-density, integrated storage server platform.
The MIS Platform uses a 4U rackmount system, and can be either a compute and storage server,
or a “Just Bunch Of Disks” expansion storage unit (MIS JBOD unit). The MIS Server Platform
can be single or dual server. Up to 5 MIS enclosures (server & JBODs) or 6 JBODs can be
mounted into a SGI Destination rack (D-Rack), as shown in Figure 1-1. (Other 3rd-party 19" racks
are also supported.) A D-Rack has space for up to 10 enclosures, however, due to floor weight
regulations, only 5-6 units may be installed in a single D-Rack. See
Features of the modular design of the MIS Platform include:
Up to 72 (3.5" or 2.5" 15mm) and a maximum of 144 (2.5" 9.5mm) storage drives in the
Server Platform
Up to 81 (3.5" or 2.5" 15mm) and a maximum of 162 (2.5" 9.5mm) storage drives in the
JBOD unit
All fit in a standard size 4U chassis: height 6.94" (176mm), width 16.9" (429.2mm), depth
36" (914.4mm).
Storage drives can be 3.5" or 2.5" (15mm or 9.5mm), SAS or SATA, rotational or SSD drives. Up
to four JBOD units can be attached to one MIS Dual Server Platform.
Warning: Rotational SAS drives and rotational SATA drives cannot be included in the
same enclosure due to vibration conflicts.
The MIS Server Platform features:
Up to 2 server modules per platform.
One or two Intel® Xeon® E5-2600 series processors per server motherboard.
!
2007-5818-003
1: System Overview
Intel Turbo Boost Technology 2.0 that automatically allows processor cores to run faster
than the base operating frequency, if the cores are operating below power, current, and
temperature specification limits (< 35ºC ambient).
Up to 8 DDR3 DIMMs (4 GB, 8 GB, or 16 GB) for a single-server motherboard
configuration, and up to 16 DIMMs for a dual-server motherboard configuration.
Up to 4 HBAs for a single server, full-height (4.25") and half-depth (3.375"), externally or
internally facing. Up to 4 HBAs (half-height, half-depth; 2 per server module) for a dual
server. There are an additional two internally facing, half-height and half-depth HBAs per
server module, used by the system. The MIS Single Server Platform can have a total of six
HBAs, where a Dual Server Platform can have a maximum of eight (including those used by
the system).
Up to 3 PCIe riser cards for single server systems (dual-servers have a mandatory 3 PCIe
risers, regardless of card count).
Up to four battery back up units for a single server module. Up to three battery back up units
per server module for a dual server platform, for a total maximum of six. (Unique BBU PCIe
technology allows the inclusion of BBUs without the consumption of any of the available
PCIe slots.)
Two boot drives per server: SAS or SATA, rotational or SSD, up to 300GB, mirrored using
RAID 1.
Dual GbE networking onboard, with an optional 2 port 10GbE, 2 port GbE, or 4 port 8Gb
FC PCIe cards (4 optional networking PCIe cards maximum, external facing only; risers 1
and 2).
Warning: Floor loading has a maximum weight allowance of 250lbs per square foot, not
including the service area. Floor loading must be less than 250lbs per square foot, including
the service area. There can be a total of 6 JBODs or 6 MIS Servers per D-Rack, or a
combined total of 5. For maximum efficiency and performance, it is suggested that the
maximum number of enclosures in a single D-Rack is 1 MIS Dual Server enclosure with 4
JBODs (two JBODs per server module). A 5th JBOD can be tolerated weight-wise, but 4 is
the suggested performance maximum.
The System Overview will first explore the “MIS Enclosure,” including “Front Grille and Control
Panels” on page 6, “Rear Panel Components” on page 7. Next, the “MIS Common Modules” on
page 10 is covered, including the power supply modules, fan modules, and StorBrick modules.
“MIS Server Platform or JBOD Unit” on page 14, discusses the presence of the “Server Module”
on page 14, its available features and associated “Boot Drives Module” on page 22, or the
!
007-5818-003 3
presence of a ninth StorBrick and “MIS JBOD I/O Module” on page 22. Finally, a “System Block
Diagram” on page 23 is given, showing a diagram of the enclosure with a dual-server
configuration (optional JBOD components show in grey), signals and power connections.
4007-5818-003
1: System Overview
Figure 1-1 SGI Destination Rack (D-Rack)
MIS Enclosure
007-5818-003 5
MIS Enclosure
The MIS enclosure, whether it is a server or JBOD, consists of a chassis, case (with front bezel
grille, control panel and rear ports).
Figure 1-2 MIS Chassis and Case
The SGI MIS chassis features a front bezel display with an internal EMI grille (Figure 1-2). The
unique bi-directional sliding rail mounts (Figure 1-3) allow the unit to be slid forwards 20" or
backwards 18" to access disk drives and other serviceable components. This also makes for an
overall safer rack design, as chassis do not need to be extended their full length to be serviced.
6007-5818-003
1: System Overview
Figure 1-3 Bi-directional rail mount
Front Grille and Control Panels
Next to the bezel grille, up to two control panels can be present on the MIS Platform, one for each
server in the MIS Server Platform, or one for each I/O unit on the MIS JBOD. Figure 1-4 shows
a single control panel and Figure 1-5 shows two control panels.
Each control panel has a Power LED, Power button, Status LED, Reset Button, Locator LED,
Locator button, Network Activity LED, Boot Drive Activity LED, and NMI Reset button (to be
used by SGI field operatives only). Indicator light meanings and button functions are explained
in, “Control Panel” in Chapter 2.
MIS Enclosure
007-5818-003 7
Figure 1-4 Single Control Panel
Figure 1-5 Dual Control Panel
Rear Panel Components
The appearance of the rear panel on the MIS chassis will depend on what modules are installed.
An MIS Platform can have up to four power supplies, each with their own AC power inputs (only
two are pictured in any of the figures here). They are high-efficiency, hot-swappable power
supplies rated at 1100 Watts.
All rear panels feature clearly silk-screened labels next to the port in question. The MIS Server
Platform (single server) in Figure 1-6, features a single server module with two USB ports, a video
port, and two NIC ports. Figure 1-7 show a MIS Server Platform with the optional four power
supplies. The MIS Server Platform (dual server) rear panel as shown in Figure 1-8 has a second
server module with its own set of USB ports, video port, and NIC ports. Figure 1-9 shows an MIS
Server Platform (single dual-server module) which features dual server construction with a single
server installed, and the option of upgrade to include a second server later. Figure 1-11 shows the
rear panel of the MIS JBOD with two I/O modules.
8007-5818-003
1: System Overview
Figure 1-6 Rear View – MIS Server Platform (single server)
Figure 1-7 Rear View – MIS Server Platform (single server, four power supplies)
Figure 1-8 Rear View – MIS Server Platform (dual server)
MIS Enclosure
007-5818-003 9
Figure 1-9 Rear View – MIS Server Platform (single dual-server module)
Figure 1-10 Rear View – MIS JBOD Unit (single I/O module)
Figure 1-11 Rear View – MIS JBOD unit (dual I/O modules)
10 007-5818-003
1: System Overview
MIS Common Modules
This section describes the common internal modules of the MIS enclosure. Designed to deliver a
high level of reliability, scalability and manageability, the MIS platform makes use of modules to
contain key components. Whether the unit is an MIS Server or a JBOD, both chassis contain the
following hot-swappable modules:
Up to four power supplies (two redundant) (Figure 1-12);
Six fan assemblies (Figure 1-14);
Capacity drives installed in StorBricks (Figure 1-15).
The power supply modules are high-efficiency, hot-swappable power supplies rated at 1100
Watts, AC Input: 100–240 VAC (50-60Hz), single or three phase. There are six hot-swappable fan
modules housing one fan with two counter-rotating impellers. And instead of the conventional
disk architecture, the unique StorBrick modules—innovative, highly-dense drive modules used to
house drive bays—allows the platform to maximize storage density.
Each MIS Server has eight StorBricks modules, and each MIS JBOD has nine, with the ninth
module taking the place of the compute server module. Each StorBrick module holds up to nine
3.5" or 2.5" (15mm), SAS or SATA, rotational or SSD drives, or, using the dual-slot drive option,
eighteen 2.5" (9.5mm), SAS or SATA, rotational or SSD drives.
Warning: Rotational SAS drives and rotational SATA drives cannot be included in the
same inclosure due to vibration conflicts.
Power Supply Module
One to four power supplies provide power for the SGI MIS server. Power supplies are configured
for N+N support. The power supplies provide 12VDC main power and 5VDC standby power. The
power supplies are hot-swappable and can be replaced under full load. Power supplies are
numbered 0-3 from the bottom up, on the rear panel of the enclosure (Figure 1-13).
!
MIS Common Modules
007-5818-003 11
Figure 1-12 Power Supply Module (rated at 1100 Watts)
Figure 1-13 Power Supply Numbering
Fan Assembly Module
Figure 1-14 Fan Assembly Module (each contains two impellers)
12 007-5818-003
1: System Overview
Six fan assemblies mounted in the middle of the chassis cool the system. Each hot-swappable fan
assembly contains two impellers. Air flows from the front to the back of the enclosure. The fan
baseboard distributes power and control signals to the fan assemblies. Firmware on the fan
baseboard monitors the fan speeds and temperatures within the enclosure. The SMS adjusts the
individual fan speeds as needed to continuously provide optimal cooling for the enclosure.
StorBrick Module
Each StorBrick Module contains up to nine 3.5" or 2.5" 15mm drives (Figure 1-15), or eighteen
2.5" 9mm drives (), mounted in the StorBrick using proprietary drive carriers (Figure 1-16). A
sliding thumb latch securely fastens the drive carriers in place (Figure 1-17, thumb latch is
pictured in blue, but is grey on the actual product). StorBricks use SAS-2 protocol, which enables
the system to use SAS and/or SATA drives (rotational disks or SSDs).
Figure 1-15 StorBrick Modules for 3.5" or 2.5" 15mm Drives (left)
and 2.5" 9.5 mm Drives (right)
MIS Common Modules
007-5818-003 13
Warning: Rotational SAS drives and rotational SATA drives cannot be included in the
same inclosure due to vibration conflicts.
Figure 1-16 3.5" 15mm Drive and Carrier
Figure 1-17 3.5" 15mm Drive Carrier (top view, with thumb latch)
Figure 1-18 Two 2.5" 9.5mm Drives and Carrier
!
14 007-5818-003
1: System Overview
Figure 1-19 2.5" 9.5mm Drive Carrier (isometric view with dual thumb latches)
MIS Server Platform or JBOD Unit
The key difference between the MIS Server Platform (Figure 1-20 or Figure 1-21) and the MIS
JBOD Unit (Figure 1-22) is the presence of the compute server module (Figure 1-23 or
Figure 1-25) and boot drives (Figure 1-29) in the Server Platform, or a ninth StorBrick
(Figure 1-15) and I/O modules (Figure 1-30) and associated midplane (Figure 1-31) in the JBOD.
Server Module
The MIS Server Platform can be single- or dual-server (Figure 1-23 or Figure 1-25) depending on
whether it has one or two compute server modules. Each compute server module can have:
Up to two Intel® Xeon® E5-2600 series processors per motherboard, with Intel Turbo Boost
Technology 2.0: if the cores are operating below power, current, and temperature specs (<
35°C ambient) limits, they automatically run faster than base operating speed.
8 DDR3 DIMMs (4 GB, 8 GB, or 16 GB) for a single-server board configuration. Up to 16
DIMMs for a dual-server board configuration,
Up to 4 HBAs for a single server, full-height (4.25") and half-depth (3.375"), externally or
internally facing. Up to 4 HBAs (half-height, half-depth; 2 per server module) for a dual
server. (See Figure 1-27 on page 20)
Up to three PCIe riser cards for a single server (dual servers have a mandatory 3 PCIe risers).
Up to four battery back up units for a single server module. Up to three battery back up units
per server module for a dual server platform, for a total maximum of six. (Unique BBU PCIe
technology allows the inclusion of BBUs without the consumption of any of the available
PCIe slots.)
MIS Server Platform or JBOD Unit
007-5818-003 15
Figure 1-20 MIS Server Platform (single server)
16 007-5818-003
1: System Overview
Figure 1-21 MIS Server Platform (dual server)
MIS Server Platform or JBOD Unit
007-5818-003 17
Figure 1-22 MIS JBOD Unit
18 007-5818-003
1: System Overview
Figure 1-23 MIS Server Module – single server
Figure 1-24 Server Module (single server) – component view
MIS Server Platform or JBOD Unit
007-5818-003 19
Figure 1-25 MIS Server Module – dual server (half height)
Figure 1-26 Dual Server Module – component view
20 007-5818-003
1: System Overview
Layout of Server CPUs, and PCIe Risers HBAs
Figure 1-27 shows the CPU and riser layout. Because of cabling restrictions in a single server,
single CPU systems, only two internal SAS HBAs are allowed. The first CPU handles Riser 1
and 2. The optional second CPU would manage Riser 3. If the second CPU is not installed, Riser 3
is non-operational.When a second CPU is installed, HBAs populated on Riser 3 are internal facing
SAS HBAs only, which connect to the StorBricks (Figure 1-28).
Figure 1-27 CPU and PCIe Riser layout
MIS Server Platform or JBOD Unit
007-5818-003 21
Figure 1-28 HBA population layout
Boot Drives Module
Each MIS Server Platform features two boot drives per server module (up to four total – mirrored
using LSI software RAID 1). These drives are SAS or SATA, rotational or SSD, up to 300GB,
used to store server data and the server operating system. Supported operating systems include:
Microsoft® Windows® 2008 R2 SP1 (not shipped with product),
Red Hat® Enterprise Linux (RHEL) 6.2,
SUSE LINUX® Enterprise Server 11 SP1, or
•VMware® ESX 5.0
22 007-5818-003
1: System Overview
Figure 1-29 Boot Drive Module
MIS JBOD I/O Module
Figure 1-30 I/O Module for MIS JBOD Unit
JBOD I/O modules slide into a midplane (Figure 1-31), which connect to the SAS controllers.
System Block Diagram
007-5818-003 23
Figure 1-31 MIS JBOD Midplane I/O Connector (right & left views)
System Block Diagram
Figure 1-32 shows the system-level block diagram for a fully populated dual-server (the optional
JBOD components are shown in grey: “STORBRICK 8” and “JBOD only”).
24 007-5818-003
1: System Overview
Figure 1-32 System-Level Block Diagram
18
44
X36
SAS
EXP
012
345
678
STORBRICK
7
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
STORBRICK
0
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
STORBRICK
1
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
STORBRICK
2
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
STORBRICK
3
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
STORBRICK
4
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
5V
POL PWR
CTL
18
44
X36
SAS
EXP
012
345
678
5V
POL PWR
CTL
STORBRICK
5STORBRICK
6STORBRICK
8
FAN0A
FAN0B
FAN1A
FAN1B
FAN2A
FAN2B
FAN3A
FAN3B
FAN4A
FAN4B
FAN5A
FAN5B
PS0PS1PS2PS3
CM0
IO0
CM1
IO1
PDU
FANB
SAS
Side
IO
SAS
Side
IO
JBOD
only
SAS signals
Power
Control signals
HOST
Side
IO
HOST
Side
IO
CM1
Boot
CM0
Boot
Control Panel
007-5818-003 25
Chapter 2
2. System Interfaces
This chapter describes the hardware and software interfaces of the MIS platforms. Both the MIS
server platform and MIS JBOD storage unit have a front control panel, disk drive LED codes, and
power supply LED codes. The control panel lights and buttons have different meanings and
functions, depending on whether the machine is the MIS Server Platform or MIS JBOD unit. The
disk drive LED codes and power supply LED codes remain the same whether the system is a
server platform or JBOD unit. Additionally, there are four programs used to initialize and monitor
the MIS machines. This chapter details the hardware interfaces, their functions and indications, as
well as the Baseboard Management Controllers (BMC) Web Console. These programs provide
power management features, environmental monitoring, etc.
Note: SGI provides features beyond those of IMPI 2.0, for instance, chassis intrusion detection,
which will gracefully power down the system if the case cover is left off for more than 15 minutes.
Control Panel
MIS Server Control Panel
The control panel (Figure 2-1) interface consists of five indicator lights and four buttons. More
information on remote functionality and the Web Console and Terminal Tool is presented at the
end of this chapter.
26 007-5818-003
2: System Interfaces
Figure 2-1 MIS Control Panel
The NMI Reset button (Non-Maskable Interrupt) should not be used, except under the direct
supervision from technical support.
Table 2-1 MIS Server Platform Control Panel Buttons and LEDs
LED/Button Description
Power LED Green LED lit means power is on.
Power button If the system is off, pushing this button powers on the system. If the
operating system is running, pushing this button shuts down the
operating system and powers off the system gracefully. If the operating
system is hung, holding the button down for 10 seconds or more will
power off the system for a hard reset.
Control Panel
007-5818-003 27
MIS JBOD Control Panel
The control panel (Figure 2-1) for the MIS JBOD is exactly the same as the MIS Server Platform.
However, some of the buttons do not have the same function as they do on the MIS Server. Since
there is no boot drive module in a JBOD, the Boot Drive Activity LED, located next to the
Network Activity LED, is present, but inactive.
Important: When there are two I/O modules on a JBOD, the top control panel connects to the
bottom I/O module on the back of the unit, and vice versa, the bottom control panel accesses the
top I/O module.
Status LED This indicator will be lit whenever there is AC power available to the
power supplies, whether the unit is on or off. Green means the system is
in good working order. Yellow indicates a problem with the system, and
service is required.
Reset button The service reset button. When pushed, this button causes the server to
reboot and, if the problem is cleared by the reset, returns the Status LED
to green.
Locator LED Blue LED is lit on the front and the back to help locate the unit in a rack
or bay.
Locator button The Locator LED will be lit blue when the Locator button is pushed.
There is a corresponding LED on the back of the server that will be blue.
When the Locator button is pushed again, the LED will go off. This
function may also be activated remotely using the Intel BMC Web
Console and pressing the virtual button, or using the Linux IPMI
Terminal Tool:
-H <ip address> -P <password> -U <user> chassis identification
NIC Activity LED The green LED will be active whenever there is any network traffic
occurring on the base board NIC ports.
Boot Drive
Activity LED
The LED is lit whenever the boot drives are being accessed.
NMI Reset button Used only under the direction of technical support personnel.
Table 2-1 MIS Server Platform Control Panel Buttons and LEDs (continued)
LED/Button Description
28 007-5818-003
2: System Interfaces
Disk Drive LEDs
Figure 2-2 shows the green/yellow and blue disk drive LEDs.
Figure 2-2 Disk Drive LEDs
Table 2-2 describes the meaning of disk drive LEDs.
Power Supply LEDs
There are two LEDs located on the face plate of the power supply, one green on top, and one
bi-color yellow/green below (Figure 2-3). Table 2-3 describes the function of the power supply
LEDs
Figure 2-3 Power Supply LEDs
Table 2-2 Disk Drive LEDs
Bi-color LED Blue LED Drive Status
Off Off Drive is off and can be removed.
Green Off Drive is on.
Yellow Off Service required.
Off/Green/Yellow On Indicates drive location.
BMC Integrated Web Console
007-5818-003 29
BMC Integrated Web Console
The control panel and various other LEDs are used to monitor the overall status of the system and
its components. Underlying the light-guided diagnostics provided through the various LEDs on
the control panel, power supplies, motherboard, etc. are the BMC/IPMI interfaces. The MIS server
supports the platform management features (environmental monitoring, power management, etc.)
provided by the Intel BMCs and IPMI 2.0. Moreover, the BMCs have features beyond those of
IPMI 2.0 (for instance, detection of chassis intrusion).
The BMC Integrated Web Console is a web-based program provided by Intel, and is used to give
general system information such as system diagnostics, server health, environmental reporting,
and event logs. Additionally, the BMC-IWC provides a remote virtual control panel for the MIS
Server, allowing for remote locating and reboot.
For more information, see the platform management documentation for the Intel S2600JF
motherboard in Intel Server Boards and Server Platforms Server Management Guide (publication
number 37830-002)
This section gives you a high-level description of each Integrated BMC Web Console page. It is
organized in sections corresponding to the four tabs in the horizontal menu. Within each section,
each menu on the left-hand side is illustrated and described in detail.
Table 2-3 Power Supply LEDs
Green LED Bi-color LED Power Supply Status
Off Off No AC power to the supply, power is off on the front
of the machine.
Off Yellow Problem indicated (voltage, fan failure, AC failure,
etc.)
On Blinking Yellow AC available, power supply in standby mode
(powered off on the front)
On Green AC available to the power supply, power supply is
on and functioning normally.
30 007-5818-003
2: System Interfaces
System Information
After login, by default, the BMC Web Console opens on the System Information page. The
System information page displays a summary of the general system information. This includes the
power status and the version of firmware, and has the following information about the server.
In the left navigation pane, There are five menu options. After System Information comes FRU
Information, System Debug Log, CPU Information, and DIMM information.
Table 2-4 System Information Details
Information Details
Host Power Status Shows the power status of the host (on/off).
RMM Status Indicates if the Intel RMM4 card is present.
Device (BMC) Available Indicates if the BMC is available for normal
management tasks.
BMC FW Build Time The date and time of the installed BMC firmware.
BMC FW Rev Major and minor revision of the BMC firmware.
Boot FW Rev Major and minor revision of the BOOT firmware.
SDR Package Version Version of the Sensor Data Record.
Mgmt Engine (ME) FW Rev Major and minor revision of the Management
Engine firmware.
Overall System Health Green/Yellow and Blue
BMC Integrated Web Console
007-5818-003 31
Figure 2-4 BMC Web Console – System Information Page
FRU Information
The Field Replaceable Unit (FRU) Information page displays information from the FRU
repository of the host system.
FRU Chassis Information includes: Type, Part/Model Number, and Serial Number. FRU Board
Information includes: Manufacturing Date, Manufacturer, Product Name, Serial Number,
Part/Model Number, and FRU File ID. FRU Product Information includes: Manufacturer,
Name, Part/Model Number, Version, Serial Number, Asset Tag, and FRU File ID.
32 007-5818-003
2: System Interfaces
Figure 2-5 BMC Web Console – FRU Information
System Debug Log
The System Debug Log page allows administrators to collect system debug information. This
feature allows a user to export data into a file that is retrievable for the purpose of sending to an
Intel engineer or Intel partners for enhanced debugging capability. Select either the “System
Debug Log” or the “System & BMC Debug Log” and press the Run button. It may take some time
for the debug information to be collected.
The files are compressed, encrypted, and password protected. The file is not meant to be viewable
by the end user but rather to provide additional debugging capability to your system manufacturer
or an Intel support engineer. Once the debug log dump is finished you can click the debug log
filename to save the results as a .zip file on your client system. The file can then be sent to your
system manufacturer or an Intel support engineer for analysis.
System Debug Log Type
The System Debug Log data is mainly used by the system manufacturer for analysis. Baseboard
Management Controller (BMC) status, BMC configuration settings, BMC Sensor readings, Power
supply data, System Event Log, sensor readings, SMBIOS tables, CPU machine check registers
BMC Integrated Web Console
007-5818-003 33
and PCI configuration space information. The System & BMC Debug Log contains regular
System Debug Log plus the BMC debug log.
Last Log
Shows the time of the last data collection. Collection times older than three minutes will be
marked as an “Old” debug log.
Encryption
The resulting zip file will be encrypted for privacy, and may only be extracted for analysis by an
authorized representative of the system manufacturer.
Generate Log
Click the Generate Log button to collect recent Debug Log data. The resulting compressed archive
will be downloaded to your system by clicking on the debug log link. You may also choose to
download the data at a later time using the debug log link. Note that it is recommended that fresh
data always be downloaded for analysis.
Figure 2-6 BMC Web Console – System Debug Log
34 007-5818-003
2: System Interfaces
CPU Information
The CPU Information page displays information on the processor(s) installed in the server.
The data in the CPU Asset Information page is collected from SMBIOS entries sent from the BIOS
to the Baseboard Management Controller at the end of POST. If there is no data available, or the
data is stale, please reset the system, allow it to complete POST, and refresh the page using the
refresh button above.
Figure 2-7 BMC Web Console – CPU Information
DIMM Information
The DIMM Information page displays information on DIMM modules installed on the host
system.
Slot Number is DIMM location on the motherboard. The location is marked as A0, B1 and so on.
BMC Integrated Web Console
007-5818-003 35
Figure 2-8 BMC Web Console – DIMM Information
Server Health
The Server Health tab shows you data related to the server's health, such as sensor readings, the
event log, and power statistics as explained in the following sub sections. Click on the Server
Health tab to select the various pages. By default, this tab opens the Sensor Readings page.
Sensor Readings
The Sensor Readings page displays system sensor information including status, health, and
reading. By default the sensor readings are updated every 60 seconds but this can be changed by
entering a value in the Set auto-refresh in seconds selection box and then pressing the Set button.
Sensor Selection drop-down box allows you to select the type of sensor readings to display in the
list. The default is set to All Sensors, with other options: Temperature Sensors, Voltage Sensors,
Fan Sensors, Physical Security, Processor, Power Unit, Memory, Event Logging Disable, System
Event, Button/Switch, Module/Board, Watchdog Sensor, Management Subsystem Health, Node
Manager, and SMI.
36 007-5818-003
2: System Interfaces
Figure 2-9 BMC Web Console – Server Health
Click Show Thresholds to expand the list, showing low and high threshold assignments. Use
scroll bar at bottom to move display left and right.
CT: Critical threshold
NC: Non-critical threshold
Click Hide Thresholds to return to original display, hiding the threshold values, showing only the
name, status and reading for selected sensors. Click Refresh to refresh the selected sensor
readings.
Event Log
The Event Log is a table of the events from the system's event log. You can choose a category
from the pull-down box to filter the events, and also sort them by clicking on a column header.
BMC Integrated Web Console
007-5818-003 37
The filters available are All Events, Sensor-Specific Event, BIOS Generated events, and System
Management Software Events. Use this page to view and save the Event log. Event Log Category
selects the type of events to display in the list. Event Log List is a list of the events with their ID,
time stamp, sensor name, sensor type, and description. Click Clear Event Log to clear the event
logs. Click on Save Event Log to download the event logs to local system.
Figure 2-10 BMC Web Console – Event Log
Power Statistics
This section shows you data related to the server's health, such as sensor readings and the event
log. Use this page to determine server power usage. In order to collect readings for this page, a
PMBus enabled power supply must be attached to the server, and the server must be in a ACPI S0
(DC power on) state to report statistics data. System Power Statistics include:
Minimum: Calculated as a minimum value of all power readings since last statistics reset.
38 007-5818-003
2: System Interfaces
Current: Present power reading.
Maximum: Calculated as a maximum value of all power readings since last statistics reset.
Average: Calculated as the arithmetic average of all power readings since last statistics reset.
Figure 2-11 BMC Web Console – Power Statistics
Configuration Tab
The Configuration tab of the BMC Web Console is used to configure various settings, such as
alerts, users, or network. It contains the following menu options in the left navigation pane: IPv4
Network, IPv6 Network, Users, Login, LDAP, VLAN, SSL, Remote Session, Mouse Mode,
Keyboard Macros, Alerts, Alert Email, Node Manager.
BMC Integrated Web Console
007-5818-003 39
IPv4 Network
Use this page to configure the network settings for server management.
Enable LAN Failover
Enabling failover bonds all available ethernet interfaces into the first LAN Channel. When the
primary interfaces lease is lost one of the secondary interfaces is activated automatically with the
same IP address.
LAN Channel Number
It lists the LAN Channel(s) available for server management. The LAN channels describe the
physical NIC connection on the server. Intel® RMM channel is the add-in RMM NIC. The
Baseboard Management channel is the onboard, shared NIC configured for management and
shared with the operating system.
MAC Address
The MAC address of the device (read only).
IP Address
Select the type of IP assignment with the radio buttons. If configuring a static IP, enter the
requested IP Address, Subnet Mask, Gateway, Primary DNS and Secondary DNS Server in the
given fields.
IP Address made of 4 numbers separated by dots as in xxx.xxx.xxx.xxx.
•‘xxx’ ranges from 0 to 255.
•Firstxxx’ must not be 0.
Caution: The RMM IP address must be on a different subnet than the baseboard IP address used
for management traffic.
40 007-5818-003
2: System Interfaces
Figure 2-12 BMC Web Console – IPv4 Network Settings
IPv6 Network
Use this page to configure the IPv6 network settings for server management.
Enable LAN Failover
Enabling failover bonds all available ethernet interfaces into the first LAN Channel. When the
primary interfaces lease is lost one of the secondary interfaces is activated automatically with the
same IP address.
BMC Integrated Web Console
007-5818-003 41
LAN Channel Number
It lists the LAN Channel(s) available for server management. The LAN channels describe the
physical NIC connection on the server. Intel® RMM channel is the add-in RMM NIC. The
Baseboard Management channel is the onboard, shared NIC configured for management and
shared with the operating system.
MAC Address
The MAC address of the device (read only).
Enable IPv6 on this Channel
This check box must be selected to enable any IPv6 network traffic on this channel.
IP Address configuration
IPv6 auto-configuration enables Stateless configuration using ICMPv6 router/neighbor
discovery.
Obtain an IP address automatically enables DHCPv6.
Use the following IP address enables static IP assignment.
IP Address
Select the type of IP assignment with the radio buttons. If configuring a static IP, enter the
requested IP Address, IPv6 prefix length, and optionally the Gateway in the given fields.
IPv6 addresses consist of 8 4 digit hexadecimal numbers separated by colons.
A :: can be used for a single sequence of two or more zero fields.
Caution: The RMM IP address must be on a different subnet than the baseboard IP address used
for management traffic.
42 007-5818-003
2: System Interfaces
Figure 2-13 BMC Web Console – IPv6 Network Settings
Users
The list of configured users, along with their status and network privilege is displayed. Use this
page to configure the IPMI users and privileges for this server.
Add User – Select an empty slot in the list and click to add a new user.
Modify User – Select a user in the list and click to modify their settings.
Delete User – Select a user in the list and click to delete.
Note: UserID 1 (anonymous) may not be renamed or deleted. UserID 2 (root) may not be renamed
or deleted; nor can the network privileges of UserID 2 be changed.
BMC Integrated Web Console
007-5818-003 43
Caution: User Names cannot be changed. To rename a User you must first delete the existing
User, then add the User with the new name.
Login
Use this page to configure login security settings for server management.
Failed Login Attempts
Set the number of failed login attempts a user is allowed before being locked out. Zero means no
lockout. Default is 3 attempts.
44 007-5818-003
2: System Interfaces
User Lockout Time (min)
Set the time in minutes that the user is locked out before being allowed to login again. Zero means
no lockout and unlocks all currently locked out users. Default is 1 min.
Force HTTPS
Enable this option to force the User to user Secure Login for Web using HTTPS protocol. It will
use the certificate uploaded under 'Configuration->SSL'.
Web Session Timeout
Set the maximum web service timeout in seconds. Timeout should be between 60 and 10800
seconds. Default timeout is 1800 seconds.
Figure 2-14 BMC Web Console – Login Security Settings
BMC Integrated Web Console
007-5818-003 45
LDAP
To enable/disable LDAP, check or uncheck the Enable LDAP Authentication checkbox
respectively.
LDAP Authentication
Check this box to enable LDAP authentication, then enter the required information to access the
LDAP server.
Port
Specify the LDAP Port.
IP Address
The IP address of LDAP server
IP Address made of 4 numbers separated by dots as in xxx.xxx.xxx.xxx.
•‘xxx’ ranges from 0 to 255.
•Firstxxx’ must not be 0.
Bind Password
Authentication password for LDAP server; the password must be at least 4 characters long.
Bind DN
The Distinguished Name of the LDAP server, e.g. cn=Manager, dc=my-domain,
dc=com.
Searchbase
The searchbase of the LDAP server, e.g. dc=my-domain, dc=com.
46 007-5818-003
2: System Interfaces
Figure 2-15 BMC Web Console – LDAP Settings
VLAN
Use this page to configure an 802.1Q VLAN private network on the specified LAN channel.
LAN Channel
It lists the LAN Channel(s) available for server management. The LAN channels describe the
physical NIC connection on the server. Intel® RMM channel is the add-in RMM NIC. The
Baseboard Mgmt channel is the onboard, shared NIC configured for management and shared with
the operating system.
BMC Integrated Web Console
007-5818-003 47
Enable VLAN
Check to enable VLAN on this channel. When enabled, the BMC only accepts packets with the
correct VLAN Identifier field. All outgoing packets are marked with that VLAN ID.
VLAN ID
Specify the VLAN ID to use. Values are from 1 to 4094. Only one ID can be used at a time, and
VLAN must first be disabled before a new ID can be configured on a given LAN channel.
VLAN Priority
Specify the VLAN Priority field to place in outgoing packets. Values are from 0 (best effort) to 7
(highest); 1 represents the lowest priority. 0 is the default.
Figure 2-16 BMC Web Console – VLAN Settings
48 007-5818-003
2: System Interfaces
SSL
The SSL Upload page shows dates for the default certificate and privacy key. Use this page to
upload an SSL certificate and privacy key, which allows the device to be accessed in secured
mode.
First upload the SSL certificate, and then device will prompt to upload privacy key. If either of the
files are invalid the device will send a notification. The device will give notification on Successful
upload. On successful upload device will prompt to reboot the device. If you want to reboot click
Ok or click Cancel to cancel the reboot operation.
Figure 2-17 BMC Web Console – SSL Upload
Remote Session
Use this page to enable/disable encryption on KVM or Media during a redirection session.
BMC Integrated Web Console
007-5818-003 49
KVM Encryption
Enable/Disable encryption on KVM data during a redirection session. Choose any one from the
supported encryption techniques.
Keyboard/Mouse Only
If KVM Encryption is set to None, the Keyboard & Mouse data can still be encrypted using
Blowfish encryption. This option has the least performance impact while still encrypting the most
important data.
Media Encryption
Enable/Disable encryption of Media data during a redirection session. Disabling encryption can
improve performance of KVM or Media redirection.
USB Key Emulation Mode
Two types of emulation are:
Floppy - Emulated USB Key will be detected as Floppy drive in remote machine.
Hard disk - Emulated USB Key will be detected as Hard disk (or) Removable drive in
remote machine.
50 007-5818-003
2: System Interfaces
Figure 2-18 BMC Web Console – Remote Session
Mouse Mode
Mouse Mode shows you which mode the mouse is currently in, and allows you to change the
mouse mode to the following options: Absolute Mode, Relative Mode, and Other Mode
Absolute Mode
Select to have the absolute position of the local mouse sent to the server. Use this mode for
Windows OS and newer Red Hat linux versions (RHEL 6.x).
BMC Integrated Web Console
007-5818-003 51
Relative Mode
Select Relative Mode to have the calculated relative mouse position displacement sent to the
server. Use this mode for other Linux releases such as SUSE (SLES) and older versions of Red
Hat (RHEL 5.x). For best results, server OS mouse acceleration/threshold settings can be reduced,
or mouse calibration in JViewer performed.
Other Mode
Select Other Mode to have the calculated displacement from the local mouse in the center
position, sent to the server. Under this mode ALT+C should be used to switch between Host and
client mouse cursor. Use this mode for SLES 11 Linux OS installation.
The mouse mode is set by clicking the Save button.
Figure 2-19 BMC Web Console – Mouse Mode Setting
52 007-5818-003
2: System Interfaces
Keyboard Macros
The Keyboard Macros page is where you can view and modify keyboard macros. Key Sequence
is the sequence of key events to playback when the macro button is pushed. Button Name is the
optional short name to appear on the button of the Remote Console. If left blank, the key sequence
string will be used as the button name. Click the Help button in the upper right corner
(Figure 2-20), to see the supported key names (Table 2-5). In this example, two macros have been
defined: Ctrl+Alt+Del and Alt+Tab (Figure 2-21).
Figure 2-20 BMC Web Console – Logout, Refresh, and Help buttons
Figure 2-21 BMC Web Console – Keyboard Macros
BMC Integrated Web Console
007-5818-003 53
Key Sequence
When using a key sequence, keep the following definitions in mind:
'+' between keys indicates the keys should be held down together
'-' between keys indicates the previous keys should first be released before the new key is
pressed
'*' inserts a one second pause in the key sequence.
Caution: Video is not updated during macro execution.
Examples:
Ctrl+Alt+Del
Ctrl+B-Enter-**Enter
Ctrl+B-Enter-*\*-Enter
Keys
Key names are either a printable character such as “a”, “5”, “@”, etc. or else one of the
non-printable keys in the table below. Names in parentheses are aliases for the same key. Numeric
keypad keys are prefixed with “NP_”.
A plain ‘*’ means a pause; use ‘\*’ for the actual ‘*’ key. The ‘\’ key must also be escaped as ‘\\’.
Table 2-5 Supported Key Names
Shift (LShift) RShift Ctrl (LCtrl) RCtrl
Alt (LAlt) RAlt (AltGr) Win (LWin) RWin
Enter Esc F1 - F12
Bksp Tab CapsLk Space
Ins Del Home End
PgUp PgDn Context (Menu)
Up Left Down Right
54 007-5818-003
2: System Interfaces
Note: Key sequences are sent to the target as scancodes that get interpreted by the target OS, so
they will be affected by modifiers such as Numlock as well as the OS keyboard language setting.
Alerts
The Alerts page allows you to configure which system events generate Alerts and the external
network destinations they should be sent to.
When one of the selected system events occurs, an alert is generated and sent to the configured
destination(s). Each LAN channel can have up to two destinations.
Globally Enable Platform Event Filtering
Global control for enabling or disabling platform event filtering. When filtering is globally
disabled through this setting, alerts will not be sent. This can be used to prevent sending alerts until
you have fully specified your desired alerting policies.
Select Events
Select one or more system events that will trigger an Alert. Clearing all events disables Alerts.
These events correspond to the IPMI preconfigured Platform Event Filters.
LAN Channel
Select which LAN Channel to configure destinations for. Each LAN Channel has it's own set of
up to two destinations. Alert destinations can be one of two types:
•SNMP Trap
Email (requires Alert Email to be configured)
NumLk NP_Div NP_Mult NP_Minus
NP_Plus NP_0 - NP_9 NP_Dec NP_Enter
PrtSc (SysRq) ScrLk Pause (Break)
Table 2-5 Supported Key Names
BMC Integrated Web Console
007-5818-003 55
The Check All button selects all events to generate Alerts. The Clear All button unchecks all
events so no Alerts will be generated. Click the Save button to save any changes made.
Send Test Alert
To test whether an alert will reach it's destination, set the LAN Channel field to the desired channel
and configure at least one destination. Then click Send Test Alerts to send a simple test alert to
the destination(s) for that Channel.
Figure 2-22 BMC Web Console – Alerts
Alert Email
Alert Email Settings allows you to configure how Alerts are sent by email to an external SMTP
Mailserver. Each LAN Channel has a separate configuration, selected through the drop-down
56 007-5818-003
2: System Interfaces
menu. The SMTP Server IP is the IP address of the remote SMTP Mailserver that Alert email
should be sent to. The IP Address is made of 4 numbers separated by dots as in
xxx.xxx.xxx.xxx where ‘xxx’ ranges from 0 to 255 and the first ‘xxx’ group must not be
0. The Sender Address is the string to be put in the From: field of outgoing Alert emails. Local
Hostname is the hostname of the local machine that is generating the alert, included in the
outgoing Alert email. The Local Hostname is a string of maximum 31 alpha-numeric characters.
Spaces and special characters are not allowed.
Figure 2-23 BMC Web Console – Alert Email Settings
Node Manager
Use this page to configure the system's Node Manager.
BMC Integrated Web Console
007-5818-003 57
Figure 2-24 BMC Web Console – Node Manager Power Policies
Node Manager Power Policies
This table lists the currently-configured policies. Selecting an item from the table will populate
the editable fields in the settings section below.
Policy Number
The policy number to add/edit/delete. Valid range is 0-255.
In the policy table, policy numbers with an asterisk (*) are policies set externally using a
non-platform domain. Changing parameters on these policies will not affect their triggers, trigger
limits, reporting periods, correction timeouts, or aggressive CPU throttling settings.
58 007-5818-003
2: System Interfaces
Enabled
Check this box if the policy is to be enabled immediately.
Hard Shutdown
Check this box to enable a hard shutdown if the policy is exceeded and cannot be corrected within
the correction timeout period.
Log Event
Check this box to enable the node manager to send a platform event message to the BMC when a
policy is exceeded.
Power Limit
The desired platform power limit, in watts.
Use Policy Suspend Periods
If enabled, you may configure policy suspend periods. Each policy may have up to five suspend
periods.
Suspend periods are repeatable by day-of-week. Start and stop times are designated in 24 hour
format, in increments of 6 minutes. To specify a suspend period crossing midnight, two suspend
periods must be used.
Policy Defaults
For all policies set through this web page, several default values will be applied:
Domain: Platform - Power for the entire platform.
Trigger: None - Always monitor after end of POST.
Aggressive CPU Power Correction: AUTO - Use of T-states and memory throttling
controlled by policy exception actions.
Trigger Limit: None.
Reporting Period: 10 seconds - This is a rolling average for reporting only. It will not affect
the average power monitored by the node manager.
BMC Integrated Web Console
007-5818-003 59
Correction Timeout: 22.555 seconds - Maximum time for the NM to correct power before
taking an exception action (i.e. shutdown or alert).
Remote Control Tab
This section of the BMC Web Console allows you to remotely monitor and control the server.
Options available in the left navigation pane are Console Redirection for remote server
management. Server Power Control shows the current power status, and allows power
operations. Finally the Virtual Front Panel is a graphic display of the front panel allowing for
remote front panel functionality.
Console Redirection
From the Console Redirection page, if available, you can launch the remote console KVM
redirection window. The remote console requires an RMM (Remote Management Module) add-in
card, otherwise the launch button is greyed-out (Figure 2-25).
60 007-5818-003
2: System Interfaces
Figure 2-25 BMC Web Console – Console Redirection (greyed-out)
Clicking on Launch Console (Figure 2-26) will prompt for download of JViewer.jnlp file. When
the file is downloaded and launched, the Java redirection window will be displayed.
Figure 2-26 BMC Web Console – Launch Console button (available)
Note: Java Run-Time Environment (JRE, version 6 update 10 or later) must be installed on client
prior to launch of JNLP file.
BMC Integrated Web Console
007-5818-003 61
Figure 2-27 Remote Console – Java redirection window
Server Power Control
This page shows the power status of the server and the following power control operations can be
performed:
Table 2-6 Server Power Control Actions
Option Details
Reset Server Selecting this option will hard reset the host
without powering off.
Force-enter BIOS Setup Check this option to enter into the BIOS setup
after resetting the server.
Power Off Server Selecting this option will immediately power
off the host.
62 007-5818-003
2: System Interfaces
Figure 2-28 BMC Web Console – Power Control and Status
Graceful Shutdown Selecting this option will soft power off the
host.
Power On Server Selecting this option will power on the host.
Power Cycle Server Selecting this option will immediately power
off the host, then power it back on after one
second.
Table 2-6 Server Power Control Actions
Option Details
BMC Integrated Web Console
007-5818-003 63
All power control actions are done through the BMC and are immediate actions. It is suggested to
gracefully shut down the operating system via the KVM interface or other interface before
initiating power actions.functionaryVirtual Front Panel
The Virtual Front Panel is a graphic representation of the front panel, providing remote
functionality virtually.
Table 2-7 System Information Details
Button Details
Power Button Power button is used to Power ON or Power
OFF
Reset Button Reset Button is used to reset the server while
system is ON
Chassis ID Button When Chassis ID button is pressed then the
chassis ID LED, on the front and rear of the
unit are lit (solid blue). If the button is pressed
again the chassis ID LED turns off.
NMI Button At present, NMI button is disabled.
Status LED Status LED will reflect the system status and
will automatically sync with BMC every 60
seconds. If any abnormality occurs in system,
then status LED will change accordingly.
Thermal fault means fault occurred in one of
Thermal sensors present in BMC. Fan fault
means fault occurred in one of the system fans.
System fault means a fault occurred because of
system errors. Power fault means fault
occurred in one of Power sensors Here, fault
means a sensor value crossed
upper-non-critical, or upper-critical value, or
lower-non-critical value, or lower-critical
value.
64 007-5818-003
2: System Interfaces
Power LED Power LED shows system power status. If
LED is green then System is ON. If LED is
grey then System is OFF.
Chassis ID LED The Chassis ID LED will be lit blue when the
Chassis ID LED button is pushed. This is the same
as the Locator LED on the physical control panel
(Table 2-1). When the Locator button is pushed
again, the LED will be lit. There is a corresponding
LED on the back of the server that will be lit blue as
well. This function can be done physically through
the Control Panel (Figure 2-1).
Table 2-7 System Information Details
Button Details
007-5818-003 65
Chapter 3
3. System Software
Overview
This chapter contains information on each of the software sets necessary for zoning MIS Servers
and JBODs, organized in four sections: “Linux Zoning Tools” on page 68, “Windows Zoning
Tools” on page 88, “CLI Zoning Tool” on page 109, and “Disk RAID Support” on page 122.
Zoning is required when multiple SAS connections are operational, in order to stop drives from
being affected by other non-owner SAS controllers (HBAs). Zoning allows the various SAS
connections to be accessible only to the drives that they own. Essentially, zoning allows an
administrator to control who can see what. When no zoning is enabled, all the SAS connections
can see all of the drives. For dual-ported SAS drives, both ports will be exposed, so the drives will
show up twice. This situation will cause conflict between the HBAs.
Zoning can be either hard or soft. In hard zoning, each device is assigned to a particular zone, and
this assignment does not change. In soft zoning, device assignments can be changed by the
network administrator to accommodate variations in the demands on different servers in the
network. T10 or PHY-based Zoning may be implemented on the MIS Server. For the MIS JBOD,
only PHY-based Zoning is supported at this time.
Zones for Linux and Windows features a GUI for ease-of-use, but is limited to 72 drives, requires
an LSI MegaRAID card, and cannot zone JBODs at this time (future releases will). The CLI
Zoning Tool is command line only, but is able to zone any number of JBODs or MIS servers.
There are two main tools for zoning, the SGI Zones application and the SGI CLI Zoning Tool.
Both zoning tools require the presence of other programs in order to operate. The Zones
application offers a GUI interface, but is not able to zone JBODs at this time (future releases will
support zoning JBODs). The CLI Zoning Tool can zone JBODs, but is a command-line only
application.
66 007-5818-003
3: System Software
Important: The Zones program is run on the server you wish to zone. JBODs are zoned through
the hardware using the CLI Zoning tool installed on a server or laptop and using an Ethernet
crossover cable to connect a server/laptop running either a Windows or a Linux operating system
and the CLI Zoning application software.
Section Guide
This Guide is organized into four sections, “Linux Zoning Tools” on page 68, “Windows Zoning
Tools” on page 88, “CLI Zoning Tool” on page 109, and “Disk RAID Support” on page 122.
Inside the first three sections you will find instructions on downloading and installing the software
needed, and the steps to prepare for, and zone MIS Platforms. The final section contains
information on the different RAID configurations available, advantages and disadvantages of each
configuration, as well as best practices.
Within the first section, “Linux Zoning Tools”, “Installing Linux Software” on page 68 gives
instructions for downloading and installing the programs required to run the Linux-based zoning
tools. The GUI software programs used are “MegaRAID Storage Manager for Linux”,
“MegaCli64 for Linux”, and “Zones for Linux”. This section also includes information in
downloading the command-line-only “CLI Zoning Tool for Linux”. Instructions for zoning using
the Linux tools are given:
“Verify Drives Seen” on page 71
“Verify Drives Seen” on page 71
“Installing a Drive through Zones for Linux” on page 74
“Creating the Drive Groups in MegaRAID for Linux” on page 77
“Formatting the Drives using YaST2 in Linux” on page 79
“Removing a Drive in Zones for Linux” on page 83
“Additional Features in Zones for Linux” on page 85 details the extra features available in Zones
for Linux including (from page 24):
“Loading .csv Configuration Files in Linux”
“Save a Configuration to .csv file in Linux”
Information on using the CLI Zoning Tool can be found in “CLI Zoning Tool” on page 109.
Section Guide
007-5818-003 67
The second section, “Windows Zoning Tools”, contains “Installing Windows Software” on
page 88. Here are instructions for downloading and installing the programs required to run the
Windows-based zoning tools. The GUI software programs used are “MegaRAID Storage
Manager for Windows”, “MegaCli64 for Windows”, “Zones for Windows”, “Python for
Windows”. This section also contains information on downloading the “CLI Zoning Tool for
Windows” on page 91. Instructions on using the Windows GUI are given:
“Verify Drives Seen in Windows” on page 92
“Verify Drives Seen in Windows” on page 92
“Installing a Drive in Zones for Windows” on page 95
“Creating the Drive Groups in MegaRAID for Windows” on page 98
“Formatting the Drives in Windows Server Manager” on page 100
“Removing a Drive in Zones for Windows” on page 106
“Additional Features in Zones for Windows” on page 107 details “Loading .csv Configuration
Files in Zones for Windows” on page 107, “Adapter Assignment Synchronization in Zones for
Windows” on page 108, and “Save a Configuration to .csv file in Zones for Windows” on
page 109.
The third section details out the “CLI Zoning Tool,” the command-line-only tool that is currently
the only tool that can zone JBODs, as well as MIS Servers. The CLI Zoning Tool requires the
presence of Python 2.6 or 2.7 to be present in order to run. Python is standard on most Linux-based
machines, and will need to be installed for Windows-based machines for the CLI Zoning tool to
run. (See “Python for Windows‚” on page 91). “Preparing to Zone using the CLI Zoning Tool‚”
on page 110 gives instruction on what is necessary before “Zoning Using CLI Zoning Tool‚” on
page 117 and “Editing the .csv File for the CLI Zoning Tool‚” on page 119.
Finally, “Disk RAID Support” on page 122 explains the different RAID arrays available and the
benefits and drawbacks of each. There are special considerations in creating RAID arrays for use
on StorBricks, that is, there are RAID configurations, namely 6+2 and 7+2, that ensures there is
no single point of failure on the StorBricks within the MIS system. “RAID Configuration Notes”
on page 129 details how to manage those concerns.
68 007-5818-003
3: System Software
Linux Zoning Tools
The following instructions are for Linux-based MIS Servers, and for Linux-based machines
installing the CLI Zoning Tool. (Operation of the CLI Zoning Tool remains the same across
platforms. See “CLI Zoning Tool” on page 109 for more information.)
Installing Linux Software
There are three programs necessary for SGI Zones for Linux management of MIS systems. These
programs give operational control over the hardware and its performance, including zoning of the
drives. They are MegaRAID Storage Manager, MegaCli64, and Zones for Linux. To run the CLI
Zoning Tool on a Linux machine, only one program is necessary (the CLI Zoning Tool itself).
MegaRAID Storage Manager for Linux
Note: MegaRAID Storage Manager is not necessary for zoning using the CLI Zoning Tool.
The MegaRAID Storage Manager is used to prepare the drives for zoning using the Zones tool,
and creating the drive groups after zoning using the Zones tool.
1. Go to http://lsi.com and search for MegaRAID Storage Manager.
2. Select and accept the latest Linux version for download and save the .tar file.
3. Change directory to where you have saved your .tar file (e.g., # cd /usr)
4. Unzip the .tar file, # tar –xf <filename.tar> (e.g.,
# tar –xf MSM_linux_x64_installer-12.01.03-00.tar)
5. Once the folder has unzipped properly, a new folder will appear called disk.
6. Change your directory into the disk folder (i.e., # cd /usr/disk)
7. Install the Dependencies, (i.e., # ./install.csh)
8. Install the MegaRAID Storage Manager rpm file, # rpm –ivh
<MegaRAID_Storage_Manager_currentversion.rpm> (e.g.,
# rpm –ivh MegaRAID_Storage_Manager-12.01.03-00.noarch.rpm)
9. Change to /usr/local directory, where MegaRAID Storage Manager will appear
(i.e., # cd /usr/local/MegaRAID\ Storage\ Manager/)
Linux Zoning Tools
007-5818-003 69
10. Issue ./startupui.sh command to start MegaRAID Storage Manager GUI (i.e.,
# ./startupui.sh).
MegaRAID Storage Manager GUI for Linux will appear.
MegaCli64 for Linux
MegaCli64 is a program required for the Zones tool to match the StorBrick to the adapters. It runs
in the background underneath Zones.
1. Go to http://lsi.com and search for MegaCli64.
2. Select and accept the latest Linux version for download.
3. Once downloaded, unzip the folder (e.g., #tar –xf <tarfilename.tar>)
4. Install MegaCli rpm file (e.g., #rpm –ivh <rpmfilename.rpm>)
Zones for Linux
Zones is proprietary SGI software, used to zone drives on the MIS Server Platform (and eventually
MIS JBOD units as well). To install the software complete the following instructions.
1. Go to http://support.sgi.com
2. Download the latest Linux version of Zones.
3. Unzip Zones.tar (e.g., # tar –xf Zones.tar). A Zones folder will appear.
4. Copy the Zones folder into the /opt directory, thus creating the /opt/Zones
directory
Warning: Zones must be installed in the /opt/Zones/ directory or it will not work.
5. Change directory into the new Zones folder (e.g., # cd /opt/Zones)
Run the Zones program by typing # python GUI.py.
Note: On first use, the Zones program will ask you to enter the paths to MegaCli64 (Figure 3-1).
The default path for MegaCli64 is /opt/MegaRAID/MegaCli. These paths can be changed at
any time by going to menu option Setup > Tools.
70 007-5818-003
3: System Software
Figure 3-1 Zones – First-use path configuration
If the path is improperly set, an error message will appear. The following error message means
that the path set is for a folder that does not contain MegaCLI64 (Figure 3-2).
Figure 3-2 Zones – Error message: Improper path
CLI Zoning Tool for Linux
The CLI Zoning Tool is proprietary SGI software, used to zone drives on the MIS Server Platform
and MIS JBOD units, and is the only tool that can zone JBODs at this time (future releases of
Zones will support zoning JBODs). To install the software:
1. Create a directory for the application: mkdir /opt/ShackCLI
2. Go to http://support.sgi.com
3. Download the latest version of ShackCLI_release_xxx.zip into the above directory
4. Extract the files: unzip ShackCLI_release_xxx.zip
Linux Zoning Tools
007-5818-003 71
To run the CLI Zoning Tool, follow the instructions given in “CLI Zoning Tool” on page 109.
Verify Drives Seen
In Linux systems, open Partitioner and verify the only disks the system sees are system drives,
labeled /opt/ (Figure 3-23 on page 82).
Note: Unconfiguring drives removes them from the system.
Linux Zones Tool
The SGI Zones tool is a proprietary software program that uses a GUI interface to zone the drives
in the MIS Server. Future releases will support zoning JBODs as well.
Open the Zones tool. In Linux, this is done by changing to the /opt/Zones folder and typing
# python GUI.py.
Figure 3-3 Linux Zones Welcome
Note: When Zones for Linux opens, leave the terminal window, from which you launched Zones,
open. This window must be left open or the program shuts down. It will be used later in
downloading session information to the StorBricks.
72 007-5818-003
3: System Software
Figure 3-4 Zones User Interface
To begin zoning, click Open Session (Figure 3-5). This queries each expander for the information
contained in the expander .bin file.
Figure 3-5 Zones – Show All, Open Session, Save Session, Download Session, and Exit buttons
Click Create New Session (Figure 3-6).
Linux Zoning Tools
007-5818-003 73
Figure 3-6 Zones – Opening a Session
The Zones program will ask you for a Session Alias (Figure 3-7). This alias will be added to a
time stamp to create folder for that session’s files: YYMMDD_HHMMSS_alias.
Figure 3-7 Zones – Enter Session Alias
Enter an alias and click Okay.
74 007-5818-003
3: System Software
Note: Aliases have a 64 character limit, and may not contain spaces or non-alpha-numeric
characters; these will be replaced with an underscore if used and a warning message will appear
(Figure 3-8).
Figure 3-8 Zones – Alias Help Warning Message
In Zones for Linux, .bin files are automatically named #.bin where # is the number of the
StorBrick in question. After the .bin file has been fetched, each file in converted into an XML
file, named using the same convention, with the file extension .xml (e.g., #.xml).
Upon opening, Zones for Linux will show a blank template for zoning. It will not show any zoning
that may be currently implemented. It does not show drives that are presently installed or zones
that are currently active. Each adapter has its own tab above the StorBrick layouts (Figure 3-9).
The number of Adapter tabs will directly correspond to the number of Adapters Zones sees in the
system.
Figure 3-9 Zones – Adapter tabs
Installing a Drive through Zones for Linux
After creating a new session, click the Show All button. This enables all boxes on all adapters
(Figure 3-10), even if there are no drives physically present. This is so a drive can be zoned to the
system that isn’t seen by the system yet.
Linux Zoning Tools
007-5818-003 75
Note: If a drive is zoned to more than one adapter, a warning message (Figure 3-11) will appear.
In Windows, this warning tells which adapter that drive is already zoned to, and a message asking
if you would like to continue. In Linux, there is just a warning message asking if you would like
to continue. It is possible to zone one drive to two adapters. This is allowed, but not suggested, as
it can cause data collisions.
Figure 3-10 Zones – Show All
Figure 3-11 Zones – Adapter Assignment Warning Message
Check the boxes for the drives you want to zone (Figure 3-12) according to the configuration of
your liking. Click Save Session. This takes the screen configuration and saves it to an .xml file
that is then converted into a .bin file.
76 007-5818-003
3: System Software
Note: The Save Session button will become enabled whenever a change has been made in the
Zones configuration.
In Linux, the name of the new .bin file is STORBRICK_#.bin where the only new element to
the file name is the STORBRICK_. In Windows, the name of the new .bin file is
sbn#_****.bin where the only new element to the file name is the n after sb.
Important: Each time the Save Session button is pushed, these files are overwritten. There is no
way to retrieve previously saved information.
Click the Download button. Once complete, this function pushes the files back into the expanders.
Select which StorBricks to push the files to, and click Okay to push the files.
Figure 3-12 Zones – Select StorBricks for download
In Linux, downloads are confirmed through the terminal window, which show a pane where you
can verify Image Validation and Checksum have passed, and then download files by
entering Y (see Figure 3-27 on page 85).
Linux Zoning Tools
007-5818-003 77
Warning: If Y is chosen and the Image Validation and Checksum have not passed,
unknown results will occur, including losing expander information, requiring a field
technician to service the machine.
Power cycle the machine.
Creating the Drive Groups in MegaRAID for Linux
Right-click on the expander and select Create a Virtual Drive (Figure 3-13).
Figure 3-13 MegaRAID – Create a Virtual Drive
A screen will pop up asking you to choose Simple or Advanced (Figure 3-14). In Simple mode,
the drives are chosen for you. In Advanced mode, you get to choose the drives, and are given
additional selections in RAID levels, allowing for spanned (00, 10, 50, 60) drives.
Figure 3-14 MegaRAID – Create Virtual Drive mode
78 007-5818-003
3: System Software
Figure 3-15 Create Virtual Drive – Drive Group Settings
Choose Write Back BBU (battery back-up unit). This mode is the safest and the fastest, and will
automatically switch from caching mode to write-straight-to-disk whenever battery power has
reached low. Write Through writes straight to disk. Write Back is a cached data flow.
Warning: If you select Write Back and power to the system is lost, data is lost.
Click Next, and a summary screen verifying settings will appear (Figure 3-16).
Figure 3-16 Create Virtual Drive – Summary
If the settings are correct, click Finish, and click Ok.
Linux Zoning Tools
007-5818-003 79
Formatting the Drives using YaST2 in Linux
Drives may be formatted using YaST2 Partitioner. In Linux, the folders that the drives will be
mounted to need to be created first. Each mount will need a new folder. Some Linux customers
will have the ability to issue the YaST2 command, bringing up a GUI to partition drives.
Otherwise, drives are formatted and mounted using command line.
1. Issue YaST2 command (i.e., # yast2) to launch the YaST 2 Server Manager GUI
(Figure 3-17).
Figure 3-17 Yast2 Server Manager GUI
2. Double-click Partitioner to launch.
80 007-5818-003
3: System Software
3. A warning message will appear (Figure 3-18). Click Yes.
Figure 3-18 YaST2 – Warning Message
4. Verify that all of your disks have appeared under Hard Disks (Figure 3-19).
Figure 3-19 YaST2 – Drives have appeared
5. Under Hard Disks, select the disk you would like to partition and click Add at the bottom of
the screen (Figure 3-20).
Linux Zoning Tools
007-5818-003 81
Figure 3-20 YaST2 – Add button
6. Select the partition size (Figure 3-21) and click Next.
Figure 3-21 YaST2 – Select Partition Size
7. Format the partition using ext3, mount the disk to your desired folder, and click Finish
(Figure 3-22).
Figure 3-22 YaST2 – Format & Mount the Drive
8. Verify the partition shows up (Figure 3-23) and click Next.
82 007-5818-003
3: System Software
Figure 3-23 YaST2 – Check for Partition
9. Click Finish (Figure 3-24).
Figure 3-24 YaST – Click Finish
It may take several minutes for the system to mount the disk.
Linux Zoning Tools
007-5818-003 83
Figure 3-25 YaST2 – Disk Mounting (in process)
Once the disk is mounted (Figure 3-25), the system will return you to the beginning YaST2 GUI.
Removing a Drive in Zones for Linux
From the Zones user interface (Figure 3-4 on page 72), you can unzone a drive from an adapter
by unchecking boxes. Once the drives are unzoned, click the Save Session button (Figure 3-5 on
page 72). This action takes the configuration shown on the screen, saves it to an .xml file, that is
then converted into a .bin file, and stores it on the expander card.
Note: The Save Session button will become enabled whenever a change has been made in the
Zones user interface.
In Linux, the name of the new .bin file is STORBRICK_#.bin where the only new element to
the file name is the STORBRICK_. In Windows, the name of the new .bin file is
sbn#_****.bin where the only new element to the file name is the n after sb.
Important: Each time the Save Session button is pushed, these files are overwritten. There is no
way to retrieve previously saved information.
84 007-5818-003
3: System Software
Click the Download Session button. A window will appear asking you which StorBricks you
would like to download to, by selecting checkboxes next to the corresponding StorBrick
(Figure 3-26), or choosing Select All. Click Okay when satisfied.
Figure 3-26 Zones – Downloading setting to StorBrick(s)
After clicking Okay, go to the terminal window that was started automatically upon launching the
Zones for Linux tool (“Linux Zones Tool Note‚” on page 71). In this window, check to see that
the values for Image Validation and Checksum are both listed as Passed. If they are, a
Y will be the default option in answer to the question: Are you sure to download file
to expander? [sic]. You will have to answer Y for each StorBrick you wish to download files
to (Figure 3-27).
If either Image Validation or Checksum have failed, the default answer will be N, which
is what you should choose if they have failed. If N is chosen, the information is not sent to the
StorBrick, and that StorBrick is skipped. (Reasons for Image Validation or Checksum not
passing could be—for example—an invalid template file or conversion from .xml to .bin
failed.)
Warning: If Y is chosen and the Image Validation and Checksum have not passed,
unknown results will occur, including losing expander information, requiring a field
technician to service the machine.
Linux Zoning Tools
007-5818-003 85
Figure 3-27 Zones – Are you sure to download files to expander?
If Image Validation or Checksum are showing as Failed, choose N to answer the
question: Are you sure to download file to expander? [sic] Once all downloads
have either been answered Y or N, if any of the downloads were answered N, you must close and
re-open the Zones program.
Warning: If there are ANY failed downloads, DO NOT REBOOT OR POWER CYCLE
THE MACHINE under ANY circumstance.
If the problem of failed Image Validation or Checksum continues, contact SGI Technical
Support (“Product Support” on page xxv of the MIS User Guide). Otherwise, once all the
downloads have passed and been answered Y, power-cycle the machine for these changes to take
effect.
Additional Features in Zones for Linux
The Zones application also includes tools for “Loading .csv Configuration Files in Linux” and
“Save a Configuration to .csv file in Linux”. Their instructions are provided here.
86 007-5818-003
3: System Software
Loading .csv Configuration Files in Linux
One option for zoning is editing zoning values in a .csv file, created through spreadsheet
programs that support .csv file extension (e.g., Microsoft Excel).
Figure 3-28 Zones – Select CSV File (Linux)
To begin in either version, select Open Session (Figure 3-6 on page 73). Select Load CSV File.
Enter an alias for the session (Figure 3-7 on page 73) and click Okay. A File Location pane will
open (Figure 3-28). Select the .csv file you wish to upload, and click Okay. The configuration
entered in the .csv file will show in the Zones GUI (Figure 3-4), with the exception that the
zoning may need to be synchronized with the hardware, explained below.
Save a Configuration to .csv file in Linux
The Zones tool also allows you to save configurations created in the GUI as a .csv file. In the
top menu bar, there is the option to Save Configuration to csv (Figure 3-29).
Figure 3-29 Zones – Save Configuration to .csv
Click that menu option, and in the drop-down menu that appears, click Save Configuration
(Figure 3-29). A Save CSV pop-up window will appear (Figure 3-30), asking you to enter a
Directory (file location) and File Name (name for the .csv file).
Linux Zoning Tools
007-5818-003 87
Figure 3-30 Zones – Save CSV pop-up
There is also the option to browse to a directory using the Browse button. This will open a
navigation pane where you can select the directory (Figure 3-31).
Figure 3-31 Zones – Select Directory navigation pane
If a directory is not chosen, or an improper directory choice is made (choosing a program instead
of a file, etc.) an error message will appear (Figure 3-32).
88 007-5818-003
3: System Software
Figure 3-32 Zones – Error: Improper Directory Selection
Once satisfied, click Okay (Figure 3-32). The values set in the GUI will be output as a .csv file
in the directory selected earlier.
Windows Zoning Tools
The following instructions are for Windows-based MIS Servers, and for Windows-based
machines installing the CLI Zoning Tool. (Operation of the CLI Zoning Tool remains the same
across platforms. See “CLI Zoning Tool” on page 109 for more information.)
Installing Windows Software
There are four programs necessary for Windows management of MIS systems. These programs
give operational control over the hardware and its performance, including zoning of the drives.
They are MegaRAID Storage Manager, MegaCli64, Zones, and Python. To run the CLI Zoning
Tool on a Windows machine, the machine must have Python installed as well as the CLI Zoning
Tool itself.
MegaRAID Storage Manager for Windows
Note: MegaRAID Storage Manager is not necessary for zoning using the CLI Zoning Tool.
The MegaRAID Storage Manager is used to prepare the drives for zoning, and to create the drive
groups after zoning.
1. Go to http://lsi.com and search for MegaRAID Storage Manager.
2. Select and accept the latest Windows version for download.
Windows Zoning Tools
007-5818-003 89
3. Download and install.
MegaRAID Storage Manager GUI for Windows will appear.
MegaCli64 for Windows
MegaCli64 is a program required for the Zones tool to match the StorBrick to the adapters. It runs
in the background underneath Zones.
1. Go to http://lsi.com and search for MegaCli64.
2. Select and accept the latest Windows version for download.
3. Download and install.
Zones for Windows
Zones is proprietary SGI software, used to zone drives on the MIS Server Platform.
1. Go to http://support.sgi.com
2. Download the latest Windows version of Zones.
3. Unzip the Zones.zip file. Make sure that destination the files will be extracted to is
c:\. A Zones folder will be created automatically, generating the directory
c:\Zones\
To run Zones, go to Start > Run, and type in c:\Zones\Zones.exe, and click Okay.
Note: On first use, the Zones will ask you to enter the paths to MegaCli64 (Figure 3-33). The
default path for MegaCli64 is C:\Program Files (x86)\MegaRAID Storage
Manager\MegaCLI\MegaCli64 and C:\Zones\XTOOLS is the default path for
XTOOLS. These paths can be changed at any time by going to menu option Setup > Tools.
90 007-5818-003
3: System Software
Figure 3-33 Zones – First-use path configuration (Windows)
If the path is improperly set, an error message will appear. The following error messages means
that the path set is for a folder that does not contain MegaCLI64, and Zones will not run properly
(or the system may crash).
Windows Zoning Tools
007-5818-003 91
Figure 3-34 Zones – Error messages from improper path configuration
If the path for the tools are set incorrectly, they must be re-set and the Zones program closed and
re-started for the correct paths to take effect.
Python for Windows
The CLI Zoning Tool requires Python, version 2.6 or 2.7, be installed on the machine that will
perform the zoning.
1. Go to http://www.python.org/download/releases/ and select the latest version of Python.
2. Download and start the installation.
3. The installation will ask which directory in which to install python (the default is
c:\Python##\ where ## is the version number).
4. On the Customize Python pane, click Next.
5. On the Complete Python Installation pane, click Finish.
CLI Zoning Tool for Windows
To install the CLI Zoning Tool software:
1. Go to http://support.sgi.com
2. Download the latest Windows version of the CLI Zoning Tool.
3. Unzip ShackCLI_release_xxx. The program will ask if you want to create a
directory name for the .zip file, click Yes.
To run the CLI Zoning Tool, follow the instructions given in “CLI Zoning Tool” on page 109.
92 007-5818-003
3: System Software
Verify Drives Seen in Windows
In Windows, open Server Manager (Figure 3-35). Verify that the only disks the system sees are
system drives, and that they are labeled c:/.
Figure 3-35 Windows Server Manager – Disk Management
Note: Unconfiguring drives removes them from the system.
Zones for Windows
The SGI Zones tool is a proprietary software program that uses a GUI interface to zone the drives
in the MIS Server. Future releases will support zoning JBODs as well.
Windows Zoning Tools
007-5818-003 93
Figure 3-36 Zones for Windows Welcome
Figure 3-37 Zones Windows User Interface
To begin zoning, click the Open Session button (Figure 3-38). This queries each expander for the
information contained in the expander .bin file.
Figure 3-38 Zones – Open Session, Save Session, Download Session, and Exit buttons
Click Open New Session (Figure 3-39).
94 007-5818-003
3: System Software
Figure 3-39 Zones – Open Session
The Zones program will ask you for a Session Alias (Figure 3-40). This alias will be added to a
time stamp to create folder for that session’s files: YYMMDD_HHMMSS_alias.
Figure 3-40 Zones – Enter Session Alias
Enter an alias and click Okay. Each expander card’s information is fetched using cmd.exe and
placed in the session folder as .bin files.
Note: Aliases have a 64 character limit, and may not contain spaces or non-alpha-numeric
characters; these will be replaced with an underscore if used and a warning message will appear
(Figure 3-41).
Windows Zoning Tools
007-5818-003 95
Figure 3-41 Zones – Alias Help Warning Message
In Zones for Windows, .bin files are automatically named sb#_****.bin where sb stands
for StorBrick, # is the number of the StorBrick in question, and **** are the last four digits of
the expander card’s SAS address. After the .bin file has been fetched, each file in converted into
an XML file, named using the same convention, with the file extension .xml (e.g.,
sb#_****.xml).
Upon opening, Zones UI will show drives that are presently installed and zones that are currently
active. A checkbox will be enabled to show that a drive is in a StorBrick. A check mark in a
checkbox that is enabled shows that the drive is zoned for that adapter. Each adapter has its own
tab above the StorBrick layouts (Figure 3-42). The number of Adapter tabs will directly
correspond to the number of adapters Zones sees in the system.
Figure 3-42 Zones – Adapter tabs (Windows)
Installing a Drive in Zones for Windows
After creating a new session, click the Show All button. This enables all boxes on all adapters
(Figure 3-43), even if there are no drives physically present. This is so a drive can be zoned to the
system that isn’t seen by the system yet.
96 007-5818-003
3: System Software
Note: If a drive is zoned to more than one adapter, a warning message (Figure 3-44) will appear.
In Windows, this warning tells which adapter that drive is already zoned to, and a message asking
if you would like to continue. It is possible to zone one drive to two adapters. This is allowed, but
not suggested, as it can cause data collisions.
Figure 3-43 Zones – Show All
Figure 3-44 Zones – Adapter Assignment Warning Message
Check the boxes for the drives you want to zone, according to the configuration of your liking.
Click Save Session. This takes the screen configuration and saves it to an .xml file that is then
converted into a .bin file.
Windows Zoning Tools
007-5818-003 97
Note: The Save Session button will become enabled whenever a change has been made in the
Zones configuration.
The name of the new .bin file is sbn#_****.bin where the only new element to the file name
is the n after sb.
Important: Each time the Save Session button is pushed, these files are overwritten. There is no
way to retrieve previously saved information.
Click the Download button. Once complete, this function pushes the .bin files back into the
expanders. Select which StorBricks to push the files to, and click Ok to push the files.
Figure 3-45 Zones – Select StorBricks for download
In Windows, a prompt will appear asking to confirm the download operation.
98 007-5818-003
3: System Software
Figure 3-46 Zones tool – Verify download
Click Yes to verify and finish downloading. Power cycle the machine. Click Ok to power off the
machine for power-cycling.
Creating the Drive Groups in MegaRAID for Windows
Power on the machine to complete power-cycle. Open MegaRAID. Right-click on the controller
and select Create a Virtual Drive (Figure 3-47).
Figure 3-47 MegaRAID – Create a Virtual Drive
A screen will pop up asking you to choose Simple or Advanced (Figure 3-48). In Simple mode,
the drives are chosen for you. In Advanced mode, you get to choose the drives, and are given
additional selections in RAID levels, allowing for spanned (00, 10, 50, 60) drives.
Windows Zoning Tools
007-5818-003 99
Figure 3-48 MegaRAID – Create Virtual Drive mode
Create the drive groups.
Figure 3-49 Create Virtual Drive – Simple Settings
Choose Write Back BBU (battery back-up unit). This mode is the safest and the fastest, and will
automatically switch from caching mode to write-straight-to-disk whenever battery power has
reached threshold. Write Through writes straight to disk. Write Back is a cached data flow.
100 007-5818-003
3: System Software
Warning: If you select Write Back and power to the system is lost, data is lost.
Click Next, and a summary screen verifying settings will appear (Figure 3-50).
Figure 3-50 Create Virtual Drive – Summary
If the settings are correct, click Finish, and click Ok.
Formatting the Drives in Windows Server Manager
Drives are formatted using Windows Server Manager (Figure 3-51). Open Server Manager—the
screen should start at Disk Management with the drives showing. If not, click Storage in the
system tree, and click Disk Management. The collection of disks/raidsets will now show.
Windows Zoning Tools
007-5818-003 101
Figure 3-51 Server Manager – Disk Management
Right click in the grey area of the first non-system disk. In the menu that appears choose Initialize
Disk (Figure 3-52).
Figure 3-52 Server Manager – Initialize Disks
A pop-up window will appear, showing all the uninitialized disks (Figure 3-53).
Warning: Be sure to select GPT (GUID Partition Table).
102 007-5818-003
3: System Software
Figure 3-53 Server Manager – Select GPT (GUID Partition Table)
Click OK. All the disks should now show as Online (Figure 3-54).
Figure 3-54 Server Manager – Disks Initialized and Online
Right click the first non-system disk. Select New Simple Volume (Figure 3-55).
Windows Zoning Tools
007-5818-003 103
Figure 3-55 Server Manager – New Simple Volume
Click Next at the New Simple Volume Wizard welcome screen (Figure 3-56). Select the size of
the volume in MB and click Next (Figure 3-57).
Figure 3-56 Server Manager – New Simple Volume Wizard
104 007-5818-003
3: System Software
Figure 3-57 New Simple Volume Wizard – Volume Size
Figure 3-58 New Simple Volume Wizard – Assign Drive Letter or Path
Choose the drive letter to be assigned or click Next for the next drive letter available (Figure 3-58).
Windows Zoning Tools
007-5818-003 105
Figure 3-59 New Simple Volume Wizard – Format Partition
Select the format settings to be used, and click Next (Figure 3-59).
Figure 3-60 New Simple Volume Wizard – Settings Confirmation
Click Finish to format the disks (Figure 3-60). New volumes will show in the Disk Management
window below the volumes (Figure 3-61).
106 007-5818-003
3: System Software
Figure 3-61 New Simple Volume in Server Manager
Removing a Drive in Zones for Windows
From the Zones user interface (Figure 3-37 on page 93), you can unzone a drive from an adapter
by unchecking boxes. Once the drives are unzoned, click the Save Session button (Figure 3-38 on
page 93). This action takes the configuration shown on the screen, saves it to an .xml file, that is
then converted into a .bin file, and stores it on the expander card.
Note: The Save Session button will become enabled whenever a change has been made in the
Zones user interface.
In Windows, the name of the new .bin file is sbn#_****.bin where the only new element
to the file name is the n after sb.
Important: Each time the Save Session button is pushed, these files are overwritten. There is no
way to retrieve previously saved information.
Click the Download Session button (Figure 3-38 on page 93). This function pushes the files back
into the expanders. Select which StorBricks to push the files to, and click Yes (Figure 3-62) to
push the files.
Figure 3-62 Zones – Verify Download
Power cycle the machine to have these changes take effect.
Windows Zoning Tools
007-5818-003 107
Additional Features in Zones for Windows
The Zones application also includes tools for loading a .csv file, adapter synchronization, and
saving a .csv file.
Loading .csv Configuration Files in Zones for Windows
One option for zoning is editing zoning values in a .csv file, created through spreadsheet
programs that support .csv file extension (e.g., Microsoft Excel). Zoning assignments done in
this way will assign zoning at the hardware level, whereas zoning through the Zones GUI zones
at the MegaRAID Storage Manager level, which may or may not correspond with the hardware.
Synchronizing the two is handled at the hardware level and not through the Zones program (see
“Adapter Assignment Synchronization in Zones for Windows” on page 108).
[
Figure 3-63 Zones – Select CSV File
To begin, select Open Session (Figure 3-38 on page 93). Select Load CSV File. The alias for the
session is the .csv file name with the “.csv” removed. A File Location pane will open
(Figure 3-63). Select the .csv file you wish to upload, and click Okay. The configuration entered
in the .csv file will show in the Zones GUI (Figure 3-37 on page 93), with the exception that the
zoning may need to be synchronized with the hardware, explained below.
108 007-5818-003
3: System Software
Figure 3-64 Zones – Error message: Canceling csv file selection (Windows)
Note: If you wish to cancel selecting a .csv file, and click the Cancel button on the Select the
Zoning CSV File, you will receive an error message (Figure 3-64). Click Okay to continue
canceling the Select .csv File operation.
Adapter Assignment Synchronization in Zones for Windows
If the adapter assignment in the .csv file loaded (hardware-level assignment) does not
correspond with how MegaRAID Storage Manager assigns the adapters (on a first-see/first-come
software-level assignment), then Zones will show what drives are present on which adapters, but
the zoning will not be there.
To fix this, select Open Session (Figure 3-39 on page 94). Select Create New Session, enter an
alias for the session, and click Ok. Next, click the Unselect All button, located in the upper right
corner of the Zones for Windows UI (Figure 3-65). Now you can assign the drives on the Adapter
tabs, to configure the tool so that it is synchronized with the hardware. Click the Step 2: Save
Session button, and then click the Step 3: Download Session button (Figure 3-38 on page 93).
Select the StorBricks on which you wish to push the new configuration (usually all, in this instance
– Figure 3-45 on page 97). Click Ok. Click Yes to power cycle the machine (i.e., let the MIS
Platform shut down gracefully and then power it back on). This will change the MegaRAID
assignment.
Figure 3-65 Zones – Select All, Unselect All buttons
CLI Zoning Tool
007-5818-003 109
Save a Configuration to .csv file in Zones for Windows
The Zones tool for Windows saves the configurations created in the GUI as a .csv file. This is
done automatically when you save the session, and the .csv file is saved in the session folder.
CLI Zoning Tool
This initial release of the CLI Zoning tool introduces proprietary SGI software, used to zone drives
on a MIS Server Platform and MIS JBOD units. The tool also supports diagnostic functions useful
to Field Service.
In the T10 implementation, SAS zoning access control is implemented by linked switch and
expander devices, with zoning enabled. These devices define a Zoned Portion of a Service
Delivery System (ZPSDS). No host device intervention is required. Each zoning expander device
within the MIS Enclosure maintains an identical zone permission table, so zone access control is
maintained across the entire ZPSDS. The difference between the expanders lies in the definition
of the PHY Zone groups that defines to which Zone Group (ZG) each of the 36 PHY’s belongs.
The Permission Table then maps the Initiator Zone Groups to the Target Zone groups.
The CLI Zoning Tool performs changes through the use of Comma Separated Values (CSV) files.
A Session is defined as the act of querying, editing, saving, and downloading the expander's binary
zoning information.
Important: JBODs are zoned through the hardware using the CLI Zoning tool; only PHY-based
Zoning is supported at this time.
The CLI Zoning Tool runs on any Windows/Linux host that has Python 2.6 or Python 2.7.1
installed. The target MIS JBOD/Server does not require an OS since the CLI Zoning Tool uses the
FanBase Ethernet connection to access the StorBricks. The CLI Zoning Tool allows the zoning of
drive groups larger than 72, hence, it is able to zone JBODs, and multiple JBODs, as single zones.
Table 3-1 is how the Zone Groups (ZG) are implemented in the MIS Enclosure. There are 256
possible ZG in the MIS Enclosure which allows for our maximum drive count of 162.
110 007-5818-003
3: System Software
A configuration file is used by the CLI application to zone the StorBricks. A set of standard
configuration files are included with the CLI Zoning Tool software package. A custom file
can be created, using a spreadsheet application and then saving it as a .csv file (see “Editing the
.csv File for the CLI Zoning Tool” on page 119).
Preparing to Zone using the CLI Zoning Tool
CLI Zoning uses the MIS-S9D proprietary network interface. This interface is to be used ONLY
when zoning. It is located at the front of the chassis at the upper right corner (Figure 3-66). The
chassis must be slid out forward at least one inch in order to connect a network crossover cable.
(See “Sliding the Chassis Forward/Backwards‚” on page 134.)
Table 3-1 Zone Group Implementation
Zone Group Description
ZG0 The Dead Zone that only talks to ZG 127
ZG0 – ZG1 Always enabled
ZG2–3 Enabled for initiators
ZG2 For Initiators that have SMP Zoning Access
ZG3 Initiators that have access to Broadcast
ZG4–7 Reserved per SAS Specification
ZG8–15 The eight possible initiators
ZG16–96 For drives 0–80 for 81 possible drives
ZG97–127 Reserved in the MIS implementation
CLI Zoning Tool
007-5818-003 111
Figure 3-66 MIS-S9D proprietary network interface
Ensure the MIS system is powered on. Use an Ethernet crossover cable to connect a server/laptop
running either a Windows or a Linux operating system and the CLI Zoning application software.
The network port connected to the server/laptop must be set to 192.168.0.xxx (10 will do).
The static IP address of the Fan Base is set to 192.168.0.3, verify connectivity to the Fan Base
with a ping command to 192.168.0.3 from the server/laptop. Verify it responds. If not it will
be necessary to power cycle the MIS server or JBOD.
Editing the ShackCLI.ini file for Linux
In the /opt/ShackCLI directory, open up the ShackCLI.ini file with an editor or vi.
The following is a sample file:
[main]
# Input filename. This must be either a pathname or a simple
# dash (-), which signifies we'll use standard in.
input_source = cli
target = 192.168.0.3
[maxsize]
# When we hit this threshold, we'll alert for maximum
# file size.
threshold = 100
[display]
show_footer = yes
# Fill up all SB infomation before going to Menu
auto_fill = no
[default]
#MIS_Variant = JBOD
112 007-5818-003
3: System Software
MIS_Variant = SERVER
storbrick = 0 1 2 3 4 5 6 7
#storbrick = 0 1 2 3 4 5 6 7 8
cmd = menu
pcsv = /opt/ShackCLI/MIS-Server_2HBA_zoning__PCSV.csv
pbcsv = /opt/ShackCLI/MIS-JBOD_1-IOMOD_zoning_PBCSV.csv
zcsv = /opt/ShackCLI/Zone_Phy_Default.csv
max_zones = 255
max_phys = 36
response_delay_default = 20
First, verify that the target IP address is 192.168.0.3 and is not commented out (i.e., there is
no # at the beginning of the line). Next, make the needed changes to the file, as follows.
1. Set the MIS Variant to the type of system to be zoned, JBOD or SERVER. (Be sure the
other is commented out.)
2. Change the StorBrick count to be either 0–7 for a server, or 0–8 for a JBOD.
3. Select the type of zoning file to be used pcsv, zcsv, or pbcsv.
4. Add the path to where the configuration file is located.
5. Unless issues develop, leave the remaining selections at default.
6. Save the ShackCLI.ini file and close.
7. Execute the CLI command: python ShackCLI.py -–ini ShackCLI.ini
-–cmd menu.
This will set the StorBricks to debug mode, and display a menu (“CLI Zoning Tool Main Menu”
on page 113).
Editing the ShackCLI.ini file for Windows
In the C:\python##\ directory, open up the ShackCLI.ini file with an editor or vi.
The following is a sample file:
[main]
# Input filename. This must be either a pathname or a simple
# dash (-), which signifies we'll use standard in.
input_source = cli
target = 192.168.0.3
[maxsize]
# When we hit this threshold, we'll alert for maximum
# file size.
threshold = 100
CLI Zoning Tool
007-5818-003 113
[display]
show_footer = yes
# Fill up all SB infomation before going to Menu
auto_fill = no
[default]
#MIS_Variant = JBOD
MIS_Variant = SERVER
storbrick = 0 1 2 3 4 5 6 7
#storbrick = 0 1 2 3 4 5 6 7 8
cmd = menu
pcsv = C:\python##\MIS-Server_2HBA_zoning__PCSV.csv
pbcsv = C:\python##\MIS-JBOD_1-IOMOD_zoning_PBCSV.csv
zcsv = C:\python##\Zone_Phy_Default.csv
max_zones = 255
max_phys = 36
response_delay_default = 20
First, verify that the target IP address is 192.168.0.3 and is not commented out (i.e., there is
no # at the beginning of the line). Next, make the needed changes to the file, as follows.
1. Set the MIS Variant to the type of system to be zoned, JBOD or SERVER. (Be sure the
other is commented out.)
2. Change the StorBrick count to be either 0–7 for a server, or 0–8 for a JBOD.
3. Select the type of zoning file to be used pcsv, zcsv, or pbcsv.
4. Add the path to where the configuration file is located.
5. Unless issues develop, leave the remaining selections at default.
6. Save the ShackCLI.ini file and close.
7. Execute the CLI command: python ShackCLI.py -–ini ShackCLI.ini
-–cmd menu.
This will set the StorBricks to debug mode, and display a menu (“CLI Zoning Tool Main Menu”).
CLI Zoning Tool Main Menu
The main menu of the CLI Zoning Tool gives the following options, as listed below and described
in Table 3-2.
1) Set Active Storbrick(s)
2) Display Current StorBrick(s) Zoning
3) Update StorBrick(s) Permissions Table From CSV
4) Update StorBrick(s) Phy Zones From CSV
5) Update StorBrick(s) Phy Based Zoning from CSV
114 007-5818-003
3: System Software
6) Change Zoning type (Server <-> JBOD)
7) Save current StorBrick(s) Zoning to CSV
8) Display StorBrick(s) Settings
9) Display CLI Settings
10) Enter StorBrick CLI (Must select a single StorBrick)
11) Reboot FanBase - will not reset StorBricks
12) Reset FanBase - Danger. This will reset StorBricks
13) Force Storbrick(s) into Debug Mode
14) Exit Storbrick(s) Debug Mode
15) Display PHY Error Counters for selected StorBricks
16) Display PHY Information for selected StorBricks
17) Display StorBrick UUID for selected StorBricks
18) Display StorBrick Firmware Revision Levels for Selected StorBricks
0) Exit CLI - back to command prompt
Table 3-2 CLI Zoning Tool Menu Options and Descriptions
Menu Option Description
1) Set Active
Storbrick(s)
This menu option allows user to select the StorBrick(s) to act
upon. The StorBricks may be entered in any order: 0 1 2 3 4 5 6 7
or 7 6 5 4 3 2 1 0, or in subsets: 0 or 0 1, etc. Storbrick numbers
must be less than or equal 7 for MIS Server and less than or equal
8 for MIS JBOD.
2) Display Current
StorBrick(s) Zoning
Displays the Zoning configuration that is currently stored in the
StorBricks.
3) Update StorBrick(s)
Permissions Table From
CSV
This menu option uses the csv file described in the ini file
under the heading ‘pcsv’ to modify the T10 Zoning Permission
Tables for the selected StorBricks. If no csv file has been
specified in the ini file the CLI Zoning Tool will prompt the
user for the name of the csv file to use.
4) Update StorBrick(s)
Phy Zones From CSV
This menu option uses the csv file described in the ini file
under the heading ‘zcsv’ to modify the T10 PHY Zone Groups
for the selected StorBricks. If no csv file has been specified in
the ini file the CLI Zoning Tool will prompt the user for the
name of the csv file to use.
CLI Zoning Tool
007-5818-003 115
5) Update StorBrick(s)
Phy Based Zoning from
CSV
This menu option uses the csv file described in the ini file
under the heading ‘pbcsv’ to modify the PHY Based Zoning
Tables for the selected StorBricks. This is the only supported
Zoning configuration for MIS JBOD and is an optional
configuration for MIS Server and MIS DC Server. Only one of
PHY Based and T10 Zoning should be implemented within an
MIS Server (although it is technically possible to mix the Zoning
types) and only one of PHY Based Zoning may be supported in
an MIS JBOD.
6) Change Zoning type
(Server <-> JBOD)
This menu option allows the user to change the zoning type from
SERVER/JBOD to JBOD/SERVER. This command will cause
the Selected STORBRICKS T10 Supported flag to be set/unset
depending what the current Zoning type is. For Example, if the
current Zoning type is PHY Based and the user selects this option
then the T10 Zoning Supported flag will be set enabling T10
Zoning to be implemented instead.
7) Save current
StorBrick(s) Zoning to
CSV
This menu option allows the user to save the current configuration
of the Enclosures Zoning in a file. This file is compatible with the
CLI commands that require a csv file to update Zoning. The csv
file format for PHY Based and T10 Zoning are identical therefore
one use of this command is to dump a MIS system’s T10 Zoning
configuration and then rewrite the same file as a PHY Based
configuration.
8) Display StorBrick(s)
Settings
This menu option simply displays information about the selected
StorBricks such as their SAS Addresses.
9) Display CLI Settings This menu option simply displays information about the CLI
Zoning Tool that is generally contained in the ini file but may
be changed in the course of operating the CLI Zoning Tool
through CLI commands.
10) Enter StorBrick CLI
(Must select a single
StorBrick)
This menu option is for Diagnostics support only.
11) Reboot FanBase -
will not reset
StorBricks
This menu option will invoke a soft reboot of the FanBase and no
StorBricks shall be reset.
Table 3-2 CLI Zoning Tool Menu Options and Descriptions (continued)
Menu Option Description
116 007-5818-003
3: System Software
Note: Options 10, 13 and 14 are to be used ONLY with the assistance of technical support.
Here are the valid Command Line commands:
usage: ShackCLI.py [-h] --cmd CMD [--sbcmd SBCMD] [--target TARGET]
[--sb SB] [--ini INI] [--zcsv ZCSV] [--pcsv PCSV]
[--zones ZONES] [--delay DELAY] [--pbcsv PBCSV]
[--mistype MISTYPE] [--v, V]
12) Reset FanBase -
Danger. This will reset
StorBricks
This menu option will invoke a cold start of the FanBase and all
StorBricks shall be reset.
13) Force Storbrick(s)
into Debug Mode
This menu option places the selected StorBrick(s) into debug
mode which allows most of the other menu options to work. At
start of day all StorBrick(s) defined in the ini file will
automatically be put in Debug Mode therefore it is not generally
necessary to run this command.
14) Exit Storbrick(s)
Debug Mode
Has no affect at this time and is reserved for future
implementations.
15) Display PHY Error
Counters for selected
StorBricks
This menu option displays all the PHY Error Counters for the
Selected StorBrick(s).
16) Display PHY
Infomation for selected
StorBricks
This menu option displays all the PHY information for the
selected StorBrick(s). this information includes the PHY’s
connected link rate, Zone Group and SAS/SATA connection
types.
17) Display StorBrick
UUID for selected
StorBricks
This menu option displays the UUID for the selected
StorBrick(s). The UUID is displayed in human readable form.
18) Display StorBrick
Firmware Revision Levels
for Selected StorBricks
This menu option displays all the Firmware Revision Level
information for the selected StorBrick(s).
0) Exit CLI - back to
command prompt
This menu option will exit the CLI Zoning Tool.
Table 3-2 CLI Zoning Tool Menu Options and Descriptions (continued)
Menu Option Description
CLI Zoning Tool
007-5818-003 117
MIS Command Line Interface optional arguments:
-h, --help show this help message and exit
--cmd CMD command to execute
--sbcmd SBCMD Command to send to SB
--target TARGET target ip address (if not 192.168.0.3)
--sb SB Index(s) of StorBrick
--ini INI Initialization file
--zcsv ZCSV CSV file name for Zoning PHY Groups
--pcsv PCSV CSV file name for Zoning Permission Tables
--zones ZONES Maximum Permission Table Entries
--delay DELAY FanBase delay before expect StorBrick result
--pbcsv PBCSV CSV file name for Phy Based Zoning
--mistype MISTYPE JBOD or SERVER
--v, V Display the version of the Shack CLI
In the menu above, this executes option 2:
python ShackCLI.py –ini ShackCLI.ini –cmd dzone
This executes option 3:
python ShackCLI.py –ini ShackCLI.ini –cmd uperm
This executes option 4:
python ShackCLI.py –ini ShackCLI.ini –cmd uphyz
This executes option 5:
python ShackCLI.py –ini ShackCLI.ini –cmd uphyb
Zoning Using CLI Zoning Tool
First, select option 2 from the Main Menu to display the current zoning. When displayed, 0
indicates that the path to that drive is disabled. An x indicates that it is active to that SAS path.
For the MIS-Server, Initiator0 = SAS lane 0 which is connected to HBA-A lane 0,
initiator1 = SAS lane 1 which is connected to HBA-B lane 0, initiator2 =
SAS lane 2 which is connected to HBA-C lane 0, initiator3 = SAS lane 3 which
is connected to HBA-D lane 0. Initator4-7 are not used at this time and will be set to 0.
Since the .csv data is organized by drives, and the drives are numbered by StorBrick, we can
start at SB[0:0]. The drive numbers in an MIS Enclosure are numbered from 0-143 for the MIS
Server and 0-162 for the MIS JBOD.
Example (truncated, lines 021–190 & 195–250, for space):
StorBrick 0 Permission Table
118 007-5818-003
3: System Software
2 2 2 1 1 1 1 1 1
Zone 5 2 2 9 9 6 5 2 2 9 9 6 6 3 3 0
Group 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0
000 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
001 ffffffff ffffffff ffffffff ffffffff ffffffff ffffffff ffffffff ffffffff
002 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000ff02
003 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000ff02
004 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
005 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
006 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
007 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
008 00000000 00000000 44444444 44444444 00000000 88111111 11111111 1111010e
009 00000000 00000000 08888888 88888888 00000001 11222222 22222222 2222020e
010 00000000 00000000 11111111 11111111 00000000 22444444 44444444 4444040e
011 00000000 00000000 22222222 22222222 00000000 44888888 88888888 8888080e
012 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000000e
013 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000000e
014 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000000e
015 00000000 00000000 00000000 00000000 00000000 00000000 00000000 0000000e
016 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000102
017 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000202
018 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000402
019 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000802
020 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000102
191 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
192 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
193 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
194 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
195 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
196 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
251 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
252 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
253 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
254 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
255 00000000 00000000 00000000 00000000 00000000 00000000 00000000 00000002
Phy Zoning for StorBrick 0
Requested Inside
Inside ZPSDS
PHY Group Persistence ZPSDS Persistent
00 220 Yes No No
01 220 Yes No No
02 220 Yes No No
03 220 Yes No No
04 011 Yes No No
05 010 Yes No No
06 009 Yes No No
07 008 Yes No No
08 017 Yes No No
09 089 Yes No No
10 088 Yes No No
11 016 Yes No No
12 091 Yes No No
13 019 Yes No No
14 094 Yes No No
15 022 Yes No No
16 092 Yes No No
17 020 Yes No No
CLI Zoning Tool
007-5818-003 119
18 095 Yes No No
19 023 Yes No No
20 024 Yes No No
21 096 Yes No No
22 021 Yes No No
23 093 Yes No No
24 090 Yes No No
25 018 Yes No No
26 220 Yes No No
27 220 Yes No No
28 220 Yes No No
29 220 Yes No No
30 220 Yes No No
31 220 Yes No No
32 220 Yes No No
33 220 Yes No No
34 220 Yes No No
35 220 Yes No No
When ready to zone, complete the following instructions.
1. Execute option 7 to make a copy of the current configuration (be sure to add .csv as
the file extension). Example: MIS-System1-zoning-092012-121103.csv
2. Edit the .csv file to the desired zone configuration. Once satisfied, save the file with a
different name so as not to over write the saved one.
3. Change the name in the ShackCLI.ini file to point it to the new file.
4. Select the update option that fits the configuration (option 3, 4, or 5).
5. Select option 12 to reset the StorBricks and invoke the zoning changes.
6. Verify that the zoning is correct by executing option 2 and reviewing the configuration
file.
7. Power cycle the MIS server and reboot any head-of-string controller of a JBOD to
refresh the information in the servers.
Once power-cycling is complete, verify the changes by executing the command:
python ShackCLI.py –ini ShackCLI.ini –cmd dzone.
If satisfied, disconnect from the MIS-S9D network interface, and slide the chassis back into the
rack.
Editing the .csv File for the CLI Zoning Tool
A configuration file made of comma separated values (a “csv” file) is used by the CLI application
to zone the StorBricks. A set of standard configuration files are included with the software
120 007-5818-003
3: System Software
package. A custom file can be created, using a spreadsheet application and then saving it as a
.csv file.
Figure 3-67 is a block diagram of MIS JBOD StorBrick SB0. The other 8 StorBrick for the MIS
JBOD repeat this, but are offset to other lanes from the I/O modules.
Table 3-3 gives a portion of a csv file. A 1 in the spreadsheet indicates that the drive in question
is accessible by the above HBA. The 0 indicates that the path is disabled.
Table 3-3 Zone Group Implementation
SB0
Drive
HBA-A
Indicator 1
HBA-B
Indicator 2
HBA-C
Indicator 3
HBA-D
Indicator 4
N/U
Indicator 5
N/U
Indicator 6
N/U
Indicator 7
N/U
Indicator 8
010000000
101000000
200100000
300010000
410000000
501000000
600100000
700010000
810000000
CLI Zoning Tool
007-5818-003 121
Figure 3-67 Block diagram of MIS-Server StorBrick SB0
122 007-5818-003
3: System Software
Disk RAID Support
The MIS Platform supports both software and hardware RAID, standard and nested. Disk
performance is improved because more than one disk can be accessed simultaneously. Fault
tolerance is improved because data loss caused by a hard drive failure can be recovered by
rebuilding missing data from the remaining data or parity drives.
The MIS Platform supports the following RAID levels:
RAID 0 (striping without parity or mirroring, Figure 3-68)
RAID 1 (mirrored without parity or striping, Figure 3-69)
RAID 5 (striping with parity, Figure 3-70)
RAID 6 (striping with dual parity, Figure 3-71)
RAID 00 (spanned drive group striped set from a series of RAID 0 drive groups,
Figure 3-72)
RAID 10 (mirrored stripes across spanned drive groups, Figure 3-73)
RAID 50 (distributed parity and striping across spanned drive groups, Figure 3-74)
RAID 60 (distributed parity, with two independent parity blocks per stripe in each RAID set,
and disk striping across spanned drive groups, Figure 3-75)
The onboard MIS server zoning application is supported in both Windows and Linux operating
systems. There is also an external FanBase CLI, where a system running either Windows or Linux
can be used to zone the JBOD in addition to MIS Servers. When LSI MegaRAID HBAs are used,
they support RAID 0, RAID 1, RAID 5 and RAID 6, along with their variants. Where ever
possible, the zoning and RAID selection for a MIS server should provide for the maximum
amount of availability in the event of a StorBrick failure. The StorBrick has four SAS lanes into
it from up to four SAS RAID HBAs. These SAS lanes are inputs to a SAS expander that connects
to the drives installed in the StorBrick. Since there are eight StorBrick in the MIS-Server, each
with nine drives, the highest availability for the data would be RAID groupings that span the
StorBricks. More information and examples provided in “RAID Configuration Notes” on
page 129.
Important: Unless specified, all systems ship as RAID 6.
Disk RAID Support
007-5818-003 123
RAID 0
Figure 3-68 RAID 0
A RAID 0 splits data evenly across two or more disks (striped) without parity information for
speed. RAID 0 provides no data redundancy. It provides improved performance and additional
storage, but no fault tolerance. RAID 0 is normally used to increase performance, although it can
also be used as a way to create a large logical disk out of a two or more physical ones. RAID 0
provides high data throughput, especially for large files in an environment that does not require
fault tolerance.
RAID 0 is useful for setups such as large read-only NFS servers, where mounting many disks is
time-consuming or impossible and redundancy is irrelevant. Any drive failure destroys the array,
and the likelihood of failure increases with more drives in the array (at a minimum, catastrophic
data loss is almost twice as likely compared to single drives without RAID). A single drive failure
destroys the entire array because when data is written to a RAID 0 volume, the data is broken into
fragments called blocks. The number of blocks is dictated by the stripe size, which is a
configuration parameter of the array. The blocks are written to their respective drives
simultaneously on the same sector. This allows smaller sections of the entire chunk of data to be
read off each drive in parallel, increasing bandwidth. RAID 0 does not implement error checking,
so any error is uncorrectable. More drives in the array means higher bandwidth, but greater risk
of data loss.
124 007-5818-003
3: System Software
RAID 1
Figure 3-69 RAID 1
A RAID 1 writes identically to two drives, thereby producing a “mirrored set” (at least two drives
are required to constitute such an array). This is useful when read performance or reliability is
more important than data storage capacity. Such an array can only be as big as the smallest
member disk. A classic RAID 1 mirrored pair contains two disks, which increases reliability
geometrically over a single disk. Since each member contains a complete copy of the data, and
can be addressed independently, ordinary wear-and-tear reliability is raised by the power of the
number of self-contained copies.
The array continues to operate as long as at least one drive is functioning. With appropriate
operating system support, there can be increased read performance, and only a minimal write
performance reduction; implementing RAID 1 with a separate controller for each drive in order to
perform simultaneous reads (and writes) is sometimes called multiplexing (or duplexing when
there are only two drives). RAID 1 is good for applications that require small capacity, but
complete data redundancy. (The server boot drive modules are configured for RAID 1.)
Disk RAID Support
007-5818-003 125
RAID 5
Figure 3-70 RAID 5
A RAID 5 uses block-level striping with parity data distributed across all member disks, creating
low-cost redundancy. In most implementations, a minimum of three disks is required for a
complete RAID 5 configuration. A concurrent series of blocks—one on each of the disks in an
array—is collectively called a stripe. The disk used for the parity block is staggered from one
stripe to the next, hence the term distributed parity blocks. RAID 5 writes are expensive in terms
of disk operations and traffic between the disks and the controller. RAID 5 striping and distributed
parity data across all drives provides high data throughput, especially for small, random access.
Fault tolerance is maintained by ensuring that the parity information for any given block of data
is placed on a drive separate from those used to store the data itself. The performance of a RAID
5 array can be “adjusted” by trying different stripe sizes until one is found that is well-matched to
the application being used.
RAID 6
Figure 3-71 RAID 6
126 007-5818-003
3: System Software
RAID 6 (block-level striping with double distributed parity) extends RAID 5 by adding an
additional parity block. RAID 6 does not have a performance penalty for read operations, but it
does have a performance penalty on write operations because of the overhead associated with
parity calculations. Any form of RAID that can continue to execute read and write requests to all
of a RAID array's virtual disks in the presence of any two concurrent disk failures. This makes
larger RAID groups more practical, especially for high-availability systems. This becomes
increasingly important as large-capacity drives lengthen the time needed to recover from the
failure of a single drive. Single-parity RAID levels are as vulnerable to data loss as a RAID 0 array
until the failed drive is replaced and its data rebuilt; the larger the drive, the longer the rebuild
takes. Double parity gives additional time to rebuild the array without the data being at risk if a
single additional drive fails before the rebuild is complete. Like RAID 5, a single drive failure
results in reduced performance of the entire array until the failed drive has been replaced and the
associated data rebuilt.
RAID 00
Figure 3-72 RAID 00
A RAID 00 drive group is a spanned drive group that creates a striped set from a series of RAID
0 drive groups. RAID 00 does not provide any data redundancy, but along with RAID 0, RAID 00
offers the best performance of any RAID level. RAID 00 breaks up data into smaller segments and
stripes the data segments across each drive in the drive groups. The size of each data segment is
determined by the stripe size. RAID 00 offers high bandwidth.
Note: RAID level 00 is not fault tolerant. If a drive in a RAID 0 drive group fails, the whole virtual
drive (all drives associated with the virtual drive) fails.
Disk RAID Support
007-5818-003 127
By breaking up a large file into smaller segments, the RAID controller can use both SAS drives
and SATA drives to read or write the file faster. RAID 00 involves no parity calculations to
complicate the write operation, which makes RAID 00 ideal for applications that require high
bandwidth but do not require fault tolerance. Figure 3-72 shows an example of a RAID 00 drive
group.
RAID 10
Figure 3-73 RAID 10
RAID 10 is a combination of RAID 0 and RAID 1 and consists of stripes across mirrored spans.
RAID 10 breaks up data into smaller blocks and mirrors the blocks of data to each RAID 1 drive
group. The first RAID 1 drive in each drive group then duplicates its data to the second drive. The
size of each block is determined by the stripe size parameter, which is set during the creation of
the RAID set. The RAID 1 virtual drives must have the same stripe size.
Spanning is used because one virtual drive is defined across more than one drive group. Virtual
drives defined across multiple RAID 1 level drive groups are referred to as RAID level 10, (1+0).
Data is striped across drive groups to increase performance by enabling access to multiple drive
groups simultaneously.
Each spanned RAID 10 virtual drive can tolerate multiple drive failures, as long as each failure is
in a separate drive group. If there are drive failures, less than the total drive capacity is available.
128 007-5818-003
3: System Software
RAID 50
Figure 3-74 RAID 50
A RAID 50 combines the straight block-level striping of RAID 0 with the distributed parity of
RAID 5. This is a RAID 0 array striped across RAID 5 elements. It requires at least 6 drives. RAID
50 improves upon the performance of RAID 5 particularly during writes, and provides better fault
tolerance than a single RAID level does. This level is recommended for applications that require
high fault tolerance, capacity and random positioning performance.
As the number of drives in a RAID set increases, and the capacity of the drives increase, this
impacts the fault-recovery time correspondingly as the interval for rebuilding the RAID set
increases. RAID 50 works best with data that requires high reliability, high request rates, high data
transfers, and medium-to-large capacity.
RAID 60
Figure 3-75 RAID 60
Disk RAID Support
007-5818-003 129
A RAID 60 combines the straight block-level striping of RAID 0 with the distributed double parity
of RAID 6. That is, a RAID 0 array striped across RAID 6 elements. As it is based on RAID 6,
two disks from each of the RAID 6 sets could fail without loss of data. Also failures while a single
disk is rebuilding in one RAID 6 set will not lead to data loss. RAID 60 has improved fault
tolerance, any two drives can fail without data loss and up to four total as long as it is only two
from each RAID 6 sub-array.
Striping helps to increase capacity and performance without adding disks to each RAID 6 set
(which would decrease data availability and could impact performance). RAID 60 improves upon
the performance of RAID 6. Despite the fact that RAID 60 is slightly slower than RAID 50 in
terms of writes due to the added overhead of more parity calculations, when data security is
concerned this performance drop may be negligible.
RAID Configuration Notes
To get the best availability, treat each drive brick as an enclosure. For a RAID 1 with only one
drive per StorBrick in the group, the loss of a StorBrick does not affect the availability of the data
(Figure 3-76).
Figure 3-76 RAID 1 with one drive per StorBrick
However, with two drives spanning a StorBrick, the loss of a StorBrick will cause the data to be
unavailable until the failed storbrick is replaced (Figure 3-77).
130 007-5818-003
3: System Software
Figure 3-77 RAID 1 with two drives spanning a StorBrick
For redundancy, choose one drive from each drive brick as you build the LUNs. Configure the
drives in a RAID group to span the StorBricks. For instance, configuring 8 drives from StorBrick
0 as a RAID5, 7+1 group will work. However, if that StorBrick (StorBrick 0) fails, all 8 drives
will become inaccessible, making that RAID group’s data unavailable until the StorBrick is
replaced.
If, however, drive 0 from each StorBrick (SB0-0, SB1-0, SB2-0, SB3-0, SB4-0, SB5-0,
SB6-0 and SB7-0) is used to make up the RAID 5, 7+1 group, and any StorBrick were to fail,
only one drive of that group would be affected. The RAID 5 algorithm would be able to recover
the data from the remaining drives and the RAID group would remain available. Configurations
varies based on needs for capacity versus protection.
There is also the option to assign both a RAID group dedicated spare drive or a global spare drive.
For example, a RAID 5 group could be a 6+1 with a dedicated spare drive. In a full up system you
would have 9 sets of these RAID groups, all spread across the StorBricks. One method for greater
capacity is available by configuring 8 groups of 7+1, one group of 6+1 and a global spare. In a
RAID 6 6+2, a spare my not be desirable at all, as one is already part of the group automatically.
For configurations that maximize storage, can allow unavailability of data for a time and/or predict
a high-success rate for StorBricks, very large RAID groups may be desirable. For typical RAID
usage, RAID 5 7+1 or RAID 6 6+2 are most likely.
Warning: Do not configure multiple drives in a RAID group to be on the same StorBrick.
Since there are eight StorBrick in the MIS-Server, each with nine drives, the highest availability
for the data would be RAID groupings that span the StorBricks. For example, a RAID 6
configuration should have eight drives, one from each StorBrick. In this configuration, if a
StorBrick failed, the data would still be available. For RAID 5 or RAID 6, with one drive per
StorBrick in a group, the loss of a StorBrick (SB3) does not affect the availability of the data
(Figure 3-78).
Disk RAID Support
007-5818-003 131
Figure 3-78 RAID 5 or 6 with one drive per StorBrick
If a larger RAID group is desired, it would have to have multiple drives on a StorBrick. Then if a
StorBrick were to fail, two drives would be unavailable. If this were a RAID 6 implementation,
the data would still be available, though another StorBrick failure or even a drive failure in that
group would cause a loss of data availability. If this were a RAID 5 implementation, the data
would become unavailable until the failed StorBrick is replaced (Figure 3-79).
Figure 3-79 Loss of a drive with multiple drives on a StorBrick does not affect RAID 6, but will impact
RAID 5
For RAID 6 with three drives of the group spanning a StorBrick, the data is unavailable until the
failed StorBrick is replaced (Figure 3-80).
Figure 3-80 Three drive loss in RAID 6 require StorBrick replacement
132 007-5818-003
3: System Software
For more on RAID configuration, see the MegaRAID guides on servinfo (http://servinfo.corp/) or
the Technical Publications Library (http://docs.sgi.com). Operating System software RAID
support: Windows Dynamic Disks, Linux mdadm, or RAID-Z.
007-5818-003 133
Chapter 4
4. System Maintenance
For warranty and safety considerations, SGI designates the following chassis components as
customer-replaceable units (CRUs):
Power supplies
•Fans
Disk drives
These components are all hot-swappable; that is, you can replace them without powering down
the storage server. A trained service technician should install and replace all other components.
This chapter describes how you replace the CRUs and check the system airflow:
“Detecting Component Failures” on page 134
“Sliding the Chassis Forward/Backwards” on page 134
“Removing the Front or Rear Chassis Cover” on page 134
“Replacing a Power Supply” on page 135
“Replacing a Fan Module” on page 136
“Replacing a Disk Drive” on page 137
“Checking the System Air Flow” on page 140
Tools Required: The only tool you will need to perform maintenance is a #1 and #2 Phillips
screwdriver.
Warning: Review the warnings and precautions listed in, “Important Information” on
page xix, before setting up or servicing this chassis.
!
134 007-5818-003
4: System Maintenance
Detecting Component Failures
In general, when a system component fails, the operating system/storage management system
(OS/SMS) receives an alert. The OS/SMS generates an alert to the monitoring application for your
storage server. The alerts include the system serial number, the suspect component, and a
summary of the fault. For most components, you should inform SGI service of the fault and
forward the information from the alerts.
In addition to the alerts, the control panel on the chassis front panel can indicate component
failures in the case of power supplies, fans, and drives. See Chapter 2, “System Interfaces”.
For more information about alert generation and management, see “Power Supply LEDs” on
page 29.
Sliding the Chassis Forward/Backwards
The cable management system of the MIS chassis allows it to be slid forward (20") or backwards
(18"). You will need to slide the chassis out to service some of its components. To slide the chassis
out, follow these steps:
1. Push the two release latches in, at the front and rear, towards the center of the chassis.
2. Pull the chassis out using the handles. The chassis will latch at the 20- or 18-inch limit.
3. To slide the chassis back in, depress the two release latches near the rail and slide it back in.
Removing the Front or Rear Chassis Cover
Important: When a chassis cover is removed, an intrusion sensor monitored by the SMS will
detect its removal. If the cover is off for more than 15 minutes or any system temperature sensor
exceeds its threshold limit, the server will perform an orderly shutdown and power-off.
As shown in Figure 4-1, the top of the chassis is bifurcated; that is, there is a front and rear chassis
cover. Except for power supply maintenance, all service actions require that you remove the front
or rear chassis cover. This section describes the steps.
Replacing a Power Supply
007-5818-003 135
1. To remove a chassis cover, first follow the instructions in “Sliding the Chassis
Forward/Backwards” on page 134.
2. Remove the single security screw from the cover.
3. Push the detent, and slide the cover out and up from the chassis.
Figure 4-1 Front & Rear Chassis Covers
Replacing a Power Supply
To replace a failed power supply, do the following:
1. Using the OS/SMS interface for your system, verify the fault (failed unit) and its location.
2. Locate the failed unit: it should have a lighted yellow service LED. See Figure 4-2.
3. Unplug the power supply that will be replaced.
4. Push the release tab on the back of the power supply.
5. Pull the power supply out using the handle.
136 007-5818-003
4: System Maintenance
6. Replace the failed power module with another of the same model.
7. Push the new power supply module into the power bay until it clicks into the locked position.
8. Plug the AC power cord back into the module and power up the server.
9. Once power supply is verified good, clear the service required status via the OS/SMS
interface.
Figure 4-2 Replacing a Power Supply
Replacing a Fan Module
To replace a fan module, do the following:
1. Using the OS/SMS for your system, verify the fault (failed unit).
2. Using the OS/SMS, set the system to a service state for the removal of the faulted fan. The
OS/SMS will turn off the fan module. It will then turn on the locator LED (blue) for that fan
module.
3. Remove the front chassis cover. (See “Removing the Front or Rear Chassis Cover” on
page 134.)
4. Locate the fan module with the illuminated blue LED (Figure 4-3).
Replacing a Disk Drive
007-5818-003 137
Figure 4-3 Replacing a Fan Module
5. Loosen the thumbscrew, pull out the faulted fan by pulling upward on both the front and rear
flanges, and replace it.
6. Once the fan module is replaced, seat the fan by pushing between the two leds until it seats.
7. Re-install the chassis cover and security screw.
8. Unlock the chassis from the extended position and push it back until it locks into the stowed
position.
9. Using your OS/SMS, return the system to a normal state and the new fan module will be
powered on.
Replacing a Disk Drive
Important: Empty drive carriers cannot be inserted into the storbricks, so slots without HDDs
will not have carriers.
!
138 007-5818-003
4: System Maintenance
To replace a failed disk drive:
1. Using the OS/SMS for your system, verify the fault (failed unit).
2. Using the OS/SMS, set the system to a service state for the removal of the faulted drive. The
OS/SMS will turn off the drive. It will then turn on the locator LED (blue) for that drive.
3. Remove the chassis cover. (See “Removing the Front or Rear Chassis Cover” on page 134.)
4. Locate the faulted drive with the illuminated blue LED and remove it from its StorBrick (or
boot drive bay). (See “Removing the Drive” on page 138.)
5. Replace the faulted drive. (See “Re-installing the Drive” on page 139.)
6. Once the drive is replaced, re-install the chassis cover and security screw.
7. Unlock the chassis from the extended position and push it back until it locks into the stowed
position.
8. Using the OS/SMS, return the system to a normal state. The new drive will be powered on.
9. Using the OS/SMS, clear the service required status. At this time the rebuild or mirroring of
the data to the new drive will begin.
Removing the Drive
As shown in Figure 4-4, the drives are mounted in driver carriers to simplify their installation and
removal from the drive bricks or boot drive bays in the chassis.
To remove the drive, perform the following steps:
1. Ensure that the drive LEDs are off (except the blue locator LED), indicating that the drive is
not in use and can be removed.
2. Unlatch the drive carrier by sliding the grey latch toward the drive and pull the drive carrier
out of the StorBrick or boot drive bay.
3. Remove the four screws that secure the drive to the drive carrier.
4. Remove the drive from the carrier.
Replacing a Disk Drive
007-5818-003 139
Figure 4-4 Hard Drive Carrier
Re-installing the Drive
To re-install a hard drive into the hard drive carrier, perform the following steps:
1. Place the hard drive carrier on a flat, stable surface such as a desk, table, or work bench.
2. Slide the hard drive into the carrier with the printed circuit board side facing down.
3. Carefully align the mounting holes in the hard drive and the carrier.
Make sure the bottom of the hard drive and bottom of the hard drive carrier are flush.
4. Secure the hard drive using the four screws (see Figure 4-4).
5. Replace the drive carrier into the chassis.
6. Push the drive carrier down to lock it place.
140 007-5818-003
4: System Maintenance
Checking the System Air Flow
To check the air flow for an MIS enclosure, perform the following steps:
1. Remove the chassis cover. (See “Removing the Front or Rear Chassis Cover” on page 134)
2. Remove the midspan: unscrew the Phillips screws from either end of the brace, and lift away
the brace.
3. Make sure there are no objects wires or foreign objects obstruct air flow through the chassis.
Pull all excess cabling out of the airflow path.
Figure 4-5 MIS Chassis Midspan Support Brace
007-5818-003 141
Chapter 5
5. Troubleshooting
This chapter describes troubleshooting for the problems listed below. Chapter 2, “System
Interfaces,” describes use of the control panel to monitor the overall system status and the status
of specific components. Chapter 4, “System Maintenance,”describes how to replace defective
components.
“No Video” on page 141
“Losing the System’s Setup Configuration” on page 141
“I/O Time-outs and MegaRAID Drivers” on page 142
“Safe Power-Off” on page 142
For help beyond what is mentioned in this document, see “Product Support” on page xxv.
No Video
If the power is on but there is no video, remove all add-on cards and cables. Use the speaker to
determine if any beep codes exist. Refer to Appendix B, “BIOS Error Codes” for details.
Losing the System’s Setup Configuration
Make sure that you are using a high quality power supply. A poor quality power supply may cause
the system to lose the CMOS setup information. If this does not fix the Setup Configuration
problem, contact your vendor for repairs.
142 007-5818-003
5: Troubleshooting
I/O Time-outs and MegaRAID Drivers
To avoid I/O time-outs with certain workloads, the megaraid_sas driver needs to have the
poll_mode_io variable set to 1. For Novell operating systems on SGI InfiniteStorage servers,
the file /etc/modprobe.conf.local needs the following line added:
options megaraid_sas poll_mode_io=1
This modification will be made on systems shipped from the factory, but if a system is installed
or upgraded in the field, this change will have to be made after installation/upgrade.
Safe Power-Off
There are several safe power-off methods for an MIS Server or JBOD. They include,
Using the OS GUI power-off button at the console screen, if a keyboard/mouse/video
monitor is connected.
Pushing and holding the Power button on the front panel (see Figure 2-1 on page 26).
When logged in via an ssh session and executing a “shutdown” or “poweroff”
command.
When logged in to the BMC and using the power control page to power off the sever.
Using the remote console screen GUI power-off button, if a KVM RMM4Lite session is
established through the BMC.
If the platform is an MIS dual-server and both servers are powered up, performing the above steps
only powers off the server with which you are working. The fans, drives and second server will
remain powered on until the second server is powered off, then all power (but standby) will be
turned off.
For a JBOD Unit, the power button on the front panel will turn off the power to that I/O module.
If a second module is installed and powered on, it, the fans and the drives will remain on until it,
too, is powered off.
007-5818-003 143
Appendix A
A. Technical Specifications
Table A-1 describes the technical specifications for the SGI MIS platform.
Table A-1 Technical Specifications
Attribute Specification
Overview
Profile 4U Standard-depth chassis
Product type SGI MIS Server Platform (single or dual server), or SGI MIS JBOD
Unit
Connectivity Up to four SGI MIS JBOD units per SGI MIS Dual Server
Mount –Standard 19-inch rack-compatible rail mount (weight-dependent)
–SGI 19-inch Destination Rack (D-Rack), 42U
–Up to 10 chassis per D-Rack
Chassis Dimensions
Height 6.94" (176 mm)
Width 16.9" (429.2 mm)
Depth 36" (914.4 mm)
Max weight 220 lbs.
Power
AC Input 100–240 VAC (50-60Hz), single or three phase
Safety –UL/CSA certified to UL6050-1
–CE/CB certified to EN60950/IEC60950
EMC –North America FCC Class A
–Europe EN55022/EN55024
Operating Environment
144 007-5818-003
A: Technical Specifications
Operating
temperature
range
–41º to 95º F (5º to 35º C)
–processor cores automatically allowed to run faster than the base
operating frequency, if the cores are operating below power, current,
and temperature specification limits (< 35ºC ambient)
Non-operating
temperature
range
-40º to 140º F (minus 40º to 60º C)
Operating
humidity range
10% to 90% non-condensing
Non-operating
Humidity
10% to 95% non-condensing
SGI MIS Server Specifications
Servers/System –One or two server modules per system
–Single- or dual-socket processors per server
Processor support –Supports Intel® Xeon® E5-2600 series processors
–Supports Intel Turbo Boost Technology 2.0
Max cores 16 per server (32 per system)
Memory –Up to 8 DDR3 DIMMs (4 GB, 8 GB, or 16 GB) for a single-server
motherboard configuration
–Up to 16 DIMMs for a dual-server motherboard configuration.
–Max 128GB per server
Boot drives –Two per server, mirrored using RAID 1
–3.5" or 2.5" (15mm or 9.5mm)
–SAS or SATA
–Rotational or SSD
–Up to 300 GB
Supported
operating
systems
–RHEL 6.2
–SLES 11 SP1
–VMware ESX 5.0
–Windows 2008 R2 SP1
Networking Up to four user-specified PCIe HBAs, full-height (4.25") and half-depth
(3.375"), externally or internally facing
Table A-1 Technical Specifications (continued)
Attribute Specification
007-5818-003 145
Expansion slots Single server: 6 x PCIe gen 2
Dual server: 4 x PCIe gen 2 per server (8 total)
RAID controllers –8 to 32 SAS ports via 8-port PCIe cards
–Support for RAID 0, 1, 5, 6, 00, 10, 50, and 60
External storage
attachment
Up to 4 SGI MIS JBOD Units via PCIe cards
Internal storage –Up to 81 SAS or SATA 15mm, 2.5" or 3.5" drives
–Up to 162 SAS or SATA 9.5mm, 2.5" drives
–Drive size and capacity can be mixed in groups of 8
–Supported internal drives: SAS or SATA, rotational or SSD
SGI MIS JBOD Specifications
Internal Storage –Up to 81 SAS or SATA 15mm, 2.5" or 3.5" drives
–Up to 162 SAS or SATA 9.5mm, 2.5" drives
–Drive type and size can be mixed in groups of 9
–Supported internal drives: SAS or SATA, rotational or SSD
Connectivity 8 or optional 16 SAS ports
Table A-1 Technical Specifications (continued)
Attribute Specification
007-5818-003 147
Appendix B
B. BIOS Error Codes
The BMC may generate beep codes upon detection of failure conditions. Beep codes are sounded
each time the problem is discovered (for example, on each power-up attempt) but are not sounded
continuously. Common supported codes are listed in Table B-1.
In Table B-1, each digit in the code is represented by a sequence of beeps whose count is equal to
the digit.
Table B-1 BMC Beep Codes
Beep Code Reason for Beep Associated
Sensors Supported
1-5-2-1 No CPUs installed or first CPU socket is
empty.
CPU Missing sensor Yes
1-5-2-4 MSID Mismatch. MSID Mismatch
sensor.
Yes
1-5-4-2 Power fault: DC power is unexpectedly
lost (power good dropout).
Power unit—power
unit failure offset.
Yes
1-5-4-4 Power control fault (power good assertion
time-out).
Power unit—soft
power control failure
offset.
Yes
1-5-1-2 VR Watchdog Timer sensor assertion VR Watchdog timer
1-5-1-4 The system does not power on or
unexpectedly powers off and a power
supply unit (PSU) is present that is an
incompatible model with one or more
other PSUs in the system.
PSU status
007-5818-003 149
Appendix C
C. Zone Permission Groups Rules
There three group controls within the expanders that must be set. These are Allow Broadcast,
Allow Zoning and Phy (Physical) Changes, and Unassigned Slot Group. The Allow Broadcast
bit is set to that the sever can send and receive broadcasts. The Allow Zoning and Phy Changes
bit is set so that hot-swapping drives is allowed. The Unassigned Slot Group bit is set so that
drives that are yet to be zoned are still recognized by the system.
The first 8 bits are the Master Group. The first bit (0) has access to bit 1 and nothing else. This is
why the Master Group 0 is always set to 0, to avoid an extra step. Master Group 1 is all FF—it
“sees” everyone. Master Groups 2 & 3 talk to all the Initiators (bits 8-15), and currently, Master
Groups 4-7 point up to Master Group 1 (this may change in the future).
The next 8 bits are the Initiator Groups. Bit 8 through 11 corresponds to Adapter 1, Paths 0
through 3 and bits 12-15 correspond to Adapter 2, Paths 4 through 7. The next 16 through 24 bits
are the zone permission groups, with the 25th bit always set for the unassigned slot. The remaining
bits are reserved for future use. See Figure C-1 and Figure C-2 for examples.
Figure C-1 Zone Permission Groups – Example 1
007-5818-003 151
Figure C-2 Zone Permission Groups – Example 2
152 007-5818-003
C: Zone Permission Groups Rules

Navigation menu