Dell EMC PowerEdge MX Modular Platform Installation, Configuration, and Management - Participant Guide
Dell EMC PowerEdge MX Modular Platform Installation, Configuration, and Management - Participant Guide
Dell EMC PowerEdge MX Modular Platform Installation, Configuration, and Management - Participant Guide
MX MODULAR
PLATFORM
INSTALLATION,
CONFIGURATION, AND
MANAGEMENT
Version 1.0
PARTICIPANT GUIDE
PARTICIPANT GUIDE
Dell Confidential and Proprietary
Copyright © 2019 Dell Inc. or its subsidiaries. All Rights Reserved. Dell, EMC and other
trademarks are trademarks of Dell Inc. or its subsidiaries. Other trademarks may be
trademarks of their respective owners.
Servers ...................................................................................................................... 44
Storage...................................................................................................................... 57
MX5016s Storage Sled ...................................................................................................... 58
MX5016s layout ................................................................................................................. 59
Storage Fabric ................................................................................................................... 60
Fabric C Switches .............................................................................................................. 61
MX5016s Simplified Block Diagram ................................................................................... 62
Compute Sled Fabric C Mini-Mezz ..................................................................................... 63
PERC/HBA Cards .............................................................................................................. 64
PERC H745P MX Enclosure Support ................................................................................. 65
Drive-Assigned Storage Configuration ............................................................................... 66
Enclosure-Assigned Storage Configuration ........................................................................ 67
Accessing Sled Insertion/Removal Handle ......................................................................... 68
MX5016s Service Tag ........................................................................................................ 69
IDSDM and BOSS ............................................................................................................. 70
Switches ................................................................................................................... 72
I/O Fabrics A and B............................................................................................................ 73
MX7000 Networking Overview ........................................................................................... 74
I/O Modules on Fabrics A and B ........................................................................................ 75
Networking IOMs ............................................................................................................... 76
MX Troubleshooting................................................................................ 306
Course Introduction
Course Overview
Course Objectives
The course objectives state the minimum levels of knowledge that you should
achieve after successfully completing this course.
Course Agenda
Introductions
MX-Series Introduction
Compute
With performance and capacity options to suit a wide range of demanding
workloads, the PowerEdge MX740c and MX840c compute sleds are well-equipped
to meet the changing data center needs.
PowerEdge MX740c
You can map individual drives to servers or share drives with multiple compute
sleds. This granular drive-level assignment lets you address databases of any
size without wasting space. Choose the appropriate PowerEdge RAID
Controller (PERC) or HBA option when mapping drives to individual or multiple
servers. Add a clustering operating system or file system of your choice when
mapping to shared drives.
Both the MX740c 2-socket and the MX840c 4-socket servers can share drives
with the MX5016s using the dedicated MX5000s SAS switch. Internal server
drives may be combined with up to seven MX5016s storage sleds on one
chassis for extensive scalability. Mix and match storage and compute nodes
within the eight available slots depending on your workload requirements.
Both the MX740c 2-socket and the MX840c 4-socket servers can share drives
with the MX5016s using the dedicated MX5000s SAS switch. Internal server
drives may be combined with up to seven MX5016s storage sleds on one
chassis for extensive scalability. Mix and match storage and compute nodes
within the eight available slots depending on your workload requirements.
High-availability features such as hot-pluggable drives and hot serviceable
extenders help prevent work interruptions. Further minimize downtime with Dell
Networking
The Dell EMC family of networking solutions provides a wide-range of top-of-rack
and embedded switching platforms for server internetworking. Using Open
Networking, Dell EMC leads the industry with the broadest set of software choices
and capabilities to match various use-cases. Dell EMC networking solutions
provide rock-solid dependability that is proven in the most demanding Enterprise
and Service Provider data center environments around the globe.
The new MX-Series Ethernet and Fibre Channel switches are open networking
switches for the MX7000 chassis. These new switches provide customers with a
highly scalable and cost-effective multichassis fabric architecture that uses a broad
set of open networking hardware, software and management options.
1. Modern design
– introducing a multichassis scalable fabric architecture that can grow with
your needs.
2. Open automation
– giving customers the power of open choice with a broad array of
hardware platform, operating systems and management services.
3. Agile transformation
– by offering optimum connectivity to servers and storage within the
chassis to accommodate demanding workloads and environments.
The initial set of MX-Series networking switches consist of the following:
MX7000 Chassis
Introduction
The PowerEdge MX7000 with its MX-series compute, storage and networking
components is the next-generation of modular architecture. Capable of tackling
both the traditional workloads of the past and the new and emerging workloads
of the future, the chassis features:
Support for 2S and 4S Skylake based servers with massive storage potential
(up to 6 local drives per 2S and 8 local drives per 4S server)
Support for 16-drive 12 Gb/s SAS storage sleds.
Comprehensive I/O options including 25Gb Ethernet, 12Gb SAS, and 32Gb
Fibre Channel.
Three I/O networking fabrics (two general purpose Fabrics A/B and one
storage specific Fabric C supporting either SAS or FC).
Hardware Comparison
M1000e MX7000
Local Chassis Up to 14x 2.5" iSCSI drives occupying 2 Up to 16x 2.5" DAS
Storage slots drives occupying 1 slot
PSUs 6x 3 KW 6x 3 KW
IS IS NOT
All new chassis, sled, and I/O Module Backwards compatible with prior
form factors generations
Let's begin by looking at the multiple components that make up the MX-series
architecture.
Here is a view of the front of the MX-series chassis. Keep in mind that this image
shows one of a multitude of possible configurations.
Chassis: 7U, 32-inches deep enclosure sharing power, cooling, and OpenManage
Enterprise – Modular Edition unified management.
drive bays.
Front features:
Numbering:
From the front, the sleds are numbered from 1 to 8, left to right. For double-
width sleds, the numbers are 1, 3, 5 and 7, left to right.
The PSUs are numbered 1–6 left to right.
The front 60 mm fans are numbered 1–4 from top to bottom.
Rear features:
Management Controller:
Redundant modules. Each module provides 2 RJ45 connectors, one micro USB
Type B, one System ID button, one status LED, one Power LED.
Cooling:
Five 80mm fan modules for sled cooling.
Optional LCD
Integrated into the far-left bezel of the MX-series chassis is an optional LCD panel
that can be applied for at-the-box management. This panel displays chassis status,
errors, and the initial IP address.
NOTE: The LCD touch panel is not a hot swappable module. Before you replace
the module, power off the enclosure and remove the AC power from the chassis.
The LCD touch panel enables you to scroll or swipe on the screen. The options
available on the LCD touch panel are as mentioned below:
Welcome Screen Enables you to select your native language and the
default LCD home page.
Alerts Enables you to view a list of all the critical and warning
alerts of the enclosure.
LCD Configuration Enables you to configure the LCD options such as View
and Modify, View only, Disabled, Present, and Not
present.
System Info Displays the Model number, Asset tag, and Service tag of
the enclosure.
The left ear is also where the optional OpenManage QuickSync hardware is
deployed, enabling IT administrators to connect to the chassis with a smartphone
or tablet using a Bluetooth, secured, wireless connection. With OpenManage
QuickSync, IT Staff can perform a range of tasks, including configuration and
diagnostics, all from a simple mobile application. IT staff can also manage the
MX7000 chassis using a traditional crash cart.
Traditional crash cart: Simple keyboard and display interface (no cable adapter):
access to chassis manager, CLI, and GUI services
Important Points
The right rack ear control panel contains a CMOS chip. In the event of a power
outage, the CMOS is used to store the basic configuration.
MM presents a connection option menu on the USB serial port to one of the
following:
System ID Button
The System ID button on the MM rear panel enables for identifying the system in a
rack. When pressed, the front panel of the chassis blinks in the identify pattern.
This button also serves as the Health status LED indicator, displaying blue for
health or blinking amber if the MM has a fault.
* - UART is a digital protocol that we use to transfer data between two devices. A
universal asynchronous receiver/transmitter (UART) is a block of circuitry
responsible for implementing serial communication. Essentially, the UART acts as
an intermediary between parallel and serial interfaces
Let us look at the chassis front again. At the bottom of the chassis front are six
3000W power supplies that provide not only enough power for 14G servers but for
future generations of servers. These can operate in various power supply or grid
redundancy modes.
The MX7000 Power Supplies use a new front-loading form factor that is developed
for the MX7000. Up to six PSUs plug into the front of the chassis. The PSUs link
directly with connectors on the Main DB. The AC connections come from the rear
of the chassis cabling to connections on the PDB.
The Platinum PSUs support 3000 W output with high line AC input. For low line
input the PSUs outputs approximately 1400 W. Mixed high line and low line AC
inputs in the same chassis are not supported. In the case this occurs, the MM
disables the PSUs with the lowest count that have different inputs.
If there are an equal number of high-line and low-line PSUs the high-line PSUs
should be enabled.
The supplies are set nominally to output 12.2 V DC at the output. There is no
remote sense that is implemented in the MX7000 distribution boards, therefore
careful attention to voltage plane layout and power distribution must be applied to
all board designs in the infrastructure.
Power supply redundancy ensures that if a power supply unit fails, there is enough
power for the chassis to continue running.
Grid redundancy uses two data center grids that are connected in such a way that
if an entire grid goes down, the chassis runs on the second power grid.
No Redundancy
This mode has no provision for redundancy - if a PSU fails the system may fault
due to excess current supplied by the remaining PSUs. This mode is typically used
in cases where the user does not have a sufficient number of PSUs to support the
installed configuration of compute sleds, IOMs, and so on.
Grid Redundancy
This mode provides for redundancy at the grid (i.e., branch) level. Configurations
for this mode are represented by “N+N,” where N is the number of power supplies
on each grid and the total available power is N * PSU capacity. Supported
configurations for this mode should include 1+1, 2+2 and 3+3.
Note: The power load is balanced between 2 grids.
PSU Redundancy
This mode provides for redundancy at the PSU level. Configurations for this mode
are represented by “N+1,” where N is the number of power supplies required to
support the configuration (the total available power is N * PSU capacity).
The PSUs are accessed from the front; however, the AC power cable connection is
at the rear of the chassis. There are six C22 inlet connectors, three on the rear left,
and three on the rear right side. Each PSU will also have an associated “AC good”
LED located next to its associated plug.
The MX7000 Power Supplies use a new front-loading form factor that is developed
for the MX portfolio.
Up to six PSUs plug into the front of the chassis. These connect directly to the main
distribution board. The AC connections come from the rear of the chassis cabling to
connections on the PDB. The Platinum PSUs support 3000 W output with high line
AC input. For low line input the PSUs output approximately 1400 W.
The supplies are set nominally to output 12.2 V DC at the output.
There are six C22 inlet power connectors located at the rear of the chassis, three
Important Points
Both 220 VAC and 110 VAC are supported but the chassis may still request/require
additional power depending upon installed component power needs.
The fans in the MX7000 are redundant and hot-swappable. This ensures cold air is
delivered to each component. Max supported fan power is 100 W.
The MX7000 has three distinct airflow paths, or zones, of the multiple airflow zone
infrastructure.
There are two different types of fan modules that are used in the MX7000 chassis,
60 mm fans in the front and 80 mm fans in the rear.
Each fan is enclosed in a mechanical assembly to make up the module.
The fan modules are hot-swappable, with 3+1 redundancy for the front fans, and
4+1 redundancy for the rear fans.
The system should always be populated with the full set of fans to support the
airflow requirements of the chassis.
The chassis is cooled from front to back, as with typical rack systems. The airflow
for the sleds and PSUs is shown below, as pictured from the side view of the
chassis. There are separate paths for the exhaust, keeping the PSU exit air
isolated from the sled exit air.
The sled air flow is controlled by the rear 80 mm fans, while the PSU air flow is
controlled by the PSU fans located toward the front of the PSUs.
The four front 60 mm fans provide cooling to the rear Fabrics A and B IOMs, Fabric
C IOMs, and the MMs.
The installation of Blank Filler Modules is only enforced through documentation and
factory build rules; there is no mechanism in the system to distinguish between the
presence of a Blank Filler Module or an empty slot.
Chassis Tub
The chassis consists of two major sections. The Main Chassis provides structure
for mounting PSUs, Sleds, and Front Fans.
The Rear Tub assembly includes the central bulkhead structure and mounting for
main chassis boards, AC plugs, and rear-end modules.
AC Status LED
x6 (not shown)
Flex
cable C22 AC
Busbars plug (x6)
FPC
signal
Right rack ear cable AC cables (x6)
Main
Right rack ear Right rack ear
PDB
cable interposer
High-speed IO for Fabrics A & B is routed directly from the Sled to the IOM via a
direct orthogonal connection.
The Main PDB handles other functions like power distribution and low-speed
signaling, other boards/cables, and FAB C.
The components that are within the Tub assembly are shown on the right of the
image.
The Main Distribution Board (DB) is the most complex of all the internal
infrastructure boards.
There are no active components, however there are many signals that are
interconnected on this board. The Main DB sits horizontally near the floor of the
chassis. From the front of the chassis, the six power supplies plug into
connectors on the bottom side of the Main DB. Also, from the front of the
chassis the eight compute sleds plug into this board for power, management
signals, and other fabric connections. From the rear of the chassis the
Management Modules plug in to this board to provide overall chassis
management and control.
Also, from the rear of the chassis the two fabric I/O Modules (Fabric C IOMs)
plug into this board.
The vertical power distribution board sits above the main distribution board,
which is oriented perpendicular to the chassis floor.
There are no active components, but many signals connections on this board.
This board connects to the main distribution board with a flex cable, and power
and ground bus bars.
From the front of the chassis, the four cooling fans plug into the vertical power
distribution board
Also, from the rear of the chassis the four fabric I/O Modules (Fab-A/B IOMs)
plug into this board for power, management signals, and other fabric
connections
A cable connects this board to the rear fan board.
At-the-Box Management
The MX-series ground breaking architecture has removed the need for separate
compute, storage, and networking management consoles used in the past. Instead,
MX-series management provides end-to-end lifecycle management of all
components with the embedded OpenManage Enterprise – Modular console.
Servers
Introduction
The PowerEdge MX740c compute sled is a two socket, full height, single width
sled. The PowerEdge MX740c provides a cost-effective balance of processor
cores, memory, and I/O making it ideal for virtualization environments, software-
defined storage and as a foundation for collaborative workloads.
This sled has 6x 2.5-inch drive bays in three rows containing two bays in each row.
All drive bays are SAS, SATA, and NVMe capable. The MX740c provides 24
memory slots.
1. Backplane
2. Backplane cable
3. Processor 1 (heat sink)
4. Processor 2 (heat sink)
5. Mezzanine card A1
6. Power connector
7. Mezzanine card B1
8. Mini Mezzanine connector
9. iDRAC card
10. BOSS connector
11. PERC connector
The PowerEdge MX840c compute sled is a four socket, full height, double width
sled. Its additional processing power, exceptionally large memory capacity, and
ability to address more local storage make it ideal for a range of database
applications, heavy transactional workloads, and enterprise applications.
This sled has 8x 2.5-inch drive bays: 4 at the top of the sled, an air intake in the
middle and four bays at the bottom of the sled. All drive bays are SAS, SATA, and
NVMe capable. The MX840c provides 48 memory slots.
iDRAC Module
Storage
Introduction
The MX5016s is a dense, highly flexible, scale-out storage sled with easy front
access that can provide flexible, customizable storage for all the servers in the
chassis.
A chassis can contain up to seven storage sleds taking up seven of the eight
chassis slots. The remaining slot in the chassis must be used for the required
compute sled. Drives can be individually mapped to one or more servers. This
granularity lets your customers add storage as needed without wasting space.
MX5016s layout
The MX5016s storage sled contain the status and activity LEDs for the individual
disk drives, and for the two expanders. The illustrations above show the
approximate location of the drive LEDs, which are physically on the Backplane
Board. The bi-color Green/Amber Status LEDs can be in one of seven different
states.
The Expander devices are the PMC PM8043B-F3EI 12 Gbps SAS expander. The
Expander device is mounted on the back side of the board, facing towards the
inside of the sled. The board incorporates a gold-finger card edge connector for
both signal and power. The signal and power connectors with which the gold finger
contacts connect are physically separate on the Backplane board. A third
connector is used to mate with the SAS Cable assembly.
Storage Fabric
Now we return to the chassis rear. Located below Fabric B is storage Fabric C with
2 side-by-side switches for redundancy. It has a small mid-plane to enable
connectivity from Fabric C to a storage network. Customers have the option of
using fabric C slots for two redundant SAS switches OR 2 redundant Fibre Channel
switches to connect to a storage network.
Fabric C Switches
Item Description
1 Information tag
5 Release lever
6 LED indicators
The illustration in this slide shows a simplified block diagram of the data path for
the MX5016s sled.
The MX7000 chassis Fabric-C interface supports two groups of x4 SAS lanes.
Each group connects to a SAS Expander inside the sled.
These SAS signals are routed through a flexible cable bundle from the Fabric-C
Board to the two SAS Expanders.
The power connector supplies 12V and Ground, as well as some control signals to
facilitate turning the sled ON/OFF and general communication with the sled.
Power and control signals are also passed through the flexible cable to the
Backplane.
The Expanders each connect to ports on all 16 HDDs. Note the first expander
connects to “port A” of the first eight HDDs and “port B” of the second eight HDDs.
The second expander is the opposite and connects to “port B” of the first eight
HDDs and port A of the other eight HDDs.
The Fabric C Mini-Mezz connects directly to the Min Distribution Board which in
turn connects to the Fabric C IOMs. Within the server, the Fabric C Mini-Mezz
connects to the system board and from there it connects to either the HBA or
PERC card.
PERC/HBA Cards
The Fabric C Mini-Mezz connects directly to the Min Distribution Board which in
turn connects to the Fabric C IOMs. Within the server, the Fabric C Mini-Mezz
connects to the system board and from there it connects to either the HBA or
PERC card.
As shown in this slide, the H745P MX resides in the compute sled and functions as
an interface to the SAS topology of the MX7000. The H745P MX is used to
configure drives that are located in the MX5016s storage sled into RAID volumes.
The H745P MX connects to a pair of MX5000s SAS I/O modules (IOMs) that
manage connections between the compute and storage sleds in the MX7000. For
more information about configuring drive assignments, see the OpenManage
Enterprise Modular User's Guide.
The H745P MX is connected to the SAS IOM, which in turn connects to one of the
SAS expanders in the storage sled. The expander in the storage sled is then
connected to all the physical drives in the storage sled, providing one path between
the H745P MX and a physical drive in the storage sled. With connections to both
SAS IOMs, the H745P MX therefore has two paths to each drive in a storage sled,
allowing for a multi-path configuration. If either a SAS IOM or a SAS expander in
the storage sled is removed, then a path to the physical drive is removed from the
SAS topology.
Enclosure-Assigned: All drives are shared among the compute sleds located in
assigned slots.
Enclosure-assigned maps the entire sled to a compute sled, however also enables
mapping of the MX5016s to more than one compute sled.
This module contains the Internal Dual SD Module (IDSDM) and vFlash SD
card that are combined into a single card module. There are two options
available:
– vFlash
– vFlash + IDSDM
– The supported capacity for IDSDM microSD cards are 16 GB, 32 GB, or
64 GB, while for vFlash the capacity is 16 GB only. The write-protect
switch is built onboard the IDSDM/vFlash module.
The IDSDM card provides the following major functions: Dual SD interface,
maintained in a mirrored configuration (primary and secondary SD)
– The IDSDM card does provide full RAID-1 functionality.
– Dual SD cards are not required; the module can operate with only one
card but will provide no redundancy.
– IDSDM enables support for SDXC (Secure Digital eXtended Capacity)
cards.
– USB interface to host system
Switches
Introduction
Each Fabric A and B allows for two IOMs (i.e., A1 and A2, or B1 and B2) for
redundancy. Each IOM connects to all 8 compute nodes through the orthogonal
connectors. The connection from the IOM to each compute node (actually the Mezz
card on the compute node) is a x8 bi-directional link. Depending on the interface,
all the lanes within the link may not be used, however the x8 provides a path for
future high-bandwidth fabric types.
The usage of mezzanine cards enables the end user flexible options from which to
choose for their I/O fabric, such as Ethernet or InfiniBand, for example.
Each Fabric A and B enables two IOMs (that is A1 and A2, or B1 and B2) for
redundancy. Each IOM connects to all 8 compute nodes through the orthogonal
connectors. The connection from the IOM to each compute node (actually the Mezz
card on the compute node) is a x8 bi-directional link.
Depending on the interface, all the lanes within the link may not be used, however
the x8 provides a path for future high-bandwidth fabric types.
Networking IOMs
1. The MX5108n is a basic but powerful Ethernet switch that is targeted for
customers with no more than two MX7000 chassis. This switch can drop to
lower speeds to connect to existing older equipment at a customer site. The
MX5108n requires OS10 and has eight server-facing 25 GbE internal ports.
2. The MX9116n Fabric Switching Engine (FSE) is a high bandwidth converged
solution for a large fabric. This model is also recommended for smaller rollouts
that require Fibre Channel capability. The MX9116n requires OS10 and has 16
server-facing 25 GbE internal ports.
3. The MX7116n Fabric Expander Module (FEM) scales fabric bandwidth across
multiple chassis. The MX7116n is an Ethernet repeater that sends traffic to the
FSE, where it is managed, and all switching happens. The combination of the
MX7116n FEM and the MX9116n Fabric Switching Engine (FSE) is what makes
scalable fabric possible.
A unique feature compared to the M1000e is that for two of the three networking
fabrics (Fabric A and B), there is no mid-plane. The vertically aligned compute
node mezzanine card connectors dock directly with horizontally aligned IOM
connectors through the assistance of mechanical guide pins. There is no mid-plane
or circuit board between these connectors. This means future networking
technology upgradability is extremely easy. There are no mid-plane bandwidth
limitations, because again there is no mid-plane.
If any settings enable compatibility, the Fabric is considered matched. User settings
may still cause incompatibility. The Fabric Consistency check is not designed to
detect/address these cases.
If an IOM supports a mix of fabrics, and two different Mezz cards (with different,
supported fabrics) are connected to the IOM - the Fabric Consistency will not flag a
mismatch.
Simulation Activity
Overview
MX7000 3D simulation
End of Module
Introduction
Introduction
MX Series Management
Important Points
1
Server Template
BIOS
RAID
NIC
iDRAC
2
Virtual I/O Pool
Virtual MAC Address Assign to one or more Server Configuration Profiles
WWN, WWPN servers or slots BIOS
iSCSI Name RAID
NIC
Virtual I/O identities
Virtual MAC Address
WWN, WWPN
3 iSCSI Name
Network
iDRAC
VLAN
Type (QoS)
Important Points
deployment process, the next available identity is fetched from the pool
for provisioning a server from the template description.
3. Network definition rounds out the PROFILE by adding the VLAN and QoS
settings. Historically these settings had to be configured in the switch as a
separate operation for every server and every switch in every chassis.
4. Slot for rip and replace.
iDRAC 9
Discussion
Discussion Notes:
iDRAC9 provides:
A powerful management processor that automates common management tasks
such as deployment, configuration, and, update, slashing server OPEX.
Innovative agent-free architecture ensures consistent management across any
environment—from containers to OSs to virtual machines.
HTML5-based GUI for easier, faster navigation without the need for plug-ins
Powerful scripting APIs to automate remote management including WS-MAN,
RACADM, and iDRAC REST API with Redfish 2016 additions.
QuickSync 2 for easier “at-the-box” deployment and trouble-shooting using your
Android or Apple mobile device
“Hi-Rez” monitoring reports detailed health status, connectivity topology, and
performance statistics for deeper insight into remote server issues.
Important Points
Regarding iDRAC9, there is no change from Rack Mount standalone 14G iDRAC.
For instance, the Dashboard Health Status is visibly different (vers. 3.21 and later).
Virtual Console is no longer available without the purchase of an Enterprise license
and iDRAC is now a separate component.
Previously, modular severs had the Virtual Console standard with the basic license,
but now, with these nodes, you must have the Enterprise DRAC license to acquire
Virtual Console.
To reiterate, you must purchase the Enterprise license to gain access to the Virtual
Console feature and while it is permitted to have a mix of Express and Enterprise
licenses in the MX7000 chassis, purchased licenses are applied per sled. Meaning,
the purchase of one (1) Enterprise license cannot be used for all sleds installed in
the chassis.
Lastly, the iDRAC on the NGM platform is now a separate card/piece and is no
longer integrated into the system board.
Important Points
Important Points
Dell performs over 8 weeks of System tests before a block release to ensure all the
components work together. The user can download as an ISO or .zip file and post
to a share where the management SW can access it, or with Internet access from
the chassis you can access the latest updates from dell.com.
Also, some updates including iDRAC can be completed immediately as it does not
impact the running applications on the server. While other updates that require a
restart can be downloaded and staged in the iDRAC to be applied on the next
scheduled server reboot, typically as part of an OS patch.
Important Points
Comprehensive RestAPI
Common code base and user experience with OME-Modular
Common code shared with Common code shared with OM-Enterprise includes:
OME-Modular includes: directory services, alert management, identity pools,
directory services, alert FW update, and RESTAPI
management, identity
pools, FW update, and
RESTAPI
Static and dynamic groups Simple, static roles - chassis admin, compute
with RBAC manager, fabric manager, storage manager, viewer
When displaying this slide, discuss at the box management options and Wireless
local access and new LCD panel.
Question
What are some of the options available for at the box management?
Answer
Question
Answer
Question
Answer
QuickAssist Technology
For more information about using iDRAC, see Dell's Integrated Dell Remote
Access Controller User's Guide at dell.com/idracmanuals.
QuickAssist Technology
Enhances security and compression performance in:
Management Module
Micro USB
Handle latch Handle
Important:
OME-M is completely software based. The lead Management Module
(MM) holds all configuration information on a database that is created
and replicated on the MMs. This configuration is done both for a
single chassis or all chassis in a group.
– The System ID button on the MX9002m rear panel allows for identifying the
MX7000 chassis in a rack. When pressed, the front panel of the chassis will
also blink in the identify pattern. This button also serves as the Health status
LED indicator, displaying blue for health or blinking amber if the MX9002m
has a fault. The rear-mounted ID button ONLY illuminates the
Chassis/Management Module ID button, it does not step through the other
devices including the left Control Panel ID Button.
The chassis supports an “At the Box” capability to use the USB serial port on the
rear of the active MX9002m to manage the MSM, an IOM, or iDRAC by using a
simple menu to select the device to manage.
1. Use a standard Android charging cable and plug it into the laptop.
2. Plug the other end of the cable into the back of the MM.
3. Configure PUTTY to connect over Serial.
4. Select the COM port.
5. Change the baud rate to 115200 and will automatically connect up (because
it's over USB it's emulated and will come up on your system).
6. Once all steps are completed, use Ctrl-A or Ctrl-X to get back to the main
Menu.
and remediation. These tasks are performed within a single console that unifies
management of tower, rack, and modular platforms. OpenManage Enterprise helps
in standardizing and supporting IT management policies and practices.
Overview
Creating Groups
Question
Once the group has been created, what would happen if the Lead chassis fails?
Answer
With this configuration you can survive the following failures and retain full
management network access to all nodes in the stack:
In this simulation, you walk through the steps to group a pair of chassis.
Once you have completed the Creating Groups Simulation Activity, inform the
instructor before moving on to the Creating Groups lab exercise.
Inform the instructor, once you have completed the MX5016s 3D Simulation
Activity.
Storage Fabric C
Within the MX7000 chassis, the internal connection of the storage subsystem is
referred to as “Fabric C.” This fabric is connected on the Main DB between sleds in
the front of the MX7000 chassis and I/O modules in the rear. The primary use
cases for this fabric will be for SAS or Fibre Channel storage connectivity, however
other technologies could use this interconnect in the future.
Enclosure-Assigned: All drives are shared among the compute sleds that are
located in assigned slots.
In this lab simulation, you walk through configuring Sled or Slot Profiles.
In this simulation, you walk through the steps to group a pair of chassis.
Once you have completed the Creating Groups Simulation Activity, inform the
instructor before moving on to the Storage Configuration lab exercise.
Extracting Logs
Important: When a log file is extracted, it is a .zip file. Once this file
has been extracted and saved, contact engineering (if needed) to
assist in deciphering these logs files.
Alert logs
Hardware logs
Host TSR/SupportAssist logs
While these logs may look new at first glance, they are similar to
those that have always been present in the iDRAC UI and legacy
CMC UI.
In this simulation, you walk through the steps to gather various logs.
Once you have completed the Combined Logs Simulation Activity, inform the
instructor before moving on to the Log Gathering lab exercise.
Firmware Update
Firmware
Question
What will happen from your viewpoint? Will we lose connection? If so, when?
Answer
Question
If you have two (2) MM modules should you update individually or together?]
Answer
Question
What effect will bundle updates have on users? In addition, should the bundle be
updated immediately, or, scheduled to run at a later time during a maintenance
period?
Answer
There are two ways of updating firmware. The first is to apply a single Dell
Update Package to a device or component. The other is to create a baseline
from a catalog to compare all the devices to the catalog and return a compliance
report to act on.
The following example of a Single device update includes the management module
including the OME-M.
After logging into the OME-Modular console as root user with password calvin,
there are five steps to a single device update:
1. Start by choosing the target device class, i.e. chassis, IOM, storage sled,
then compute
2. Select target(s) then click “Update Firmware”
3. Select individual package as the update source, browse, and select the
appropriate Dell Windows Update Package (DUP), follow via progress bar
4. Verify the FW version to be applied and select target(s) to be updated
5. Select Run now and then “Finish” for chassis, IOM, and storage sleds. For
servers, stage all but the last update as iDRAC will apply all staged updates
in one reboot
Dell EMC Engineering recommends applying your system updates in the following
sequence. While the file versions may change, the procedures outlined will
continue to stay true.
1. OME-Modular
2. MX5000s
3. MX5016s
4. MX5108n
5. Compute sleds
When adding a catalog one of the following three catalog source options can be
used:
Here we show an example of the Catalog Share Type as HTTPS. Other options
include:
NFS
CIFS
HTTP
HTTPS
Tip: Use the Add option on the Configuration > Firmware > Catalog
Management option to download the catalog from dell.com/support.
Refer to dell.com for the most current firmware update information.
Click the Catalog checkbox and click on View Report to reveal details.
1. Create Catalog
2. Create Baseline from Catalog
Firmware - Rollback
The Rollback Firmware window enables you to roll back a firmware update to the
previous version before the update. The rollback option is available if:
Backup Chassis
The Backup Chassis feature enables you to create a backup of the chassis
configuration, including those of its member Compute Sleds, and use it later. You
must have the Chassis Administrator privileges to back up the chassis.
Backups are saved to either an NFS or CIFS share. Select the location type where
you want the chassis configuration backup file.
To restore from Backup: OME-M Home > More Actions > Restore. The Restore
Chassis feature enables you to restore a chassis configuration, if the backup file is
created from the same chassis. The filename extension must be .bin.
Important:
You must have the Chassis Administrator privileges to restore the
chassis configuration. Also, the custom SSL certificate is not restored
after the chassis is restored.
After the simulation, complete the OME-M Navigation Discovery Hunt lab exercise.
In this simulation, you walk through the steps to update the firmware. (You will be
performing an actual MX7000 firmware update in a later lab exercise.)
Once you have completed the Firmware Update Simulation Activity, inform the
instructor before moving on to the OME-M Scavenger Hunt lab exercise.
14G PowerEdge servers are equipped with iDRAC9, unlike 13G PowerEdge
servers which used iDRAC8.
iDRAC Licensing
Accessing the iDRAC uses the same methods that have been used on previous
PowerEdge server, both modular and rack/tower systems.
Many new features are introduced with 14G hardware --- here is an overview of
the iDRAC9 changes over the iDRAC8.
iDRAC9 provides:
Industry-leading automation of server management via its rich APIs
A simpler customer support experience with less time required to resolve
issues.
Best-in-class GUI experience via its new HTML5-based console.
Simplified methods to deal with tedious, repetitive management tasks by
using simple easy-to-use tools.
The flexibility to support the diverse management requirements of
different markets and use cases.
Best in class at-the-box management using superior support of popular
mobile devices.
iDRAC9 Password
Factory Default:
Secure Default Password (14G Default)
Legacy Default Password (Pre-14G Default)
For the 14G PowerEdge servers and iDRAC9, a new option for the factory default
password for iDRAC is offered. This option is referred to as the secure default
password for iDRAC.
Dashboard
The System tab provides system information and iDRAC details and at a
glance status of the systems. You can access more details about the system
through this tab.
The Storage tab provides details on the storage components. Summary
information and information about controllers, physical disks, virtual disks,
and enclosures are accessed from here.
The Configuration tab is where settings for items such as power
management, virtual console, licenses, systems, storage configuration,
BIOS, and server configuration profile may be configured.
The Maintenance tab includes the lifecycle log, job queue, system update,
system event log, troubleshooting, diagnostics, and SupportAssist.
The iDRAC tab displays the details of the iDRAC settings. It includes
configuration of the Network Settings, IPv4 settings, and the iDRAC Service
Module. Options for connectivity, services, and users are also available.
Unpinned Menu
The unpinned menu shows the different subheadings under each major menu item.
Dashboard Actions
Use the Dashboard to Power On System or Shutdown the system, Identify System,
and perform More Actions (as in Turns on the System Lockdown Mode) using
iDRAC.
Power On System
– Power On System allows you to turn on the system. Success message is
displayed after the system is on. You can further shut down the system
using the following options:
Graceful Shutdown
– Shuts down the operating system and powers off the system. Graceful
shutdown requires an Advanced Configuration and Power Interface (ACPI)
aware operating system, which allows system directed power management.
– NOTE: Configure shutdown option on the operating system before you
perform a graceful shutdown using this option. If you use this option without
configuring it on the operating system, it reboots the managed system
instead of performing a shutdown operation.
Power Off System
– Powers off the server
NMI (Non-Masking Interrupt)
Lockdown mode is a licensed feature, and it is only available for iDRAC enterprise
customers only.
After the System Lockdown mode is enabled, users cannot change any
configuration settings. System Settings fields remain disabled with some
exceptions.
After the System Lockdown Mode is enabled, the message in the yellow box is
displayed in the iDRAC UI. Also notice that with the System Lockdown Mode
enabled, access to the Lifecycle Controller is disabled.
Storage
Controller functions may be configured and managed without accessing the BIOS.
Which includes, configuring virtual disks, applying RAID levels and hot spares for
data protection. Other controller functions such as rebuilds, and troubleshooting
can be initiated from the iDRAC. Data may be protected by configuring data-
redundancy or assigning hot spares.
Controllers:
– Most operating systems do not read and write data directly from the
disks, but instead send read and write instructions to a controller. The
controller is the hardware in your system that interacts directly with the
disks to write and retrieve data.
– A controller has connectors (channels or ports) which are attached to
one or more physical disks or an enclosure containing physical disks.
The physical devices connected to the controller must have the latest firmware.
Power Control:
Use this section to view the power status of the server and perform remote power-
control operations.
Use this page to view or configure the Virtual Console settings. Using the Virtual
Console, the remote system using the keyboard, video, and mouse on the local
management station may be managed to control the corresponding devices on a
remote managed system. Up to a maximum of six simultaneous Virtual Console
sessions may be run. The Virtual Console may be used with the Virtual Media to
perform remote software installations.
ActiveX: Starts the ActiveX viewer on a system running Windows and a Java plug-
in on a system running Linux. ActiveX viewer works only on Internet Explorer.
Java: Starts a Java viewer.
HTML5: Starts an HTML5 virtual console if the virtual console feature is enabled
and the Virtual Console Type is set as HTML5.
NOTE: While using HTML5 to access vConsole, the language must be consistent
across client and target keyboard layouts, client OS, client browser, and the target
OS. For example, all must be in English (US), or any of the supported OS
languages.
ActiveX:
– Starts the ActiveX viewer on a system running Windows and a Java plug-in
on a system running Linux. ActiveX viewer works only on Internet Explorer.
Java:
– Starts a Java viewer.
HTML5:
– Starts an HTML5 virtual console if the virtual console feature is enabled and
the Virtual Console Type is set as HTML5.
A view of the window that opens when HMTL5 is selected as the viewer type.
Virtual Media,
– Accessed through the Virtual Console viewer, provides the managed system
access to media connected to a Remote system on the network.
Using Remote File Share
– Specify an image file on a network share and make it available to the
managed system's operating system as a Virtual drive by mounting as a CD,
DVD, or Floppy using Network File System (NFS) or Common Internet File
System (CIFS).
Virtual Media and RFS is mutually exclusive.
Use this page to view and filter the events, generate alerts, or specify the actions
when an event occurs. An event occurs when the status of a system component is
outside of predefined thresholds.
If an event matches an event filter and this filter is configured to generate an alert
(Email, SNMP trap, IPMI alert, remote system log, Redfish event, or WS event),
then an alert is sent to one or more configured destinations. If the same event filter
is configured to do a task (such as reboot, power cycle, or power off system), the
action is performed. Only one action may be set for each event.
Storage settings are based on one single controller (at a time), and only one job
per controller can be scheduled at a time
Batch changes may be made into one job by adding them to the Pending
Operations.
Click Apply when ready to start or schedule the job. Pending operations persist
until the job is created or they are discarded.
Use this page to configure the BIOS settings. The BIOS fields are grouped under
menus.
Maintenance
iDRAC Settings
Overview
– Use this section to view the iDRAC details, network settings, and IP settings.
Connectivity
– Use this page to configure the network properties, common iDRAC settings,
IPv4 and IPv6 properties, IPMI settings, VLAN settings, and manage SSL
certificates.
Services
– Services - Use this page to view or change the interface settings for local
configuration, Web server, SSH, Telnet, remote RACADM, SNMP agent,
and automated system recovery agent.
Users
– Use this page to add or edit users, configure directory services and log in
methods, view active sessions, and so on.
Settings
– Use this page to create manual backups of server settings, create
automated backup tasks, or to import server profiles
Management Module
– This page shows the Health Status, the firmware version and the IP Address
of the Management Module.
Simplicity and ease of use are key tenets of Group Manager, which provides the
ability to see a list of servers and quickly identify a server with an error and access
that server with a single click.
Group Manager has three primary views: Summary, Discovered Servers, and Jobs.
The Summary view provides a searchable dashboard for the health monitoring and
inventory of all systems in the local group.
The Discovered Servers view shows any iDRACs on the link local network that
have been auto-discovered and have not been on-boarded to a group.
The Jobs view shows the progress and history of group configuration actions.
Group configuration actions include: setting up email alerting, adding a local iDRAC
user, removing a local iDRAC user, changing a local iDRAC user password,
changing the group passcode, changing the group name, and on-boarding newly
discovered iDRACs.
Display the Group Manager key points slide. Review the key points of Group
Manager as outlined on the slide. Point out the following information and ask the
following question.
Important information:
Group Manager is built into the iDRAC. No software is necessary to install for the
feature to work. It runs completely on the iDRAC.
Question
What do you think up to 100 iDRACs means on the second bullet of the slide?
Answer
Question
Answer
The iDRAC Service Module feature is available in the iDRAC Express and
Enterprise licenses.
If you access the iDRAC iSM setup section before the iSM is installed, you will see
a screen similar to this slide screenshot.
The iSM can either be installed using the iDRAC option as seen in this screenshot
or you can download the iSM file from the support site and install it from within the
server operating system
Note: Only use an iSM file that is supported with 14G servers.
Lab Activities
In this lab simulation, you will complete lab activities covering Navigating the
iDRAC9 interface and the iDRAC Virtual Console.
You will now complete the Navigating the iDRAC9 and the iDRAC Virtual Console
lab exercises.
Lifecycle Controller
The Lifecycle Controller is the engine that provides advanced system management
functions to perform tasks such as deploy, configure, update, maintain, and
diagnose through a graphical user interface. Like the iDRAC, the Lifecycle
Controller is embedded into the server’s system board and is available outside of
the OS, as it can only be accessed during the power-up sequence.
The Lifecycle Controller and the iDRAC are linked closely together and work side
by side. Some of the Lifecycle Controller functions are linked to the iDRAC license,
so the number of Lifecycle Controller functions are dependent on the iDRAC
license type.
Accessing the Lifecycle Controller is done by pressing <F10> when the prompt
appears during the power-up sequence. The system will then go through the rest of
the power-up sequence before booting to the Lifecycle Controller.
Firmware Update
While you can update the firmware through the iDRAC interface, you can also
update it using the Lifecycle Controller. Just like the iDRAC, you can roll back to a
previous version. The list of available updates is shown here, but keep in mind that
with the IDRAC, you can access it even while the operating system is running;
whereas with the Lifecycle Controller you have to power-cycle the server. Even
when you update using the iDRAC, it is actually the Lifecycle Controller that
performs the update, as the Lifecycle Controller is running the whole time the
iDRAC is operational. You can obtain the updates from a number of different
sources, including the Dell FTP server, local drives such as USB or DVD, or even a
network share.
Hardware Configuration
The Operating System Deployment option is a key feature that allows users to
choose from a range of operating systems to install on the managed server. This
includes the ability to configure RAID during the OS installation. The interface is
similar to the System Build and Update Utility DVD that some Dell customers may
be familiar with.
To install an OS, first select the desired one. The Lifecycle Controller will then
check its driver base from its own updatable OS driver pack before instructing you
to insert the OS installation DVD. It will then install the OS and insert the drivers at
the correct time. This results in a fully functional operating system without concern
about which drivers are required or how to get them.
Platform Restore
The Platform Restore option is used to push the configuration back to individual
hardware parts or the entire system board. This is a great option for ensuring that if
things go wrong, you can get your server up and running again as quickly as
possible.
The configuration of every hardware part installed in your server is stored within the
Lifecycle Controller, including the firmware and BIOS code. If something goes
wrong—like if a network card fails and the technician arrives on-site and replaces
the card—then what happens? The card would have to be reconfigured manually—
that is, if the person who knows how it should be configured is around.
That is where the Parts Replacement option comes in. It allows you to decide what
happens if a hardware part fails and needs to be replaced. Instead of manually
reconfiguring the new part, you can set the Lifecycle Controller to automatically
push the configuration back out to the new part. This includes re-flashing the
firmware with the code that was on the old part.
This is a great tool but let us take it one step further. What if the system board
fails? Since the Lifecycle Controller is located on the system board and holds all
the configuration data and firmware versions, you would have just lost all of that
information. But have you?
This is where the Server Profile comes in. When you create a Server Profile, it
backs up all the Lifecycle Controller configuration information, the BIOS, and all the
firmware code onto a USB stick, hard drive, or network share. When the new
system board is installed, you simply access the new Lifecycle Controller and
import the Server Profile. All the backed-up configuration information is now
restored to the new Lifecycle Controller, which in turn pushes all of that information
out to the hardware parts. That saves you time and effort from manually
reconfiguring devices like the iDRAC and the BIOS.
The Lifecycle Controller also pushes out the firmware and BIOS code if required.
However, you must back up the Server Profile before the system board fails. The
Platform Restore function is also dependent on having an iDRAC Enterprise
license.
Lab Activity
You will now complete lab exercise on Navigating the Lifecycle Controller.
You will now complete the Navigating the Lifecycle Controller lab exercise.
Redfish
Introduction
Redfish
What is Redfish
Depending upon the scale of an IT installation, using the Graphical User Interface
(GUI) of OpenManage Enterprise or a single iDRAC may be desirable. However,
when deploying thousands of servers around the world, this mode of operation may
call for the writing of programs/scripts, which can operate on 1000s of servers at a
time. This type of large scale environment would depend on APIs like Redfish.
What is Redfish?
– It is the first open effort toward building a single API for all hardware. It
replaced IPMI-over-LAN and is a modern, scalable, secure management
API.
– It enables seamless, programmable insight, control, and ultimately
automation of hyperscale architectures.
– A foundational component of Intel Rack Scale Architecture
Basically, like the CLI, IPMI, and SMNP, Redfish is just another way to
interface with various platforms.
Key benefits include:
Increased simplicity and usability
Encrypted connections and heightened security
A programmatic interface that can easily be controlled through scripts
Redfish uses JavaScript Object Notation (JSON) and the Open Data Protocol
(Odata). JSON is human-readable and simpler than XML.
The URI is the primary unique identifier of resources. Redfish URIs consist of three
parts as described in RFC3986: Part one defines the scheme and authority of the
URI, part two specifies the root service and version and part three defines a unique
resource identifier.
The URI is the primary unique identifier of resources. Redfish URIs consist of three
parts as described in RFC3986: Part one defines the scheme and authority of the
URI, part two specifies the root service and version and part three defines a unique
resource identifier.
Redfish Operations
Auto Config, RACADM, WS-Man and Redfish SCP operations via HTTP/HTTPS
in addition to CIFS and NFS.
SCP operations via local file streaming.
SCP JSON format for export/import in addition to .XML format.
End of Module
Initial Setup
Introduction
Initial Setup
ESD
Due to the increased density of semiconductors used in recent Dell products, the
sensitivity to static damage is now higher than in earlier Dell products. For this
reason, some previously approved methods of handling parts are no longer
applicable.
Rack Installation
Weight of Chassis:
PowerEdge MX7000 180 lbs. (82 kg) 400 lbs. (182 kg)
Rack installation:
1. Once the Static Rails are installed, with two people, pull the four enclosure
handles on the MX7000 chassis upward, and lift the enclosure.
2. Align the rear of the enclosure with the rails and slide the enclosure into the
rack. When the rear enclosure handles are near the rack flange, press the blue
release buttons on the sides of each enclosure handle, and push the handles
down until it locks into place.
3. Slide the enclosure into the rack until it is firmly seated.
4. Tighten the captive screws on the front panel of the enclosure.
5. Install the sleds, rear modules, power supplies, and fans.
Question
Answer
Question
Would you connect the Management Module (MM) to network cables before you
configure the LCD?
Answer
Important Points: Static rails do not support the ability to service the
enclosure while in the rack and are not compatible with a Cable
Management Arm (CMA). A strain relief bar (SRB) is provided with
each rail kit to organize and secure the power cables exiting the back
of the server.
For the MX7000 chassis front KVM, MM uses a standards-based KVM client to
attach to the iDRAC KVM Server of a requested sled, and then renders the
received video frames within the MMs G200 graphics controller. The MMs USB
host controller receives keyboard/mouse events and redirects them as a virtual
USB keyboard/mouse to the iDRAC of the compute sled.
There is not a KVM selector button (as seen, for example, in the FX2 chassis),
instead it uses an On-Screen Display (OSD) overlay to display a list of the installed
sleds or Chassis Consoles for the user to select from. The MM supports redirecting
the chassis console to the front panel DisplayPort video similar to the M1000e
blade KVM solution. This chassis console is listed as the last node in the OSD for a
user to select it when needed. After the chassis console is selected, the user can
log in to it and execute supported chassis management commands.
KVM Disable
– KVM is enabled by default, however it can be disabled. If disabled, no KVM
functionality is enabled.
Chassis Console Disable
The images shown in this slide depict examples of what you should observe when
connecting to the KVM.
Steps:
1. Select the OME-Modular port from the On-Screen Display by clicking with the
mouse or using the arrow keys plus enter.
2. When prompted for credentials, enter the password on the chassis luggage tag
on the front right.
The LED-Only control panel is composed of five Status LEDs, an LED Status Bar,
and a system ID button. The LED behaviors of the five status LEDs are called out
in LED-Only Control Panel slide.
If the customer has the installed LED-Only Control Panel, here are the options to
configure the MX platform:
Using the Serial Port/USB connector on the controller
Using the KVM and selecting from the menu in the MM.
– Enter credentials and use the following racadm command to configure:
racadm setniccfg -s <IPv4Address> <netmask> <IPv4 gateway>
Through QuickSync when available
The LED Status Bar is illuminated in Blue/Amber based on the chassis state and
behaviors are called out in the following table:
When the System ID button is pressed, the identify state alternates between blink
active and inactive. When blink is active, the blue ID LEDs on one of the following
components will blink:
The chassis LED Status Bar (or circular ID button/indicator on LCD option)
A sled and associated devices of the same sled
The MM firmware rotates through the installed sleds, moving to the next compute
sled when the button is pressed.
The optional LCD Panel is a touch screen display (supporting button press, scroll,
and flick gestures), which provides system information, QuickSync 2 buttons,
status, and error messages to indicate if the system is functioning correctly or if the
system needs attention. The LCD is vertically oriented. It supports touch and has
no physical buttons (all user interaction is touch based).
Below the LCD are two status LEDs that light up when the user is not interacting
with the LCD. To extend the life of the LCD backlight, the LCD goes black when not
in use and one of the two LEDs are then turned on. The LCD is not hot-pluggable
and requires the chassis to be fully powered off for servicing.
There are two LED illuminated ICONs located below the LCD (but still part of the
touch panel) which are:
Activate LCD/Identify/Alert ICON – Circle icon illuminates blue or amber as
follows:
– Solid blue when an alert is not present
– Blinks blue when a chassis Identify is in progress.
– Blinks amber when an alert(s) is/are present.
– When the LCD display area is off, pressing this circled area turns on the
LCD.
QuickSync 2 Status - White icon is evident when the Bluetooth/WiFi is active. It
is solid when the transceiver is turned on and blinking when there is traffic.
The Main menu provides access to the LCD functionality. By default, the main
menu is the default home screen, but users can choose an alternative view from
the Welcome wizard or from the Settings page. When users are not interacting with
the LCD, it is turned off. When a user touches the LCD screen while it is off (black),
the LCD turns on.
On initial power up an animation indicating that the LCD is starting is shown until
the chassis manager has initialized and can then respond to user touches. If the
LCD loses communication with the MM, it displays a message on the LCD
indicating it is unable to communicate with the Management Module.
2. Select Network Settings. A list is displayed showing the IPv4 address, IPv6
address, DNS name, MAC Address, and VLAN ID of the chassis controller
(MSM)
3. Click the IPv4 address to show the current values for the extended IPv4
network settings. The IPv6, VLAN ID, and DNS name are not selectable.
a. IPv4 address
b. Subnet mask
c. Gateway
Optionally, if a static IP address is desired:
Select Settings from the main menu
Select Network Settings
Next, select Edit (Note: Edit is at the bottom of the network settings list if the
LCD is in View and Modify mode)
IPv4
Select Static IP
While changing from a DHCP address, confirm the change. The IP address
configuration area is displayed
– Select the first through fourth octet one at a time and configure the value
Select Next
The Subnet Mask configuration area is shown next
– Select the first through fourth octet one at a time and configure the values
Select Next
The gateway address configuration area is shown
– Select the first through fourth octet one at a time and configure the value
Select Save.
The IP, subnet mask, and gateway address values are validated.
a. If invalid, an error message is shown to the user
b. If closed, you are returned to the network Settings page
c. Navigate back to correct the error as needed
If all values are valid, a wait animation is shown while the network is being
updated
You are returned to the network Settings screen showing the updated current
values
Simulation Activity
In this simulation, you walk through the steps to use the Left Rack Ear LCD Panel.
Inform the instructor, once you have completed the Using the Left Rack Ear LCD
Panel Simulation Activity.
The subset of RACADM commands available can be displayed via racadm help.
To set a static IPV4 address the command: racadm setniccfg would be used.
Alternatively, a serial interface can be used by connecting the micro USB port on
the rear of the management module to a laptop and opening a terminal emulator
with settings of 115200 baud, 8 data bits, and 1 stop bit.
The initial screen would then enable selection of a login to the management
module, networking I/O modules, or servers (iDRAC). Selecting the management
module would prompt for the credentials on the luggage tag and provide the same
RACADM command options as via the KVM.
When first logging into the web UI, a getting started wizard is displayed for initial
chassis configuration. It is always possible to restart this wizard by selecting Initial
Configuration in the actions bar.
7. Group Definition
If multiple chassis are wired together before running the Getting Started
Wizard (GSW), a list of available chassis to add to the group is displayed.
After one or more chassis are wired together, you can log in to the chassis
that you want to promote to lead.
The wizard then prompts to name the group plus how and which settings to
replicate from the lead to the members. The recommendation is to
automatically propagate all settings.
Simulation
In this simulation, you walk through the steps to deploy the chassis using the
Chassis Deployment Wizard.
Inform the instructor, once you have completed the Chassis Deployment Wizard
Simulation Activity.
Power on Sequence
The MM communicates with IOMs, compute, and storage sleds via the
Fabric Resource Director (FReD).
FReD is a microcontroller-based reference circuit that provides a consistent
management interface between the MM and the other hardware.
FReD communicates with MM through an I2C connection.
FReD provides the following functions:
DHCP:
DHCP is the default IP address mode. When using DHCP, the switch obtains its IP
address, subnet mask, and default gateway from the DHCP server. The DHCP
client can only connect to a DHCP server that is on the same subnet as the switch.
If your DHCP server is not on the same subnet as the switch, use a static IP
address.
Static:
Dell EMC recommends configuring the switch module through the Compute Sled
chassis management GUI, namely, Dell EMC OpenManage Enterprise - Modular.
Use this management GUI to configure the switch IP address and perform
advanced configuration tasks. By default, the IP address mode is set to DHCP. Use
this GUI to assign a static IP address. To configure the FC switch module, follow
these steps:
1. Log in to OME-M
2. Select Devices > I/O modules
– The I/O modules display.
3. Select the I/O module name, IP address, or View Details on the right side of
the display.
Tip: The Fibre Channel Switch Module was used as an example. For
additional information on how to configure a specific switch, and/or
how to configure via Serial connection, refer to the switch model
Installation guide for further instruction.
Simulation Activity
In this simulation, you walk through the steps to configure the I/O Module in OME-
M.
Inform the instructor, once you have completed the Configuring I/O Modules in
OME-M Simulation Activity.
Simulation Activity
Inform the instructor, once you have completed the Setting Up iDRAC Simulation
Activity.
Identity Pools
Identity Pools
Identity pools are used for template-based deployment on servers to virtualize the
network identity for: Ethernet, iSCSI, FCoE, and FC.
You can create a maximum of 5000 Identity Pools in each of these categories.
The server deployment process fetches the next available identity from the pool
and uses while providing a server from the template description. You can then
migrate the profile from one server to another without losing access to the network
or storage resources in your environment.
You can edit the number of entries in the pool. However, you cannot reduce the
number of entries less than those assigned or reserved. You can also delete the
entries that are not assigned or reserved.
You can perform the following tasks on the Identity Pools page:
Server Templates are created and deployed after fabrics are created and deployed
to the fabric IOMs to provide connectivity between servers and fabrics.
To create a template, from OME Modular click Home > Configuration > Deploy >
Create Template.
Slot configuration
Setup configuration - The setup configuration covers Quick Deploy, Security,
Network Services. Customer Network Access, and Multi-Chassis Management.
You can also perform the following tasks on the Deploy page:
– Create templates
– Edit templates
– Clone templates
– Export templates
– Delete templates
– Edit network
– Deploy templates
– View template details - Select the check box corresponding to a template
from the list to see View Details on the right side. Click View Details to view
the Template Details page.
Lab Activity
You will now complete the PowerEdge Server Templates lab exercise.
Compute Sleds
1 2 3 4 5
11 10 9 8 7 6
MX740c:
1. Backplane
2. Backplane cable
3. Processor 1 (heatsink)
4. Processor 2 (heatsink)
5. Mezzanine card A1
6. Power connector
7. Mezzanine card B1
8. Mini Mezzanine connector
9. iDRAC card
10. BOSS connector
11. PERC connector
6 7 8 9 10 11 12 13
14
4 15
3
16
2
17
1
18
19
27 26 25 24 23 22 21 20
MX840c:
1. Memory module sockets
2. TPM module connector
3. SATA connector
Simulation Activity
Inform the instructor, once you have completed the MX840c 3D Simulation
Activity.
Discuss OS Installation. Be aware that there are no differences when installing the
OS on the MX7000 platform, as compared to previous 14G modular platforms.
Poll the class to ask how many of them have installed an OS on a 14G modular
platform. (Take note of those who have minimal to no experience with hands on OS
installation knowledge.) Take this opportunity to share your OS installation
experience and ask participants to share any good to know, or gotchas that they've
experienced when installing the OS and how they found a work around to their
potential issue (if applicable).
Discussion points: (Guide the participant, but do not provide answers. Also,
participant should be aware of when to use the following, that is, remote or local
install).
Question
When reloading the OS what methods can be used? Ask participants to provide the
answer.
Answer
Question
What iDRAC9 licensing changes have been made that affect how the OS is
installed?
Answer
End of Module
MX Networking Hardware
MX networking overview:
MX - Modular Series:
9 = Speed/Type.
1 = Generation.
16 = Number of internal ports.
n = Device Type.
PowerEdge naming convention:
8 can be 2 or 4 CPUs.
9 is 4 CPUs.
2. The second number indicates the generation:
0 for 10th generation.
1 for 11th generation and so on.
3. The third number indicates the make of the CPU:
0 for Intel.
5 for AMD.
The server sleds are installed vertically on the front of the chassis and I/O
modules (IOMs) are installed horizontally on the back of the chassis.
The connector on the mezzanine card attaches directly to the connector on the
IOM installed at Fabric A and B. No midplane is required.
The server sleds have two mezzanine cards: One for the A fabric and one for
the B fabric.
Each mezzanine card has two connectors: The A mezzanine card has
connections to the A1 and A2 fabrics and the B mezzanine card has
connections to the B1 and B2 fabrics.
What are the supported I/O Modules for the following Fabrics?
Fabric A and B
1. Ethernet Switch:
Fabric C
1. 12 Gb SAS:
2. Fibre Channel:
8 x 25 GbE server-facing ports, 2 x 100 GbE uplink ports, 1 x 40 GbE port and 4
x 10GBase-T ports.
Supported in Fabrics A & B.
Supports all PowerEdge MX Ethernet Cards.
Supports Open Networking, Full Switch, and SmartFabric operating modes.
< 800 ns latency, 960 Gbps switching fabric.
NVMe over Fabric Ready.
Pass-Through Modules
Key Capabilities:
Cable types
Key features:
Direct Attach Cable (DAC) with optical connectors on the end—also called
Twinax. FEC Error Correction adds latency, so these cables are supported at
short lengths of two meters and under.
Active Optical Cable (AOC) is a fiber optic cable that is optimized for the length
of the cable. AOC is more expensive than DAC but less expensive than
standard transceivers and passive fiber.
Multi-Mode Fiber (MMF) 50-um Large core fiber 100 m (300 m) reach. Easy to
attach components. Transceivers are low cost. Fiber 3x the cost of SMF.
Single-Mode Fiber (SMF) 9-um Tiny core fiber. 2/10Km reach. Hard to attach
components. Transceivers are expensive. SMF costs less than dental floss.
Optic Connectors
Optic connectors:
Introduction to QSFP28-DD
Key features:
QSFP28-DD ports are backwards compatible and support QSFP28 and QSFP+
Optics and cables.
QSFP28 ports are backwards compatible and support QSFP+ Optics and
cables.
QSFP28-DD ports are slightly deeper than QSFP28 with a second row of
contacts.
Full-Switch Mode
Overview
Scalable Fabric Architecture makes the MX9116n Fabric Switching Engine (FSE)
available to all chassis in the system by linking it to MX7116n Fabric Expander
Modules (FEM).
MX7116n Fabric Expander Modules key features:
Topology Architecture
What is the recommended FSE placement (split across two chassis for
redundancy)?
Choose one configuration and tick the best answer.
Configuration 1: __________
Configuration 2: __________
Minimum Requirements
Deploy a SmartFabric
Define VLANs:
To define VLANs using the OME-M console, perform the following steps:
1. Open the OME-M console.
2. From the navigation menu, click Configuration > Networks.
3. In the Network pane, click Define.
4. In the Define Network window, complete the following:
To define VLANs using the OME-M console, perform the following steps:
1. Open the OME-M console.
2. From the navigation menu, click Devices > Fabric.
a.
Enter SmartFabric in the Name box.
b. Optionally, enter a description in the Description box.
c. From the Design Type list, select 2x MX9116n Fabric Switching Engine in
different chassis.
d. From the Chassis-X list, select the first MX7000 chassis.
e. From the Switch-A list, select Slot-IOM-A1.
f. From the Chassis-Y list, select the second MX7000 chassis to join the
fabric.
g. From the Switch-B list, select Slot-IOM-A2.
h. On the Summary page, verify the proposed configuration and click Finish.
Define uplinks:
After initial deployment, the new fabric shows Uplink Count as ‘zero’ and shows a
warning sign. The lack of a fabric uplink results in a failed health check. To create
uplinks, follow these steps:
1. Open the OME-M console.
2. From the navigation menu, click Devices > Fabric.
3. Click on the fabric name, SmartFabric.
4. In the Fabric Details pane, click Uplinks.
5. Click on the Add Uplinks button.
6. In the Add Uplink window complete the following:
a. Enter Uplink01 in the Name box.
b. Optionally, enter a description in the Description box.
c. From the Uplink Type list, select Ethernet.
d. Click Next.
e. From the Switch Ports list, select Ethernet 1/1/41 and Ethernet 1/1/42 for
both MX9116n FSEs.
f. From the Tagged Networks list, select VLAN0010.
Switch Firmware
MX SmartFabric Deployment
Overview
The video demonstrates, in the correct order, the steps to deploy SmartFabric
Services:
1. MX7000 chassis access configuration.
2. OME-M Chassis Deployment Wizard
3. I/O Module settings
4. Creating a Multi-Chassis Management group
5. Z9100-ON Leaf Switch configuration
6. Defining VLANs in OME-M
7. Creating a SmartFabric
Only the specific Compute Sleds, I/O modules, and switches used in the demo are
shown and described in the video. The equipment used includes:
Dell EMC PowerEdge MX7000 Modular Chassis
Dell EMC PowerEdge MX740c compute sled
Dell EMC MX9116n Fabric Switching Engine (FSE)
Dell EMC MX7116 Fabric Expander Module (FEM)
Dell EMC MX9002m Management Module
Two different Leaf Switch examples:
A Scalable Fabric spans multiple chassis and allows them to behave like a
single chassis from a networking perspective.
A Multi-Chassis Management group allows multiple chassis to be managed like
a single chassis.
Click to play video: Compute and I/O Modules Used In Deployment Example
Inter-Chassis Cabling
The inter-chassis cabling used in this deployment example is shown in this figure.
Additional inter-chassis cabling provides redundancy to all available MMs in both of
the chassis.
Using the S3048-ON switch, Port GB1 of C1MM1 and C2MM2, connect to the
management network.
The initial out of box configuration uses the LCD screen to assist in setting up the
chassis for access.
After selecting the language, tap Main Menu, and then Save.
Use the password printed on the Express Service Tag, which is located on the front
right side of the chassis.
The Chassis Deployment Wizard can be used to perform initial configurations and
set up.
Various configurations such as setting the time, importing a profile, network IOM
access settings, and so on can be completed using this wizard.
From the Settings screen, choose Network, then click to remove the check from the
Enable DHCP option.
In the required fields, enter the IP address, Subnet mask, and Gateway, then click
Apply.
Deployment Overview
SmartFabric Deployment
SmartFabric Description.
List the SmartFabric deployment steps.
Deployment Steps
This video shows the steps to create a Mult-Chassis Management (MCM) group
using OpenManage Enterprise Modular (OME-M).
This video shows all the required networking connections to create a SmartFabric:
Compute sleds to MX9116n
VLTi connection between Chassis 1 MX9116n and Chassis 2 MX9116n
Connection between MX7116n Fabric Expansion Modules and MX9116n Fabric
Switching Engines
Connections between both MX9116n Fabric Switching Engines and Z9100-ON
Leaf Switches
VLT connection Between Leaf Switch 1 and Leaf Switch 2
This video reviews the configuration of the Z9100-ON Leaf Switches using the
Command Line Interface.
Leaf Switch configuration is not part of the MX7000 ICM class. Several switch
models are supported. This information is provided to show one example of
Leaf Switch configuration.
Leaf Switch configuration is necessary to support the SmartFabric Uplinks.
To support the uplinks, VLT connections between the Leaf Switches and the
MX9116n FES in each chassis must be prepared.
VLANs on the Leaf Switches must match the VLANs created on the
SmartFabric.
For detailed steps to configure Leaf Switches, refer to the Dell EMC PowerEdge
MX SmartFabric Deployment Guide.
Define Uplinks
Define Uplinks
This video shows the commands used to validate the Z9100-ON Leaf Switch
configuration.
This video shows the commands used to validate MX9116n switch operating in
SmartFabric mode.
This section discusses how to deploy SmartFabric using a different Leaf Switch
option.
In this example, two Cisco Nexus 3232C switches are used as Leaf Switches.
For additional information, see the Dell EMC PowerEdge MX SmartFabric
Deployment Guide.
Click to play video: Topology: Cisco 100 GbE Leaf Switches with MX7000 Chassis
This section reviews the configurations on the Cisco 100 GbE leaf switch.
Click to play video: Cisco 100 GbE Leaf-1 Command Line Interface
In this simulation, practice using OME-M to locate switch and port information.
In this simulation, practice using OMEM to configure breakout, VLANs, and Identity
Pools.
In this simulation, practice using OME-M to create a chassis group and review
chassis group information.
In this simulation, practice using OME-M to create a fabric and add an uplink to the
fabric.
End of Module
MX Troubleshooting
Blanks All Empty slots in the MX7000 chassis must be populated with
blanks (Sled, IOM, EC, and PSU). Required for proper cooling of
the chassis and components.
Power The number of power supplies that are required is dependent upon
Supply system configuration and wanted redundancy mode. The minimum
quantity must be 2. For grid redundancy, power supplies must be
populated in the following order: 1+1 (1 and 4), 2+2 (1, 2 and 4, 5)
3+3 (all slots)
Power Cords One C21/C20 power cable must be connected to the C22 plug
corresponding to each populated PSU.
Control Right Control Panel is always required, and at least one left control
Panel panel option (LCD or LED) must be present on the MX7000 chassis.
Features and functionality at both the host and fabric levels to greatly raise
levels of scaling.
CPU and fabric integration necessary for the increased computing density,
improved reliability, reduced power, and lower costs required by significantly
larger HPC deployments.
Fabric tools to readily install, verify, and manage fabrics at this level of
complexity.
Storage Storage sled can be populated in any sled slot. Requires at least one
Sled compute node in the chassis to map storage to. Also requires at least
1 Fabric-C SAS IOM to be present and powered ON.
LCD Alerts
Located below the LCD are two LEDs that light up when the user is not interacting
with the LCD. To extend the life of the LCD backlight, the LCD goes black when not
in use and one of the two LEDs are then turned on.
The LCD is not hot-pluggable and requires the chassis to be powered off for
servicing.
The two LEDs illuminated ICONs located below the LCD are Activate
LCD/Identify/Alert ICON and Circle ICON, which are still part of the touch panel.
Circle icon illuminates blue or amber as follows:
Displayed Errors
The Hardware Logs tab displays the hardware log of the chassis and the
components in the chassis. You can sort the logs that are based on the
following parameters:
Severity
Message ID
Start date
End Date
Category
You can export all the logs or selected logs as an excel sheet or file containing
comma-separated values (.csv) to your local system or network drive.
Severity
Displays the severity of the hardware log. The possible options are:
All
Warning
Unknown
Critical
Fatal
Info
Message ID
Displays the identification number of the message.
Timestamp
Displays the date and time of generation of the hardware log.
Category
Displays the severity of the hardware log. The possible options are:
All
Audit
Configuration
Storage
System Health
Updates
Work Notes
Description
Displays the description of the hardware log.
SupportAssist Collection
Also, SupportAssist Enterprise can monitor hardware issues that may occur on
devices that you are managing by using OpenManage Essentials, Microsoft
System Center Operations Manager (SCOM), or OpenManage Enterprise - Tech
Release.
Hardware Troubleshooting
Hardware Troubleshooting
Troubleshoot
Commands Function
Rear LEDs
There are two LEDs on each MX9002 at the rear of the chassis: 1) MM Power LED
(green only), 2) MM Status Button/LED (bicolor: Blue and Amber), also serves as
an ID button.
The Power and Status LED (which color depends on the status) states are as
follows:
Failed MM (hardware failure prevents MM Power LED OFF, Status LED OFF.
from uboot on or not able to start).
Failed MM - MMC powers on, uboot starts Power LED OFF, Status LED
boot but is not able to boot operating Amber-solid.
system partitions, or MM fully boots and
detects a failure such as the switch failure,
MM failure, VR failure, and so on).
The combo Status LED/Button on the rear of the chassis changes to AMBER when
any of the Front Panel iconic indicators shows AMBER. When the EC module is in
Identify State, the combo Status LED/Button shall always blink BLUE and overrides
any other Status LED state.
Profiles:
The restore chassis feature enables you to restore a chassis configuration, if the
backup is created from the same chassis. You must have chassis administrator
privileges to restore the chassis configuration.
The backup chassis feature enables you to create a backup of the chassis
configuration, including those of its member sleds and use it later. You must have
chassis administrator privileges to back up the chassis configuration.
Failover:
When a Management Module fails, hardware control from the chassis is
temporarily offline while new Management Module goes active.
While the Management Module is offline or failing over, no new sleds, IOMs, or
PSUs can be powered on.
When a new active Management Module is declared, the chassis returns to normal
operation.
1. Power off the system, including any attached peripherals, and disconnect the
system from the electrical outlet.
2. Remove the Management Module.
3. Move the jumper on the Management Module from pins 2 and 3 to pin-1 and 2.
4. Replace the Management Module.
5. Reconnect the system to its electrical outlet, and turn on the system, including
any attached peripherals.
6. Power off the system, including any attached peripherals, and disconnect the
system from the electrical outlet.
7. Remove the Management Module.
8. Move the jumper on the Management Module from pin-1 and 2 to pins 2 and 3.
9. Replace the Management Module.
10. Reconnect the system to its electrical outlet, and turn on the system, including
any attached peripherals.
11. Assign a new system and/or setup Management Module password.
The existing Management Module passwords are not disabled (erased) until the
system boots with the jumper on pin-1 and 2.
However, before you assign a new system and/or setup Management Module
password, you must move the jumper back to pins 2 and 3.
Important:
If you assign a new system and/or setup Management Module
password with the jumper on pins 1 and 2, the system disables the
new Management Module password the next time that it boots.
NOTE: If storage or other peripheral sleds are installed that do not support Fabric
A/B connections, the corresponding ports on the I/O modules mapping to these
slots are not active. This is also the case for any compute sled that does not have a
Mezzanine that is installed for the particular IOM (A or B) in the system.
NOTES:
IOM Redundancy is for multipaths to storage where if an IOM fails the other IOM
will still have a direct path to the drives mapped from the storage array.
MX5000s Locations
Item Description
1 Information tag
5 Release lever
6 LED indicators
MX5000s
Identification Flashes blue for one second and then Indicates that the SAS
indicator turns off. IOM is active.
Flashes amber for 2 seconds and turns SAS IOM card failure.
off.
Force Active
– Users can perform a failover of the IOM by performing a Force Active on the
Passive or Degraded Switch.
– Performing a Force Active on the SAS IOM is considered a disruptive
operation and should only be used as necessary.
– When the user performs a Force Active, the SAS IOM is marked Active and
the associated Storage Configuration is applied to the chassis.
Clear Config
Troubleshooting Power
Sled Overview:
- Slot Number/Type
- Device Name
- Priority
- Power Consumption
- Minimum Power
- Peak Power
Fabric Overview:
- Slot
- Device name
- Power Consumption
- Minimum Power
- Peak Power
1. Press the power button to ensure that your system is turned on. If the power
indicator does not glow when the power button is pressed, press the power
button firmly.
2. Plug in another working power supply unit to ensure that the system board is
not faulty.
3. Ensure that no loose connections exist. For example, loose power cables.
4. Ensure that the power source meets applicable standards.
5. Ensure that there are no short circuits.
6. Have a qualified electrician check the line voltage to ensure that it meets the
needed specifications.
Troubleshooting power supply unit problems
MX7000 Redundancy
Get the same power on results, regardless of what is turned on first. After an
AC Cycle, allows all devices that were on to power on again.
When possible, warn and turn on, rather than denying power-on.
The MX7000 chassis supports various types of redundancy. The user may select
these through the user interface. These are described in the following subsections.
Grid Redundancy
This mode provides for redundancy at the grid (i.e., branch) level. Configurations
for this mode are represented by “N+N,” where N is the number of power supplies
on each grid and the total available power is N * PSU capacity. Supported
configurations for this mode should include 1+1, 2+2 and 3+3.
PSU Redundancy
This mode provides for redundancy at the PSU level. Configurations for this mode
are represented by “N+1,” where N is the number of power supplies required to
support the configuration (the total available power is N * PSU capacity). Supported
configurations for this mode should include 1+1, 2+1, 3+1, 4+1, and 5+1. This is
typically used in cases where the user does not have separate grids to feed the
chassis, or the power that is required by the configuration exceeds the capacity of
three PSUs.
No Redundancy
This mode has no provision for redundancy. If a PSU fails, the system may fault
due to excess current supplied by the remaining PSUs. This mode is typically used
in cases where the user does not have enough PSUs to support the installed
configuration of compute sleds, IOMs, and so on.
PSU Pairing
The PSUs are configured to support dual grid AC sources, and are connected in
such a way that PSUs 1–3 are associated with one grid, while PSUs 4–6 are
associated with the second grid. The AC plugs on the rear of the system are set up
as three plugs on each side of the chassis, for obvious grouping to a rack PDU.
The PSUs support Rapid-ON which enables a PSU to go into a “sleep” mode while
system power consumption is low enough such that its partner provides all the
power. This enables for more efficient power utilization when overall system power
is low. The partner can then “wake” the sleeping PSU by asserting a WAKE signal
when system power draw increases.
Each PSU has an LED with “power” symbol that lights up when power cable at rear
is inserted. Then there is an LED for AC. Both of those seem to indicate that AC
power is connected, and it is not clear why both are needed.
NOTE: After installing a power supply unit, allow several seconds for the system to
recognize the power supply unit and determine if it is working properly.
PSU mismatch ON for one second, five 4 Hz blinks and OFF (nonrepeating
cycle).
PSU firmware Default state, 4-Hz blink during the update, return to default
update state.
AC source available ON
AC source OFF
unavailable or
power cable
unplugged
DC output available ON
DC output OFF
unavailable
Troubleshooting Temperature
System cover, cooling shroud, EMI filler panel, memory module blanks, or back
filler bracket is not removed.
Ambient temperature is not higher that the system-specific ambient
temperature.
Ensure that the Sleds, IOMs, and PSU blanks are installed in the empty bays.
External airflow is not obstructed.
A cooling fan is not removed or has failed.
The expansion card installation guidelines have been followed.
Troubleshooting Fans
There are two different types of fan modules that are used in the MX7000 chassis,
60-mm fans in the front and 80-mm fans in the rear.
The fan modules are hot-swappable, with 3+1 redundancy for the front fans, and
4+1 redundancy for the rear fans.
The system should always be populated with the full set of fans to support the
airflow requirements of the chassis.
The rear fan board, or 80 mm Fan Interposer, sits parallel to the chassis floor. The
rear fans plug into this board from the rear of the chassis. This board connects
internally to the vertical power distribution board, through a signal power cable.
This cable supports power for the five rear fan modules, up to a maximum of 225 W
per fan.
Within the MX7000 chassis, there are several infrastructure boards that are
interconnected through cables to provide the power and communication paths
between modules. These boards are supported within the chassis sheet metal and
are not hot-serviceable.
The Main Distribution Board (DB) is the most complex of all the internal
infrastructure boards. There are no active components, however there are many
signals that are interconnected on this board.
From the front of the chassis, the six power supplies plug into connectors on the
bottom side of the Main DB. Also, from the front of the chassis the eight compute
sleds plug into this board for power, management signals, and other fabric
connections.
From the rear of the chassis the Management Modules plug in to this board to
provide overall chassis management and control. Also, from the rear of the chassis
the two fabric I/O Modules (Fabric C IOMs) plug into this board.
The vertical power distribution board sits above the main distribution board, which
is oriented perpendicular to the chassis floor. There are no active components, but
many signals connections on this board. This board connects to the main
distribution board with a flex cable, and power and ground bus bars. From the front
of the chassis, the four cooling fans plug into the vertical power distribution board.
Also, from the rear of the chassis the four fabric I/O Modules (Fabrics A/B IOMs)
plug into this board for power, management signals, and other fabric connections.
A cable connects this board to the rear fan board.
Some redundant components are hot plug serviceable, enabling zero system
downtime for failures and serviceability. The other components are cold service,
meaning that the system must be powered down for the component to be replaced,
creating downtime and impacting availability of the system.
Lab Activity
You will now complete the PowerEdge MX7000 Troubleshooting lab exercise.
iDRAC Health
The iDRAC Dashboard should be one of the first places that you look when
troubleshooting. The System Health section should give you a high-level overview
after which you can then examine to gather more detail.
Also the Virtual Console can be useful for taking control of the server.
iDRAC Logs
The iDRAC creates an error log against any issues it detects in the system. These
error logs are a vital part of any troubleshooting procedure because it allows the
administrator to look back and observe what has been happening within the
system.
For example, a server may have an intermittent issue that never occurs when the
support technician is onsite. Using the logs, the technician can see what errors
occurred together with the time that the error occurred, a message ID and a
summary message. Also, the error can be expanded to display additional
information to help diagnose the issue.
Also, using the icon, the technician can quickly scan down to locate any warning or
critical failures, plus a filter can be used to narrow down the error codes to a
particular field such as storage.
iDRAC Troubleshooting
The iDRAC contains a troubleshooting section which is available when you select
Troubleshooting from the system tree, as you can see from this screenshot.
These features are covered in the following slides.
NOTE: The Video Capture feature is available with iDRAC Enterprise license.
To configure the boot capture video settings, select one of the following options
and click Apply.
The boot capture timestamp is the time that the boot capture sequence is
completed - that is either when the boot capture file size has reached 2 MB or
when the host system is rebooted.
The list displays the currently active boot capture file. While the update is in-
progress, click Refresh to view the latest timestamp for the boot capture file.
You can play the files directly from iDRAC or save it to a location on your system.
Use this page to view the last system POST code (in hexadecimal) prior to booting
the operating system of the managed system. POST Code helps to detect pre-
video stops responding, report fatal errors, and analyze the system failures during
BIOS POST, particularly the No POST No Video situations.
The fatal error codes are used to report all the fatal POST errors.
The Intrusion relates to the chassis intrusion probes (or switches) and provides
information such as the server cover being removed.
The Last Crash Screen displays a view of the most recent crash screen, which
provides information about the events leading to the system crash. Last system
crash information is saved in the iDRAC memory and is remotely accessible. The
Last Crash Screen feature is available with iDRAC Express and Enterprise
licenses.
The Reset iDRAC will perform a reset and not lose any settings - A normal reboot
operation is performed on the iDRAC. After reboot, refresh the browser to
reconnect to iDRAC.
The Reset iDRAC to Default Settings gives you the following options:
You can perform a hard or soft iDRAC restart without turning off the server:
Hard reset — On the server, press and hold the ID LED button for 15 seconds.
Soft restart — Using iDRAC Web interface or RACADM.
In the Command field, enter one of the following diagnostic commands and click
Submit. The result is displayed.
Command Function
Use this page to generate a SupportAssist collection of the server and then export
the collection to a location on the management station (local) or to a shared
network location such as Common Internet File System (CIFS) or Network File
Share (NFS).
The collection is generated in the standard ZIP format. You can send this collection
to technical support for troubleshooting or inventory collection.
Automatic:
Using iDRAC Service Module that automatically invokes the OS Collector
tool.
Manual:
Lifecycle Controller allows you to create a copy (image file) of the server's profile
on the vFlash SD card installed on the server. The server profile that is backed up
on a vFlash SD card contains the server component configuration and firmware
that is installed on various components on the server where the card is installed.
The backup image file does not contain any operating system or hard-disk drive
data.
For better security, Lifecycle Controller allows you to remove the vFlash SD card
and keep it in a safe location, or you can copy the server profile (backup image)
that is stored on the vFlash SD card to any USB drive or an external network share.
Whenever the firmware is corrupted, configuration changes are incorrect, or the
system board is replaced, you can use the backup image to restore the server to its
previously stored profile. The following features are available to maintain a server:
Allows you to import an iDRAC license from a network share or a USB drive.
If the current adapter on the system is NPAR enabled and is replaced with a
new adapter, after the host server is turned on, press <F2> and select
System Setup > Device Settings and ensure that the NPAR is enabled.
NPAR must be enabled on the new adapter before using the Part
Replacement feature.
To apply part firmware and configuration to replaced parts:
Disabled
– The feature is disabled, and the current configuration is not applied if a
part is replaced.
Apply always
– The feature is enabled (default setting), and the current configuration is
applied if a part is replaced.
Apply only if firmware match
– The feature is enabled, and the current configuration is applied only if the
current firmware matches with the firmware of a replaced part.
Lab Activities
Please inform your instructor before completing the MX7000 Firmware Update lab
exercise.
Module Summary
Summary
Course Conclusion
Course Conclusion
Final Questions