Dell M1000e Install Admin Troubleshooting

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 78

Dell M1000e Install Admin Troubleshooting

The M1000e chassis can hold:

 Either 8 full-height blades or 16 half-height blades, or 32 quarter-height blades

 Blade Portfolio
 M-Series blades combine powerful computing capabilities with optimized
connectivity that works virtually with any network, storage, or management
infrastructure. The image shows the latest blade servers that are supported by
the M1000e chassis.

PowerEdge Server Numbering


A PowerEdge server number is an alphanumeric combination, where the first letter
indicates the type of chassis. It can be either a Rack, Tower, or Modular chassis.
The first digit specifies the power of the server, based on the number of sockets. The
second digit helps to identify the generation of the server, starting from the
10th generation. The last digit helps to identify the type of processor.
Simplified Manageability with CMC
The Dell Chassis Management Controller (CMC) is a systems management hardware
and software solution for managing multiple Dell blade chassis. It is a hot-pluggable
module that sits at the back of Dell PowerEdge M1000e chassis. The CMC has its own
microprocessor and memory, and it is powered by the modular chassis into which it is
plugged.
You can configure the M1000e chassis either with a single CMC or with redundant
CMCs. In redundant CMC configurations, if the primary CMC loses communication with
the M1000e chassis or the management network, the standby CMC takes over chassis
management.

 CMC automates and simplifies configuration of new hardware in a one-to-many


manner.
 CMC’s virtual file share enables a one-to-many deployment of updates and
images.
 FlexAddress feature replaces the factory-assigned World Wide Name/Media
Access Control (WWN/MAC) IDs with chassis-assigned WWN/MAC IDs for a particular
slot, an optional upgrade.
 Firmware update of server components such as BIOS, Network Controllers,
Storage Controllers, and so on can be done across multiple servers in the
chassis.
 CMC and Lifecycle Controller also provide the ability to update various
components on one-to-many basis, without additional software.

M1000e Power Subsystem


The chassis takes in AC power and distributes the load across all active internal PSUs.
The chassis can automatically switch off any unused PSUs using the Dynamic Power
Supply Engagement (DPSE) feature.
The M1000e supports 2360 W, 2700 W, and 3000 W power supplies. The 2700W dual-
voltage power supply is capable of supporting both 208–240V and 110–120V.
Dell does not support mixing 230 V and 115 V within the same chassis, but does
support any combination of of 2360 W, 2700 W, and 3000 W power supplies while
running them both at 230 V.
The following redundancy policies are configurable with or without dynamic PSU
engagement:
In the N+N power supply configuration, the system provides protection against AC grid loss or
power supply failures. If one power grid fails, three power supplies lose their AC source, and the
three power supplies on the other grid remain powered, providing sufficient power for the
system to continue running. PSUs in slots 1, 2 and 3 are in the first grid while PSUs in slots 4, 5
and 6 are in the second grid.

In the N+1 configuration only power supply failures are protected, not grid failures. The
likelihood of multiple power supplies failing at the same time is remote.

In the N+0 configuration there is no protections against power failure.

Power Management Features


The M1000e System offers many advanced power managements features controlled by
the CMC and negotiated along with the iDRAC on every server module.
Power redundancy policy is a configurable set of properties that determine how Dell
CMC manages power to the chassis.
Power Budgeting and Power Capping are two power management features.
The M1000e includes an advanced power budgeting feature that is controlled by the
CMC and negotiated along with the iDRAC in every blade server.

 When a blade server is installed and its power button is pressed, the button
starts to slowly blink green.
During this time the iDRAC interrogates the blade server’s components to calculate the
power requirements, and these requirements are fed back to the CMC.
If enough power is available, the blade server will then power up, and blade server’s
power button will display solid green.
 The iDRAC provides the CMC with its power envelope requirements before
powering up the blade server. The power envelope consists of the maximum and
minimum power requirements that could keep the server operating. iDRAC’s initial
estimate is based on a worst-case model where all components in the blade server
draw maximum power and are often higher than the actual blade server requirements.

The M1000e allows you to specify a System Input power cap to ensure that the overall
chassis AC power draw stays under a given threshold.
The CMC first ensures enough power is available to run the

 Fans
 I/O Modules
 iKVM (if present)
 CMC
This power allocation is called the Input Power Allocated to Chassis infrastructure.
Power cap is set at the chassis level for our blade servers.

Servers are allocated power based on their Server Priority setting, with priority 1 servers
getting maximum power, priority 2 servers getting power after priority 1 servers, and so
on. Lower priority servers may get less power than priority 1 servers. It is based on
System Input Max Power Capacity and the user-configured setting of System Input
Power Cap.
If total power budget stays below the value of the System Input Power Cap, the CMC
allocates servers a value less than their maximum requested power.
The system administrator can also set priorities for each server module. The priority
works in conjunction with the CMC power budgeting and iDRAC power monitoring to
insure that the lowest priority blades are the first to enter any power optimization mode.

Dynamic Power Supply Engagement


Power utilization is monitored in M1000e and up to four PSUs could be moved to
Standby state depending on the overall system load.
When it's enabled and set to a Power Supply Redundancy configuration, the enclosure
always keeps an extra PSU powered on and marked Online in addition to the PSUs
required to turn on the enclosure (min 2 to power on the enclosure).
In a six PSU configuration, a minimum of two power supply units are always turned on.
Since an enclosure in the Power Supply Redundancy configuration always has an extra
PSU engaged, the enclosure can withstand the loss of one online PSU. The enclosure
can also have enough power for the installed server modules. The loss of the online
PSU causes a standby PSU to come online. Simultaneous failure of multiple PSUs may
result in the loss of power to some server modules while the standby PSUs are turning
on.
Power supplies in M1000e
When Dynamic Power Supply Engagement is enabled and set to Grid Redundancy
configuration all power supplies are engaged when the chassis is turned on. Power
utilization is monitored, and if system configuration and power utilization enable, PSUs
are moved to the Standby state. The Online status of PSUs in a grid mirrors that of the
other grid. Hence, the enclosure can sustain the loss of power to an entire grid without
interruption of power to the enclosure.
An increase in power demand in the Grid Redundancy configuration causes the
engagement of PSUs from the Standby state. This maintains the mirrored configuration
that is needed for dual-grid redundancy.

Fan Redundancy
M1000e Chassis has 9 standard hot pluggable, redundant fan modules.

The 9 fans are distributed evenly across the enclosure. The speed of each fan is
individually managed by the CMC. Fans are individually balanced to ensure smooth
operation and maximum throughput at lower speeds.

 Fans are N+1 redundant. This means that failure of any single fan will not impact
system uptime or reliability.
o Any fan failure will immediately be visible through the chassis LCD or the
CMC GUI.
o Failure of more than one fan will not automatically result in shutting down
of blade servers.
This is because the blade servers have their own self-protection mechanisms to prevent
them from running too hot. Multiple fans failure can be because of the configuration,
ambient temperature, and its workload.
 The CMC controls every element of the cooling subsystem.
However, the CMC depends on the server iDRACs to feed back temperature
information from within the blades, and the CMC is controlled by firmware.
Dynamic fan control results in to approximately 25% less airflow than 16 x 1U servers.
This results in less noise and longer life cycles for the cooling components.

 Airflow Management Design


 Computational Fluid Dynamics (CFD) is used to design the laminar airflow within the
chassis. It uses traditional front-to-back cooling.

The Server Modules are cooled with traditional front‐to‐back cooling. The front of the
system is dominated by an inlet area for the individual server modules.
The cooling process is as follows:

 Air passes through the server modules and then through vent holes in the mid-
plane.
 Air is drawn into the fans, which exhaust the air from the chassis.

The I/O Modules use a bypass duct to draw ambient air from the front of the system to the I/O
Module inlet. This duct is located above the server modules. This cool air is then drawn down
through the I/O Modules in a top to bottom flow path and into the plenum between the midplane
and fans, from where it is exhausted from the system.

The Power Supplies that are located in the rear of the system, use basic front‐to‐back
cooling. But they draw their inlet air from a duct located beneath the server modules.
This insures that the power supplies receive ambient temperature air.

iKVM Module
The local access KVM module for M1000e server chassis is called the Avocent®
Integrated KVM Switch Module (iKVM).
Some of the key features of iKVM are:

 It acts as an analog keyboard, video, and mouse switch that plugs into the
chassis and active CMC's CLI.
 Uses the On-Screen Configuration and Reporting (OSCAR) user interface to
select one of the servers or the Dell CMC command line you want to access.
 It assigns an order of precedence to each type of connection, so that when there
are multiple connections, only one connection is available while others are disabled.
The order of precedence for iKVM connections is as follows:
 Front panel
 Analog Console Interface
 Rear panel

Important: The ports on the control panel on the front of the chassis are
designed specifically for the iKVM, which is an optional module. If you do not
have the iKVM, you cannot use the front control panel ports.

Midplane
The Dell PowerEdge M1000e mid-plane is a passive board (no electrical components
other than connectors) that serves as the conduit for power, fabric connectivity, and
system management infrastructure. Also, it enables airflow paths for the front to back
cooling system through ventilation holes.

All M1000e mid-plane routing is fully isolated, supporting all chassis power, fabric,
system management, and fault-tolerance requirements.
The version 1.1 midplane allows for 10gb connectivity on Fabric A, while the original midplane
version 1.0 only supports 1gb connectivity.

Introduction to Fabrics
A Fabric is defined as a method of encoding, transporting, and synchronizing data
between multiple devices. Examples of Fabrics are GbE, Fibre Channel, or InfiniBand.
Fabrics are carried inside the PowerEdge M1000e system between server modules and
IOMs through the mid-plane. They are also carried to the outside world through the
physical copper or optical interfaces on the IOMs.
The fabrics are independent of each other.
A Fabric is defined as a method of encoding, transporting, and synchronizing data
between multiple devices. Examples of Fabrics are GbE, Fibre Channel, or InfiniBand.
Fabrics are carried inside the PowerEdge M1000e system between server modules and
IOMs through the mid-plane. They are also carried to the outside world through the
physical copper or optical interfaces on the IOMs.
The fabrics are independent of each other.
Fabric A is a redundant Gigabit Ethernet (GbE) fabric, supporting I/O module slots A1 and A2.
The integrated Ethernet controllers in each blade dictate Fabric A as a 1-to-10Gbps Ethernet-
only fabric.

Modules that are designed specifically for Fabric B or Fabric C cannot be installed in slots A1 or
A2, as indicated by the color-coded labeling on the faceplate of each module.

Fabric B is a 1-to-40Gbps redundant fabric, supporting I/O module slots B1 and B2. Fabric B
currently supports 1GbE or 10GbE, DDR/QDR/FDR InfiniBand, and 4Gbps, 8Gbps, or 16Gbps
FC modules.

To communicate with an I/O module in the Fabric B slots, a blade must have a matching
mezzanine card installed in a Fabric B mezzanine card location. Modules designed for Fabric A
may also be installed in the Fabric B slots.

Fabric C is a 1-to-40Gbps redundant fabric, supporting I/O module slots C1 and C2.
Fabric C currently supports 1GbE or 10GbE, DDR/QDR/FDR InfiniBand, and 4Gbps,
8Gbps, or 16Gbps FC modules.
To communicate with an I/O module in the Fabric C slots, a blade must have a
matching mezzanine card installed in a Fabric C mezzanine card location. Modules
designed for Fabric A may also be installed in the Fabric C slots.

Fabric Mapping
The M1000e I/O is fully scalable to current generations of server modules and I/O
Modules. There are three redundant multi‐lane fabrics in the system, as illustrated in the
image.
Fabric A is dedicated to Gigabit Ethernet.
Fabrics B and C are identical, fully customizable fabrics, which are routed as two sets of
four lanes from mezzanine cards on the server modules to the IOMs in the chassis rear.
Supported bandwidth ranges from 1 Gbps to 10 Gbps per lane depending on the fabric
type used.
In the adjoining figure:

 Fabric A connects to the integrated Ethernet LAN on Motherboard (LOM) or


Network Daughter Card (NDC) on each blade server. Therefore, Fabric A can only be
used for Ethernet traffic.
 Fabrics B and C both support Ethernet, Fibre Channel, or InfiniBand technology
modules in IOM slots B1, B2, C1, and C2.
 To communicate with an IOM in the B1 and B2 (Fabric B) or C1 and C2 (Fabric
C) slots, a blade must have a matching mezzanine card that is installed in the B or C
mezzanine card position within the blade server.
I/O Modules (IOMs)
The M1000e enclosure supports three layers of I/O Fabric.

 Fabric A only supports Ethernet


 Fabrics B and C support Ethernet, Fibre, and InfiniBand
You can install up to six hot-swappable I/O modules in the enclosure, using:

 Fibre Channel switches


 Fibre-Channel pass-throughs
 Infiniband switches
 Ethernet switches
 Ethernet passthrough modules
The illustration following shows the fabric layout in the chassis.
I/O Module Stacking
A powerful feature of M1000e, known as Stacking helps to stack multiple switches
together.

 In switch stacking,
 All switches in a stack share a single management console and forward
packets between their ports as a single switch.
 A stacked set of switches appear to all external connections as a single
logical switch
 Modular switch stacking can be done for:
 Higher Server Throughput
The prime benefit of this cabling scheme is that it enables increased throughput. The
uplink connection from one switch in the stack forwards network traffic from all the other
switches in the stack.
Configure using stacking cables and connect:

 Port xg2 of the stacking module of each switch to port xg1 of the stacking module
of the neighboring switch
 The first and last switch in the stack to complete the loop
Modular stacking configuration reduces the number of uplinks that need to be
connected to the stack
Increased Server Availability

To ensure maximum uptime of network resources, servers should be connected to:

 The network via fully redundant paths


 At least two logically discrete switches
Each switch should connect separately to the network.

When a stack spans across two or more chassis, the number of required uplinks per
server can be greatly reduced, minimizing total cost of ownership (TCO). Gain network
and application performance by allowing peer-to-peer network traffic between servers
connected to the same switch stack to transit across the stack without passing through
the distribution switches.

This figure shows two M6220 switch stacks deployed across two chassis. Each switch
stack should have its own set of uplinks to the same distribution switches as the other
switch stack. Each server should have a teamed pair of Ethernet ports configured for
switch fault tolerance, with one port connected to each stack.
The pair of stacks should be configured with the same connectivity to the servers and to
the distribution switches upstream.
CMC Overview
The Dell Chassis Management Controller (CMC) is a systems management hardware
and software solution for managing the M1000e chassis and installed components. It is
a hot-pluggable module that is installed in the rear of the Dell PowerEdge M1000e
chassis.
The M1000e must have at least one CMC and can support an optional redundant
module. Each CMC occupies a slot accessible at the rear of the chassis. Redundancy is
provided in an Active–Standby pairing of the modules, and failover occurs when the
active module has failed or degraded. The CMC interface ports are the stacking port,
10/100/1000 Ethernet ports, and one serial port. The CMC serial port interface provides
common management of up to six I/O modules through a single connection.

CMC Features
Some of the important features of CMC are:

 It provides remote management capabilities and power control functions for


modular server chassis and all its components. The secondary optional CMC module
provides failover protection and hot-plug replacement.
 CMC can be configured to send email alerts or SNMP trap alerts for warnings or
errors that are related to temperatures, hardware misconfigurations, power outages,
and fan speeds.
 CMC has its own microprocessor and memory which is powered by the modular
chassis. It also provides secure remote management access to the chassis and
installed modules.

 The CMC provides a mechanism for centralized configuration of the following:


 The enclosure’s network, security, and power settings
 I/O switches and iDRAC network settings
 Checks I/O fabric consistency between the I/O modules and blades and
disables components if necessary to protect the system hardware
 Enclosure-level real-time automatic power and thermal management.
 Monitors and manages system power requirements
 Monitors and automatically controls cooling fans based on actual ambient
and internal temperature measurements

CMC Managed Cooling


CMC Failover
Failover is transfer of control when failure or fault is detected

 A CMC Failover can occur when:


 The active CMC fails or is removed.
 The network cable is removed from the active CMC.
 A CMC firmware flash is initiated on the active CMC.
 The racadm cmcchangeover or the racadm racreset commands are issued.
 What happens when a CMC failover occurs:
 All external iDRAC sessions and all active CMC sessions disconnect and
must be reestablished.
 Internal iDRAC sessions move to the new active CMC.
 Blades and server nodes continue to operate.
The active CMC will failover for all of the reasons that are stated above. When the
active CMC fails, it sometimes takes a couple of minutes for the failover to be
completed, during which the fans run at 100 percent in fail-safe mode.

CMC Fail-Safe Mode


The M1000e chassis enables the fail-safe mode to protect the blades/ server nodes and
I/O modules from failures. The fail-safe mode is enabled when no CMC is in control of
the chassis.

 During the CMC failover period or single CMC management loss:


 You cannot power on newly installed blades/ server nodes.
 Existing blades/ server nodes cannot be accessed remotely.
 Chassis cooling fans run at 100% for thermal protection of the
components.
 Blade/ server nodes performance reduces to limit power consumption until
management of the CMC is restored.

FlexAddress
FlexAddress is an optional feature on M1000e which overrides server-assigned
addresses with chassis-assigned addresses.

 FlexAddress enables the CMC to assign WWN/MAC addresses to a particular


slot and override the server factory-assigned addresses.
 If the server node is replaced, the slot-based WWN/MAC addresses
remain the same.
 This feature eliminates the need to reconfigure Ethernet network
management tools and SAN resources for a new server node.
FlexAddress is delivered on a Secure Digital (SD) card that must be inserted into the
CMC to activate the feature.
If a blade server is replaced, the FlexAddress for the slot remain the same for the given
server slot. If the blade server is inserted in a new slot or chassis, the server-assigned
WWN/MAC is used unless that chassis has the FlexAddress feature that is enabled for
the new slot. If you remove the blade server, it reverts to the server-assigned address.
You do not need to reconfigure deployment frameworks, DHCP servers, and routers for
various fabrics for identifying the new server.
FlexAddress Assignment in CMC
Every blade server is assigned unique WWN and/or MAC addresses as part of the
manufacturing process.
Without FlexAddress, if a blade server had to be replaced with another blade server,
then

 The WWN/MAC addresses are changed


 Ethernet network management tools and SAN resources have to be reconfigured
to identify the new blade server
FlexAddress enables the CMC to assign WWN/MAC addresses to a particular slot and
override the factory addresses. Therefore, if the blade server is replaced, the slot-based
WWN/MAC addresses remain the same. This feature eliminates the need to reconfigure
Ethernet network management tools and SAN resources for a new blade server.
Also, the override action occurs only when a server node is inserted in a FlexAddress-
enabled chassis; no permanent changes are made to the blade server. If a blade
server is moved to a chassis that does not support FlexAddress, the factory-assigned
WWN/MAC addresses are used.
Extended Storage- FlexAddress SD Cards
If the M1000e chassis is shipped with an optional second CMC. The second CMC has
an SD card that supports only Extended Storage.
When are SD cards needed?

 Optional SD card storage is required for holding update images larger than 48
MB (iDRAC firmware and driver packs).
 SD card storage is required to store the BIOS configurations.
 If the configuration is a redundant CMC, then both CMCs need the storage
option.
 The SD card storage can be satisfied with either a FlexAddress or a CMC
Extended Storage card.

Daisy-Chaining CMCs
CMC has a second Ethernet port for connection to other CMCs in the rack. CMC connects to
the management network to manage all blade servers. This saves port consumption on external
switches. If you have multiple chassis in a rack, you can reduce the number of connections to
the management network by daisy-chaining chassis together. This reduces the connections that
are required to one.

Important: Never connect both ports on a single CMC to the same external


switch, as this will cause loops in your network.
In the adjacent illustration, the CMC has two RJ45 Ethernet ports, labeled Gb1 (the uplink port)
and STK/Gb2 (the stacking or cable consolidation port). With basic cabling, you can connect the
Gb1 port to the management network and leave the STK/Gb2 port unused.

When daisy-chaining chassis together, Gb1 is the uplink port and STK/Gb2 is the stacking
(cable consolidation) port. Connect the Gb1 ports to the management network or to the
STK/Gb2 port of the CMC in a chassis that is closer to the network. You must connect the
STK/Gb2 port only to a Gb1 port further from the chain or network.

Caution: Connecting the STK/Gb2 port to the management network without first
configuring for redundancy in the CMC can have unpredictable results. Cabling
Gb1 and STK/Gb2 to the same network (broadcast domain) can cause a
broadcast storm.

Daisy-Chain Cabling Guidelines


Cabling guidelines to daisy chain CMC modules from chassis to chassis:

 CMC Ethernet port Gb1 is the uplink port. It is the uplink to the management
network, or used to connect to the STK in the adjacent chassis.
 The CMC Ethernet port that is labeled STK is the daisy chain port. It connects
only to CMC port Gb1 on the adjacent chassis. Do not connect this cable directly to the
management network.
 Up to four chassis units can be daisy chained.
 Chassis units can be daisy chained in both redundant and non-redundant
deployments:
 In a redundant CMC deployment, cable together all CMC modules in the
CMC primary slots. Cable together all CMC modules in the CMC secondary slots. Do
not connect the primary daisy chain with the secondary daisy chain (do not “cross-
cable” the two sets of CMCs).
 In a non-redundant CMC, cable together all CMC modules in the CMC
primary slots.

 Multi-chassis Management
 Multi-chassis management is the capability to select chassis configuration properties
from Lead Chassis and push those properties to group of chassis or Members.
 It is done for admin reasons, and user can configure the lead chassis and then copy
those settings and propagate them to the member chassis.
 In Chassis Properties Propagation, the admin can select the categories of lead
configuration properties to be propagated to member chassis. In the setting categories
you can choose what you want identically configured, across all members of the chassis
group. For example, if you select Logging and Alerting Properties category, this enables
all chassis in the group to share the logging and alerting configuration settings of the
lead chassis.

 RACADM Commands
 The Dell Remote Access Controller Admin (RACADM) utility is a command line tool that
enables for remote or local management of Dell Servers using the iDRAC or DRAC.
RACADM provides similar functionality to the iDRAC/DRAC Graphical User Interface
(GUI). The Dell Chassis Management Controller (CMC) can also be managed remotely
with RACADM.

RACADM commands can be run remotely from a management station and/or locally on
the managed system.

RACADM commands enable you to view managed system information, perform power
operations on the managed system, perform firmware updates, configure settings and
more. Because RACADM is run from a command line interface (CLI), system
administrators can create scripts that control and update Dell systems in a one-to-many
fashion.

Updating Firmware
A vital part of maintaining a healthy chassis is the ability to update the CMC version.
Notice on this screen that you can update either the active or standby CMCs individually
or together. If you update them both together, you will be navigated to the version
location, and then it is transferred across to the CMCs, where it is validated. Once the
validation is complete, the update begins on both and is displayed on this screen.

During this time, the CMCs maintain control of the chassis. However, at the end, both
CMCs have to reset, at which time the fans ramp up to 100 percent until one of the
CMCs comes back online and takes control. It now becomes the active CMC and looks
after the chassis, bringing up the standby CMC gradually.

Therefore, it is normal to get an error stating that CMC redundancy is lost for a few
minutes after you hear the fans ramp down from 100 percent. Also notice that you can
update the iKVM firmware and initiate updating the iDRAC firmware from this page.
The following software components are included with CMC firmware package:
 Compiled CMC firmware code and data
 Web interface, JPEG, and other user interface data files
 Default configuration files

Power on System
Powering on the system involves following steps.

 Powering On Chassis
Pressing the chassis power button turns on all components that are related to the
chassis, and it affects the main power bus within the chassis. Components such as the
iDRACs, IOMs, and iKVM will begin to power up. The blade server will not power on
immediately.
To manually turn off the system, push and hold the power button for 10 seconds to
forcefully shut down the chassis and all systems.
 Powering On Blades
Press the power button on the chassis. The power indicator should display a green
LED.
Blades can be powered on/off manually with the power button that is located on the
front of each blade server. Pressing the button will power on the blade and begin POST.
Pressing and holding the power button for approximately 10 seconds forcefully powers
off the blade.

Initial Configuration
The CMC’s default IP address is 192.168.0.120.
Initial Configuration involves assigning an IP to the CMC. There are five ways to
perform initial configuration.
 Dynamic Host Configuration Protocol (DHCP) - The CMC retrieves IP
configuration (IP Address, mask, and gateway) automatically from a DHCP server on
your network. The CMC will always have a unique IP Address allotted over your
network.
 LCD Configuration Wizard - You can configure the CMC using LCD
Configuration Wizard only when the CMC is deployed or the default password is
changed. If the password is not changed, using LCD to reconfigure the CMC, might
cause possible security risk.
 RACADM CLI using a null modem cable - Serial connection using a null
modem cable.
o Cable: Null Modem
o Bits per second: 115200
o Data bits: 8
o Parity: None
o Stop bits: 1
o Flow control: None
 RACADM CLI using iKVM - If the chassis has the iKVM, press <Print Screen>
and select blade number 17. Blade number 17 is a direct local connection to the CMC.
 Web GUI - Web GUI provides remote access to CMC using a graphical user
interface. The Web interface is built into the CMC firmware and is accessed through the
NIC interface from a supported web browser on the management station.

14G Memory Overview


Some of the key features of the PowerEdge 14G Memory subsystem are:

 It can support 12 DIMMs per installed CPU.


 24 memory sockets split into two sets of 12 sockets, one set per processor
 The minimum RAM size is 1x 8-GB memory module per processor.
 Supported DIMM types are:
 Registered DIMMs (RDIMM)
 Load Reduced DIMMs (LRDIMM)
 Nonvolatile dual inline DIMM-Ns (NVDIMM-Ns)
 It provides optimized and mirrored mode support (Advanced ECC Mode not
supported).

BOSS
In the 14th generation of PowerEdge server, Dell EMC has introduced Boot Optimized Storage
Solution (BOSS) as a new boot device. The BOSS device provides a hardware mirror option so
that the user can create a RAID 1 volume. This helps to install the operating system on the LUN
from a redundancy viewpoint. Besides the RAID1 volume, this controller can operate in pass-
through mode (Non-Raid).

NVDIMMs
NVDIMM is nonvolatile memory with added flash storage and a battery backup system
for all data persistence. NVDIMM Persistent Memory is a disruptive Storage Class
Memory technology that enables unprecedented performance improvement over legacy
storage technologies.
In an event of a power outage, the data is backed-up to flash thus retaining the memory
contents during the system power loss. NVDIMMs integrate nonvolatile NAND flash
memory with dynamic random access memory (DRAM) and dedicated backup power on
a single memory subsystem.

14G PowerEdge Memory Modes


In this mode, the memory controllers run independently of each other. This mode
permits a larger total memory capacity but does not support Single Device Data
Correction (SDDC) with x8-based memory modules.
It is recommended to populate all three channels with identical memory but each
channel can have different size DIMM. The larger DIMM should be installed in the first
slot and configuration has to be the same across all three channels. In a dual processor
configuration, the memory configuration for each processor must be identical. Optimizer
mode is the only mode to support mixed memory sizes.
Only Optimizer mode is used for each processor.
Memory Mirroring
The PowerEdge system supports memory mirroring, if identical memory modules are
installed in the two channels closest to the processor (CH 0 & 1). Mirroring can be
enabled in the System Setup program. The total available system memory is one-half of
the total populated memory, any operating system will only show half of the physical
memory installed.

Fabrics
A fabric is a method of encoding, transporting, and synchronizing data between multiple
devices. Fabrics are carried inside the PowerEdge M1000e system between server
modules and IOMs through the mid-plane. They are also carried to the outside world
through the physical copper or optical interfaces on the IOMs.
Fabrics are the main components in the PowerEdge M1000e architecture. The features
of fabrics are:

 It enables the M1000e chassis to pass data between the blade servers and the
IOMs.
 The M1000e chassis can house six I/O modules, enabling a greater diversity of
roles for all the enclosed blade servers.
 The six I/O slots are connected to three fabrics, and each fabric connects
to two slots.
 The fabrics are independent of each other.
 IOMs are used as pairs, with two modules servicing each server module fabric
and providing a fully redundant solution.
The fabrics in a M1000e enclosure are classified into three types. The image shows the
fabric mapping in M1000e enclosure.

 Fabric A: It connects to the integrated Ethernet LAN on Motherboard (LOM) or


Network Daughter Card (NDC) on each blade server. Therefore, Fabric A can only be
used for Ethernet traffic.
 Fabrics B: It supports Ethernet, Fibre Channel, or InfiniBand technology
modules in IOM slots B1 and B2.
 Fabrics C: It supports Ethernet, Fibre Channel, or InfiniBand technology
modules in IOM slots C1 and C2.
I/O Fabric Architecture
A Fabric is the interconnect architecture that is used by the M1000e chassis to pass
data between the blade server and IOMs.
In this architecture, Fabrics B and C are identical and fully customizable Fabrics. They
are routed as two sets of four lanes from mezzanine cards on the server modules to the
I/O Modules in the rear of the chassis. Supported bandwidth ranges from 1 Gbps to 10
Gbps per lane depending on the fabric type used.
The following image represents the I/O Fabric Architecture using half-height blades.
Network Daughter Card
The Network Daughter Card (NDC) is a custom form-factor card that contains a
complete NIC subsystem.
Some of the key features of NDC are:

 The PowerEdge blade server systems support NDC cards instead of the
traditional LAN on Motherboard (LOM) design. Some blade server models can house up
to two NDCs.
 The feature is marketed as a Dell Select NIC Adapter.
 The NDC has the typical features and behavior of a traditional LOM subsystem. It
adds the benefit of flexibility by enabling customers to choose their favored network
types, speed, and vendors.
Blade Mezzanine Cards
The M-series blades have two or four mezzanine card slots enabling the user to install
mezzanine cards. Some of the functionalities of mezzanine cards are:

 It enables the communication with fabrics B & C.


 It is designed to connect through 8-lane PCIe to the server module’s chipset in
most cases.
 Optional mezzanine card may have either one dual port ASIC with 4- or 8-lane
PCIe interfaces or dual ASICs, each with 4-lane PCIe interfaces.
 They provide support in routing the external fabrics through high-speed, 10Gbps-
capable air dielectric connector pins through the planar, and mid-plane.
 They isolate the differential pair with ground pins, and signal connector columns
are staggered to minimize signal coupling.
Important: For the blade to boot, the mezzanine card must match the fabric as
determined by the I/O module (IOM) installed in the chassis. If the card does not
match the fabric, the blade will not boot. In this situation, no damage occurs on
either the IOM or mezzanine card, but a fabric mismatch message is generated
in the chassis logs.

M-Series Blade I/O Fabrics


There are three different types of M-Series blades that are connected to I/O Fabrics.
They are:

 Quarter-Height Blades
o It requires IOM with 32 internal ports (M6348 or Dell Force10 MXL) to
connect all LOM ports on all blades
o It requires 2 x 32-port IOMs to connect the two LOM ports on each blade.
o It consists of one Fabric B or Fabric C mezzanine card.
 Half-Height Blades
o It has one Select Network Adapter or LOM.
o It consists of one Fabric B mezzanine card and one Fabric C mezzanine
card.
 Full-Height Blades
o It has two Select Network Adapters or LOMs.
o It consists of two Fabric B mezzanine cards and two Fabric C mezzanine
cards.

Platform Restore
The Platform Restore option is used to restore individual hardware parts or the system
board configurations.
This option enables you to decide what happens if a hardware part fails and needs to be
replaced. Instead of manually reconfiguring the new part, you can set the Lifecycle
Controller to automatically push the configuration back out to the new part. This
includes reflashing the firmware with the code that was on the old part.
In the case of the system board containing the Lifecycle Controller fails, you can make
use of the Server Profile, which includes a back up of all the Lifecycle Controller
configuration information onto a USB stick, hard drive, or network share. Now, when the
new system board is installed, you can access the new Lifecycle Controller and import
the Server Profile.
Port-Mapping Half-Height Blades
The port-mapping feature enables the user to redirect a communication request from
one Port (IP address and Port number) to another port on the opposite side of the
gateway. The data packets are traversed through a network gateway, such as a router
or firewall.
Each LAN on Motherboard (LOM) on a blade has two port connections. For a half-height blade
with Dual Port Adapters, only the first LOM port (NIC 1 and NIC2) is active. The second
port(NIC3 and NIC4) gets disabled during system boot. All the IOMs have the same port
mapping for half-height blades.

An IOM with 32 internal ports, will only connect with 16 internal ports when using dual port
adapters.

Important: The image on the slide displays the port mapping for fabric A. If a
mezzanine card is installed, then you can determine which internal ports on the
switch will connect to the ports on the mezzanine adapters.
For a half-height blade with Quad Port Adapters, both the LOM port connections (LOM1
and LOM2) are active.
If LOM1, connection 1 connects to an IOM adapter A1 on a port 5, then the LOM2,
connection 1 connects to IOM adapter A1 on a port 21(n+16).
Port-Mapping Full-Height Blades
For a full-height blade with Quad Port NDC Adapters, both the LOM port connections
(LOM1 and LOM2) are active. All six IOMs have the same port mapping for full-height
blades.

If LOM1, connection 1 connects to an IOM adapter A1 on a port 5, then the LOM2,


connection 1 connects to IOM adapter A1 on a port 13(n+8).

An IOM with 32 internal ports, will connect to all 16 internal ports when using quad port
adapters.
Internal Dual SD Module/vFlash SD
The Internal Dual SD Module (IDSDM) card provides two SD card slots and a USB interface
that is dedicated for the embedded hypervisor.
The vFlash is a storage device that is part of the iDRAC subsystem. The iDRAC can use the
vFlash card for storage purposes, backing up files, and other functions. iDRAC also uses the
vFlash to store system configurations and as a buffer when updating the firmware.
System Setup
To access the System Setup, reboot your server and access the System Setup utility
using the <F2> hot key, when requested.
The System Setup utility consists of three options, which include:

 System BIOS - Configuration of BIOS settings such as system information,


memory settings, processor settings, system profile settings, and so on.
 iDRAC Settings - Configuration of the iDRAC parameters can be done using
Unified Extensible Firmware Interface (UEFI). You can enable or disable various iDRAC
parameters by using the iDRAC Settings utility.
 Device Settings - Configuration of the device settings such as RAID Controller,
and Integrated NICs can be done.
Devices which conform to the Human Interface Infrastructure (HII) in UEFI Specification
are displayed under the Device Settings page.
 Storage

Dell PS-M4110 Blade Array


The M4110 is a storage array in blade form factor.
The key components of M4110 are:

 EqualLogic iSCSI SAN for the Dell M1000e Blade chassis


 14 hot-plug 2.5” disk drives
 Redundant 10Gb connections available through Fabric A or B
Important: Maximum of four M4110 arrays can be installed in a single M1000e
chassis.

Storage Array Environments


Typical use cases for a storage blade array include:

 Virtualization environments
 SAN applications (Exchange, database)
 High-performance cluster and grid environments
 Front-end applications (web applications/Citrix/terminal services)
 File sharing access
 Web page serving and caching
 SSL encrypting of web communication
 Audio and video streaming

The M1000e CMC manages the PS-4110.


PS-M4110 Front View
The power LED is for status only. It is not a button. The M4110 will power up and down
with the M1000e simultaneously.

PS-M4110 Drawer
You can open the array inner drawer while the operating member gains the access to
hot-swap components. Steps to open:
Push the array's front panel and release it quickly to unlatch the array’s inner drawer.
When the drawer is unlatched, a Caution label will be visible.
The front panel is not designed as a handle. It can break if treated roughly. When opening the
array inner drawer, do not pull on the front panel. Grip and pull the drawer by its top, bottom,
or sides.

Push the release button located on the side of the PS-M4110 array, which releases the latch
that secures the array's drawer to its outer housing.
This prevents the array's drawer from opening accidentally during handling when outside of the
M1000e enclosure.
Drive
Identification
Drives are numbered from 0 to 13 from the front of the M4110 to the back.
Type 13 Control Modules
The PS-M4110 uses a Type 13 control module.

Components of Type 13 Control Modules


Key components:

 Ethernet port
 A 10Gb/s iSCSI Ethernet port (Ethernet 0) is used for communication on
one of the two redundant fabrics.
 Management port
 Ethernet port 1 can optionally be set up as a management port.
 Status and Power LEDs
 Indicate status of the control module. ACT LED for activity and PWR LED
for power.
 Serial port
 It enables you to connect a computer directly to the array, without network
access.
 Micro SD card
 A field-replaceable micro SD card containing the PS Series firmware.
 Release button and latch
 It releases the control module from the array for replacement. The release
lever has a switch that detects activation and prompts the array to save data to non-
volatile storage (data residing in cache memory) .

 Controller Module Up-Close


 To remove the controller module, release the latch by pressing the red circle.

When the user configures the blade array to work with fabric B, the user cannot specify to use
B1 or B2 fabric.

The PS-M4110 will automatically chooses either B1 or B2 fabric, so when the user has many
blade arrays inside the chassis configured for fabric B then B1 and B2 must be stacked
otherwise the two blade arrays might not talk to one another.
Switch Configurations
When using a PS-M4110 inside an M1000e enclosure, the IO modules must be
interconnected (stacked or use a link aggregation).

 If fabric B is configured, the B1 and B2 IOMs must be stacked or use a link


aggregation (LAG).
Stacking allows multiple switches management as a single entity while LAG uses
multiple parallel network connections to increase the network throughput.
The redundant fabric IOMs must be connected using interswitch links (stack interfaces or LAGs).
The links must have sufficient bandwidth to handle the iSCSI traffic.

Identify Failed Drive


When installing PS-M4110 drives, observe the following guidelines and restrictions:

 Replace a failed drive as soon as possible to provide the highest availability.


 Install only drives of the same type, speed, and spin rate in an array.
Refer to the following table for more information about PS-M4110 LED status.

Replace Drive
Replacement steps:

 Remove the bad drive by pressing the drive release button.


 Fully open the drive release latch.
 Remove the drive by grasping the drive sides.
 Install the new drive by lifting the drive release latch to its fully open position.
 Hold the drive by its carrier and insert the drive into an empty slot.
 When the drive is fully inserted as far as it can go, rotate the drive release latch
to its fully closed position.
 Push down on the latch until you hear and feel a click.

Replace a Control Module


Replacement steps:

 Ensure replacement is a non-active controller.


 Remove the control module by pressing the control module release button.
 Fully open the control module release latch.
 Remove the control module by grasping it from the sides.
 Install the control module:
o Lift the control module release latch to its fully open position.
o Hold the control module by its carrier and insert it into the slot.
o When the control module is fully inserted, push down on the latch until it
click

Special Handling: Control Module


Important points to note:

 If you are replacing a failed control module, remove the micro SD card from the
failed control module and install it in the replacement control module. This will make
sure that the new control module is running the correct firmware.
 If two control modules are installed in the array, but only one is shown in the GUI
(or CLI), make sure you have allowed enough time (two to five minutes) for the two
control modules to boot and synchronize.
 When synchronization completes, a message is displayed on the serial console
(if connected), and the ACT LED on the secondary module will be orange.

Micro SD Card Replacement Procedure


Replacement steps:

 Remove the control module from the array.


 Remove the SD card from the control module.
o Push the micro SD card in its slot to release the spring mechanism. The
micro SD card will be partially ejected from the housing.
o Gently pull the card straight out of the slot.
 Install the replacement SD card in the control module.
o Align the micro SD card so the arrow on the card points towards the micro
SD card slot in the control module.
o When correctly positioned, the notch on the card corresponds to the slot.
o Firmly press the card into the slot until it clicks into place.
 Insert the control module into the array.

PS-M4110 Management Methods


There are four management methods used to connect to a PS-M4110 array:

 Group Manager via the iDRAC


 Serial Console Ports by physical connection
 Telnet/SSH via CMC
 M1000e CMC/iDRAC

 Connect to the CMC Using Telnet


 Use SSH or Telnet to connect to the IP address of the CMC and log in.

Connect Using the iDRAC port


You can connect to the iDRAC port of the PS-M4110 with a RACADM command:

 Connect server-x (x=slot number)


 Login to the M4110 active controller module

Fabric Configuration
By selecting the M4110 in the system tree, you can view and change certain settings.
As you can see below, you can also change the fabric that the M4110 connects to.
If the member is already installed, you can change the fabric it communicates on.
PS-M4110 Management Network
The M4110 management network path is via the CMC network port.

In-Band Deployment/Management
In-Band Management allows configuration management through the network locally.
The M4110 can be:

 Deployed from CMC Console Interface – Management Workstation.


 Managed from EqualLogic Group Manager – Host Server.
Common Network Terms
Defined as a method of encoding, transporting, and synchronizing data between
multiple devices.

 Fabric is transported within the converged system between server modules and
I/O Modules through the midplane.
 Fabric is carried to the outside world through the physical copper or optical
interfaces on the I/O Modules or PCIe cards.
 Examples of fabrics are GbE, Fibre Channel, and InfiniBand.
One example of GbE network fabrics is the 10GbE Pass-Through Modules for Dell
M1000e Blade Enclosures as shown in the picture below.

Lane
Defined as a single fabric data transport path between I/O end devices.
In modern high-speed serial interfaces, each lane comprises one transmit and one
receive differential pair.

In reality, a single lane has four wires in a cable or traces of copper on a printed circuit
board:

 Transmit positive signal


 Transmit negative signal
 Receive positive signal
 Receive negative signal
Differential pair signaling provides improved noise margin for these high-speed lanes.

Various terminologies are used by Fabric standards when referring to lanes:

Link

Defined as a collection of multiple Fabric lanes used to form a single communication


transport path between I/O end devices. The differentiation has been made here
between lane and link to help prevent confusion over Ethernet’s use of the term link for
both single- and multiple-lane fabric transports.

 Some Fabrics such as Fibre Channel do not define links, as they simply run
multiple lanes as individual transports for increased bandwidth.
 A link as defined here provides synchronization across the multiple lanes, so they
effectively act together as a single transport.
 Examples are 4-lane (x4), 8-lane (x8), and 16-lane (x16) PCIe, or 4-lane
10GBase-KX4.
PCIe, InfiniBand, and Ethernet define Fabric lanes as Link.
Port

Defined as the physical I/O end interface from a device to a link, a port can have single or
multiple lanes of Fabric I/O connected to it.

M-Series I/O Modules


You have learned the different network terminologies. Now, let us go through the in-
depth knowledge of M1000e supported Interconnect types.
Converged Ethernet

 10/40Gb Switch MXL


 10 Gb Plug and Play M I/O Aggregator
 10 Gb Basic M8024-k
 10 Gb Pass-Through Module

 Ethernet 1Gb/10Gb
 1Gb Pass-Through Module
 1/10Gb High-density M6348
 1/10Gb Basic M6220

 Fibre
 Brocade M6505
 Brocade M5424
 8/4 Gb Fibre Channel Pass-Through Module

 Inifiniband
 Mellanox M4001F
 Mellanox M4001T

MXL: 10/40-GbE Blade
MXL 10/40-GbE Blade as a converged network deployment helps improve data center
flexibility with configurable I/O choice.

Two FlexIO Modules


Two QSFP+ Ports
Some of the key features include:

 56-port design
 Two Flex I/O bays enabled choice (Modules can be different)
 PVST+ protocol for easy integration into Cisco environments
 Two FCoE options:
 Native Fibre Channel uplinks with FC FlexIO module (FCoE only on
internal ports to the servers)
 FCoE transit to top of rack switch with IOM acting as a FIP Snooping
Bridge

MXL: 10/40-GbE Blade Configuration


Manual Initial Configuration
When you initially log in to the switch, by default, you are logged in to EXEC mode.
EXEC mode enables you to view settings and enter EXEC Privilege mode, which is
used to configure the device.
The MXL blade uses industry standard CLI, such as:

 range command selects multiple ports


 switchport command selects Layer 2-port configuration
 flowcontrol command selects buffering ports for storage use
 mtu command defines maximum packet size for Jumbo frames; select 9216
packet size for ports
 no shutdown command enables ports

M8024k
M8024k is a fully modular and managed Layer 2/3 Ethernet switch:

One FlexIO Module


Four SFP/SFP+ ports
 24-port design
 Flex I/O fully modular design enables connectivity choices including SFP+ and
10GBASE-T
 Provides connectivity for the latest 10Gb-KR NICs and CNAs (Converged
Network Adapters), including support for Switch-Independent Partitioning
 Supports DCB (protocols PFC and DCBx)
 Converged features (not supported in Simple Switch Mode):
 FCoE Transit Switch using FIP Snooping Bridge
 Stack up to six devices using SFP+ fixed ports or SFP+ module
10 Gb Ethernet Pass-Through
An Ethernet Pass-Through module exposes each Ethernet port to external network
switches.

 16 ports correspond to 16 server blades.


 16 external 10 GbE SFP+ ports.
 Pass-Through Module connects to top-of-rack FCoE switches and Converged
Network Adapters (CNAs) in individual blades.
 There is a transparent connection between blade servers and external LAN.
10Gb Ethernet Pass-Through Configuration
Important: Pass-through module can help connect different blades in the M1000e to
different switches or isolated networks. Use a basic switch if switching is to be
contained within the M1000e enclosure.
1/10 Gb Ethernet Pass-Through Module
The 1GbE Pass-Through Module provides transparent connection between LAN and
server blades. Key features:

Cables
CAT 5

 16 ports correspond to 16 server blades.


 Supports 10/100/1000Mb connections with all 1Gb Broadcom adapters.
 Ethernet media speed is configured through the blade LOM firmware or by the
operating system.
Option Modules: M6348
M6348 is a managed Layer 2/3 Gigabit Ethernet switch. Key features:

 48-port switch.
 Stackable with rack-mount PowerConnect 7000 Series.
 Supports Dell Simple Switch Mode.
M6348 Configuration
Important: M6348 works with all 1Gb Mezzanine cards and LOMs. For optimized use (full
internal-port utilization), pair with: Quad-port GbE Mezzanine cards or Quad-port Fabric A
adapters.
Option Modules: M6220
M6220 is a key component of the Flex I/O architecture of the M1000e. Key features:

 24-port Ethernet switch.


 Optional 10 Gb uplinks and resilient stacking.
 Supports Dell Simple Switch Mode.
M6220 Configuration

Fibre Channel
The Fibre Channel (FC) networks are used to:

 Attach host servers to storage in small to medium businesses and in large


enterprises.
 Increase the number of servers that can access a single storage array.
 Increase the amount of storage each host can access and the storage I/O
operations per second performance of each server.
The Fibre Channel protocol is used to transport small computer systems interface
(SCSI), commands, and associated data to access block-level storage. Storage Area
Networks (SANs) are built using specialized Fibre Channel switches. As these networks
grow into multiple switch networks, they form Fibre Channel Fabrics. These Fabrics use
mesh or partial mesh topologies.
A SAN:

 Transports SCSI commands and disk data between servers and storage arrays
on the network.
 Can be formed on either a Fibre Channel or standard Ethernet TCP/IP network
infrastructure.

Fibre Channel in M1000e


Fibre Channels are used in Storage Area Network (SAN) applications for higher speed
rate, and lossless delivery of raw block data.
Fibre Channel switches from left to right:

 Brocade M6505
 Brocade M5424
 8/4 Gbps FC Pass-Through

Brocade M6505 (16 Gb Switch)


The Brocade M6505 adds 16-Gbps Fibre Channel connectivity across servers, storage,
and SAN as needed:

 24 Fibre Channel ports


 Zero footprint, hot-pluggable design with no additional fans or power supplies
 Auto-sensing and speed-matching connections with 16/8/4 Gbps to Fibre
Channel devices
 Complete redundancy, up to four switches per chassis
 Dynamic Ports on Demand (PoD) and “pay-as-you-grow” port upgrades for 12-
port configurations
Brocade M5424 Configuration
Configurable as Brocade full fabric switch or Access Gateway Mode for multi-vendor
inter-operability.
Comparison: M5424 and M6505
8/4 Gbps FC Pass-Through Module
8/4 Gbps is a Fibre Channel Pass-Through Module for the M1000e.

 8 Gbps, 4 Gbps, or 2 Gbps connections.


 16 ports correspond to 16 server blades.
 Module is unmanaged – all management occurs through firmware or exterior
switches.

 InfiniBand
 InfiniBand is a protocol and switched network infrastructure that is a high-speed
replacement for the internal PCI bus that is used in servers. It is configured as a
direct memory access interconnection between two or more servers. The
InfiniBand network becomes the high-performance server clustering interconnect
for clustered systems running high-speed computer applications.
The InfiniBand, Fibre Channel over IP(FCIP) and Fibre Channel over Ethernet(FCoE) are other
examples of various protocols and network infrastructures available to augment and enhance
existing end-to-end connectivity

Mellanox Blades
Mellanox blades in an M1000e use the InfiniBand protocol:

 For high-performance computing (HPC) and low latency applications.


 Available in redundant switch configuration.
 Full non-blocking throughput.

CMC Indicators
Troubleshooting Non-Responsive Chassis
Management Controllers
If you cannot log in to the CMC using any of the interfaces (the Web interface, Telnet,
SSH, remote RACADM, or serial), you can verify functionality by observing the CMC
LEDs. Facing the front of CMC as it is installed in the chassis, there are two LEDs on
the left side of the card.
The top green LED (item 10 in illustration) indicates power. If it is not on:

 Verify that you have AC present to at least one power supply.


 Verify that the CMC card is seated properly. Release or pull the ejector handle,
remove the CMC, reinstall the CMC ensuring that the board is inserted all the way and
the latch closes correctly.
 NOTE: A normal CMC boot or reset takes over a minute to fully boot into its operating
system and be available for login. The blue LED is enabled on the active CMC. In a redundant,
two-CMC configuration, only the top green LED is enabled on the standby CMC.

 The bottom LED (items 8 and 9 in the figure below) is multi-colored. When CMC
is active and running, and there are no problems, the bottom LED is blue. If it is amber,
a fault is detected. The fault may be caused by any of the following three events:
 A core failure. In this case, the CMC board must be replaced.
 A self-test failure. In this case, the CMC board must be replaced.
 An image corruption. In this case, upload the CMC firmware image to
recover the CMC.

Power Supply indicators


Fan Module Indicators
iDRAC Troubleshooting and Video Capture
The iDRAC9 contains a troubleshooting section. From here, you can view the last three
power-on procedures in video format. You can also run a selection of remote diagnostic
commands, view the last crash screen, or run a SupportAssist Collection.

 The iDRAC Video Capture feature automatically records the boot process and
stores the last three boot process recordings.
 Video Capture
 The iDRAC Video Capture feature automatically records the boot process
and stores the last three boot process recordings.
 A boot cycle video logs the sequence of events for a boot cycle.
 A crash video logs the sequence of events leading to the failure.
 Boot Capture Video Settings
 Disable - Boot capture is disabled.
 Capture Until Buffer Full - Boot sequence is captured until the buffer size
has reached.
 Capture Until End of POST - Boot sequence is captured until end of
POST.

 SupportAssist Collection
 This page accesses SupportAssist and creates SupportAssist collections.
 This integration enables you to use other SupportAssist features. iDRAC
provides application interfaces for gathering platform information that enables
support services to resolve platform and system problems.
 iDRAC helps you to generate a SupportAssist collection of the server. You can
and then export the collection to a location on the management station (local) or
to a shared network location.
 The collection is generated in the standard .ZIP format. You can send this
collection to technical support for troubleshooting or inventory collection.
iDRAC Diagnostics
This page is used to diagnose issues that are related to the iDRAC hardware using network
diagnostic tools. Here are two sections of iDRAC Diagnostic.

 Reset iDRAC
 Resets the iDRAC.
 A normal reboot operation is performed on the iDRAC.
 After reboot, refresh the browser to reconnect and log in to iDRAC.
 Reset iDRAC to Default Settings
 This action resets the iDRAC to the factory defaults.
 Choose any of the following options:
o Preserve user and network settings.
o Discard all settings and reset users to shipping value.
o Discard all settings and reset user name and password.

iDRAC Logs
 Lifecycle Log
 This page enables you to view and export the Lifecycle Controller log entries.
 By default, the latest 100-log entries are displayed.
 You can filter the log entries based on category, severity, keyword, or a date
range.
 System event log
 This page enables you to view, clear, or save the log events that occur on the
blade server.
 You can configure the iDRAC to send emails, SNMP traps when specified events
occur.

Easy Restore
The Easy Restore function automatically backs up vital information to the restore Serial
Peripheral Interface (rSPI) card.

If the system board is ever replaced, a screen is presented giving you the option of
automatically restoring certain types of information:

 Automates the system board replacement process


 Stored in flash memory in the restore Serial Peripheral Interface (rSPI) card
 Restores:
 Service Tag
 iDRAC license
 BIOS, iDRAC, and LOM/NDC configuration

Server Profile
The Server Provide backup on a vFlash SD card contains the server component
configuration and firmware that is installed on various components on the server. The
backup image file does not contain any operating system or hard-disk drive data. It is
used when the system board is replaced to restore all the configuration and firmware
revisions.
The Server Profile backs up the Lifecycle Controller configuration information, the BIOS,
and all the firmware code on to a local drive or network share.
When the new system board is installed, you access the new Lifecycle Controller and
import the Server Profile.
The backed-up configuration information is now restored to the new Lifecycle Controller,
which in turn pushes all of that information out to the hardware parts. That saves you
time and effort from manually reconfiguring devices like the iDRAC and the BIOS.
The Lifecycle Controller also pushes out the firmware and BIOS code. However, you
must back up the Server Profile before the system board fails.
Use the Part Replacement feature to automatically update a new part to the firmware
version or the configuration of the replaced part, or both. The update occurs
automatically when you reboot your system after replacing the part.
NOTE: Part Replacement does not support RAID operations such as resetting
configuration or recreating Virtual Disks.

Blade Server Diagnostics Utility


Dell Embedded Diagnostics are also known as the enhanced Preboot System
Assessment (ePSA) diagnostics. Having them embedded as part of the Lifecycle
Controller subsystem enables quick access without booting to the operating system.
As a foundation for determining functionality of a freshly installed device at a hardware
level, there are embedded diagnostic hardware tests which enable diagnostic tests to
be quickly run.

 Can be accessed through either:


 Lifecycle Controller (<F10> during POST)
 Boot Manager (<F11> during POST)

Minimum Hardware Configuration


The following is the minimum hardware configuration that is required to power on the M1000e
chassis:

 One CMC
 Three Power Supply Units
 Nine fan modules

 Minimum CMC Firmware Versions for


Blade Servers
 The following table lists the minimum CMC firmware version that is required to
enable the listed Blade Servers:

You might also like