The IBM 9119 is a Power System E880 with POWER8 system node that
uses 8-core symmetric multiprocessing (SMP) processor chips with 512
KB of L2 and 8 MB of L3 cache per core, DDR3 CDIMM memory, dual memory
controllers, and an industry-standard Gen3 PCIe I/O bus designed to
use 32 lanes organized in two sets of x16. The peak memory and I/O
bandwidths per system node have increased over 300% compared to
POWER7+ servers. The two primary system building blocks are one
system control unit and one or more system nodes. Additional I/O
support is provided with a 19-inch PCIe Gen3 I/O expansion drawer and
an EXP24S SFF Gen2 expansion drawer. The processors, memory, and base
I/O are packaged within the system nodes. The system nodes are rack
based.

Model abstract 9119-MHE

The IBM 9119 Power System E880 Model MHE features: System with
processor, memory, and base I/O, Up to sixty-four 4.356 GHz POWER8
processor cores, Up to 8 TB of 1600 MHz DDR3 CDIMM memory, eight PCIe
G3 x16 I/O expansion slots per system node enclosure, a maximum of 16
per system; system control unit, providing redundant system master
clock and redundant system master Flexible Service Processor (FSP) and
support for the Op Panel, the system VPD, and the base DVD; 19-inch
PCIe Gen3 4U I/O expansion drawer and PCIe FanOut modules, supporting
a maximum of 48 slots; PCIe Gen1, Gen2 and Gen3 adapter cards
supported in both the system node and I/O expansion drawer; EXP24S SFF
Gen2-bay Drawer with twenty-four 2.5-inch form-factor SAS bays;
Dynamic logical partition (LPAR) support for adjusting workload
placement of processor and memory resources; Active Memory Expansion
that is optimized onto the processor chip; No-charge elastic processor
and memory days with initial system order; 90-day temporary elastic
CoD processor and memory enablement features; and Power Enterprise
Pools, supporting unsurpassed enterprise flexibility for workload
balancing and system maintenance.
Back to top

The IBM Power System E880 server is designed as an industry-leading
enterprise class system with outstanding price/performance,
mainframe-inspired reliability and availability features, flexible
capacity
upgrades, and innovative virtualization technologies. The new Power
E880 model MHE features the following:

System with processor, memory, and base I/O

Up to one hundred twenty-eight 4.35 GHz POWER8 processor cores (up
to 64 cores in 2014)

Up to 16 TB of 1600 MHz DDR3 CDIMM memory (up to 8 TB in 2014)

Eight PCIe Gen3 x16 slots per system node enclosure, a maximum of
32 in a 4-node enclosure (up to 16 per 2-node system in 2014)

System control unit, providing redundant system master clock and
redundant system master Flexible Service Processor (FSP) and support
for the Op Panel, the system VPD, and a DVD

A new 48-core system node option (#EPBD) is added to the existing
32-core system node for the Power E880 (9119-MHE) server. One, two,
three, or four of the 4.02 GHz 48-core system nodes provides a
48-core, 96-core, 144-core, or 192-core server. All the system nodes
in the server must be the same gigahertz and feature number.

Each 5U 48-core system node has 32 CDIMM memory slots providing up
to 4 TB of memory capacity and 8 PCIe Gen3 x16 slots providing up to
256 GB of I/O bandwidth. Memory configuration guidelines and
supported PCIe configuration options and maximums are identical to the
Power E880 32-core system node guidelines and options. Connection
from the system node to the system control unit uses special cables
ordered using features ECCA (first node), ECCB (second node), ECCC
(third node), and ECCD (fourth node).

The 48-core Power E880 system node requires a minimum of 8 cores
which must be permanently activated (static activations), just like
the 32-core system node minimum. The rest of the cores can use
static, mobile, mobile-enabled, Power Integrated Facility for Linux
(IFL) or remain inactive or use temporary activations. Processor core
activation features are:

1-core static activation (#EPBM)

1-core mobile activation (#EP2T)

1-core mobile-enabled activation (#EPBR)

4-core Power IFL package (#ELJG)

Temporary Capacity on Demand features are available for the 48-core
system node server using:

90-day enablement (#EP9T)

1-day elastic billing feature using AIX/Linux (#EPJQ)

1-day elastic billing feature using IBM i (#EPQR)

100-day elastic billing feature using AIX/Linux (#EPJS)

100-day elastic billing feature using IBM i (#EPJT)

No-charge elastic processor and memory days are available with the
initial system order or with a same-serial-number model upgrade. You
can receive 15 processor days and 240 GB days for every active and
inactive processor core initially. Order this by using the
appropriate quantity of features EPJ5 and EMJ9.

Like the 32-core Power E880 server, same-serial-number upgrades
from the Power 780 server are available to the 48-core Power E880
server. Additional feature conversions to the 48-core processor and
activation features are announced for this purpose.

Firmware level 8.3, or later, is a prerequisite for the 48-core
system node.

Power 770 upgrade to Power E880

Previously, Power 770 (9117-MMD) upgrades to a Power E870
(9119-MME) server were supported. Now Power 770 (9117-MMD) upgrades
to a Power E880 (9119-MHE) server are also supported. This gives
Power 770 "D model" clients who choose the E880 server additional
growth options in processor cores, gigahertz, maximum memory, and
maximum I/O. Multiple feature conversions are introduced to provide
the underlying ordering capability for this upgrade.

Power E870/E880 enhancements

The PCIe I/O slot maximum is doubled for both the Power E870 and
E880 servers. Previously, a maximum of two PCIe Gen3 I/O Drawers per
system node was announced. Now up to four drawers per node can be
attached. Using two 6-slot Fan-out Modules per drawer provides a
maximum of 48 PCIe slots per system node. With two system nodes, up
to 96 PCIe Gen3 slots (8 I/O drawers) are supported. With a 4-node
Power E880 server up to 192 PCIe Gen3 slots (16 I/O drawers) are
supported. Firmware 8.3, or later, is required to use the doubling of
I/O drawers.

Additional PCIe I/O drawer configuration flexibility is provided to
the Power E870 and E880 servers. Previously, only zero or two PCIe
I/O drawers could be attached to a system node. Now zero, one, two,
three, or four PCIe I/O drawers can be attached per system node. Plus
a "half" drawer consisting of just one PCIe fan-out module in the I/O
drawer is also supported, allowing a lower-cost configuration if fewer
PCIe slots are required. Thus a system node supports the following
half drawer options: one half drawer, two half drawers, three half
drawers, or four half drawers. Because there is a maximum of four
EMX0 drawers per node, a single system node cannot have more than four
half drawers. A server with more system nodes can support more half
drawers up to four per node. A system can also mix half drawers and
full PCIe Gen3 I/O drawers. The maximum of four PCIe Gen3 drawers per
system node applies whether a full or half PCIe drawer.

Drawers can be added to the server at a later time, but system
downtime must be scheduled for adding a PCIe3 Optical Cable Adapter or
a PCIe Gen3 I/O drawer (EMX0) or fan-out module (#EMXF).

An additional 20-meter AOC cable (#ECC9) is announced for the Power
E870 and E880 servers, providing additional cabling and distance
flexibility.

The number of PCIe adapters supported in the Power E870/E880 system
node is more than doubled. This provides additional configuration
flexibility. The following additional adapters, which had been
previously announced on other servers, are now supported in the system
node:

PCIe LP POWER GXT145 Graphics Accelerator (#5269)

PCIe LP 4Gb 2-Port Fibre Channel Adapter (#5276)

PCIe2 LP 2-Port 4X IB QDR Adapter 40Gb (#5283)

PCIe2 LP 2-port 10GbE SR Adapter (#5284)

PCIe2 LP 3D Graphics Adapter x1 (#EC41)

PCIe2 LP 4-Port (10GbE+1GbE) SR+RJ45 Adapter (#EN0T)

PCIe2 LP 4-port (10GbE+1GbE) Copper SFP+RJ45 Adapter (#EN0V)

PCIe LP 2-Port Async EIA-232 Adapter (#EN28)

PCIe LP 10Gb FCoE 2-port Adapter (#5270)

PCIe LP 4-Port 10/100/1000 Base-TX Ethernet Adapter (#5271)

PCIe LP 2-Port 1GbE SX Adapter (#5274)

PCIe LP 4-Port Async EIA-232 Adapter (#5277)

PCIe LP 2-Port 1GbE TX Adapter (#5281)

PCIe2 LP 2-Port 10GbE RoCE SFP+ Adapter (#EC27)

PCIe3 LP 4-port (10Gb FCoE & 1GbE) LR&RJ45 Adapter (#EN0N)

The first eight adapters above were listed in an October 2014
statement of direction (SOD). This SOD listed 10 low-profile
adapters; however, IBM no longer plans to announce support of the
following two low-profile Ethernet adapters in the Power E870/E880
system node:

PCIe LP 10GbE SR 1-port Adapter (#5275)

PCIe2 LP 4-Port 10GbE&1GbE SR&RJ45 Adapter (#5280)

Thus the combination of the above eight newly supported adapters
plus the two adapters that will not be supported reflect the modified
completion of the SOD. Note that these two low-profile adapters that
will not be supported have equivalent full-high adapters which are
supported in the PCIe Gen3 I/O Expansion Drawer.

In addition to the above adapters being supported, two new PCIe3
adapters are supported in the system node. These are a 4-port 10Gb
Ethernet (#EN16/#EN18) adapter and a 2-port 10Gb Ethernet
(#EC37/#EC2M) adapter.

Four additional PCIe adapters that were previously supported on
other servers are now supported in the PCIe Gen3 I/O drawer:

10Gb FCoE PCIe Dual Port Adapter (#5708)

PCIe2 4-Port (10GbE and 1GbE) SR and RJ45 Adapter (#5744)

PCIe 380 MB Cache Dual Port 3Gb SAS RAID Adapter (#5805)

PCIe3 4-port (10Gb FCoE and 1GbE) LR and RJ45 Adapter (#EN0M)

In addition to the above four adapters, two new PCIe3 adapters are
supported in the PCIe Gen3 I/O Drawer. These are a 4-port 10Gb
Ethernet adapter (#EN15/#EN17), and a 2-port 10Gb Ethernet adapter
(#EC2N/#EC38).

With additional PCIe Gen3 I/O drawers, the maximum number of PCIe
adapters is increased. Associated with the increase in the number of
SAS adapters in an increase to the EXP24S HDD/SSD drawer (#5887). The
E870 now supports up to 3,072 drives with a new maximum of 128 EXP24S
drawers. The E880 now supports up to 4,032 drives with a new maximum
of 168 EXP24S drawers. The maximum of 16 EXP24S drawers per PCIe Gen3
I/O drawer due to cabling considerations remains unchanged.

PCIe Gen3 adapters

4-port 10Gb Ethernet Adapter

The PCIe3 4-port 10GbE Adapter doubles the number of 10GbE ports
per adapter, saving PCIe slots on POWER8 servers. The adapter is
supported in full-high slots such as the PCIe Gen3 I/O drawer and the
4U scale-out system units and in the Power E870/E880 system node
slots. It does not fit in the 2U scale-out server low-profile slots.
In addition to Ethernet NIC support, it also provides SR-IOV NIC
support, which can allow further virtualization and slot savings.

The adapter has four SFP+ ports. All four ports use SR optical
fiber cabling or all four use copper twinax cabling. Optical
transceivers are included with the SR adapter feature. Copper twinax
transceivers are included on the separately ordered copper cables.

IBM i, AIX, and Linux support this adapter, both with and without
VIOS. IBM i supports the use of this adapter for IBM i LAN console.
NIM and Linux Network Install function is supported. See the feature
code description for prerequisite software levels.

New 2-port 10Gb NIC/RoCE Adapter

The PCIe3 2-port 10GbE NIC/RoCE Adapter is a refresh of the current
10Gb NIC/RoCE Adapter, helping to ensure adapter availability for
POWER8 servers. The adapter can be configured for either NIC or for
RoCE capability. NIM and Linux Install is supported, an enhancement
compared to the existing PCIe2 NIC/RoCE adapters
(#EC27/#EC28/#EC29/#EC30).

RoCE can provide a performance and efficiency boost for the
applications using the interface. The adapter is available either for
SR optical fiber cabling or for copper twinax cabling. Optical
transceivers are included with the SR adapter feature. Copper twinax
transceivers are included on the separately ordered copper cables.

Note: -1 SR-IOV announced February 2015 for Power E870/E880 system
node. Now available in other POWER8 servers.

Note: -2 Adapter is only available in Power E870/E880 system node, not
2U server.

These adapters each have four ports, and all four ports are enabled
with SR-IOV function. The entire adapter (all four ports) is
configured for SR-IOV or none of the ports is.

The client can chose to configure as SR-IOV and assign 100% of a
port to one partition. The FCoE capability of the adapter is not
supported when using SR-IOV. SR-IOV can provide simple virtualization
without VIOS with greater server efficiency as more of the
virtualization work is done in the hardware and less in the software.
SR-IOV can also provide bandwidth quality of service (QoS) controls to
help ensure that client-specified partitions have a minimum level of
Ethernet port bandwidth and thus improve the ability to share Ethernet
ports more effectively.

As a QoS example, assume there are three partitions: Partition A,
B, and C. If Partition A is assigned a minimum of 20% of the bandwidth
of a port, Partitions B and C cannot use more than 80% of the
bandwidth unless Partition A is using less than 20%. Partition A can
use more than 20% if bandwidth is available. This applies to outbound
NIC traffic from the port. It does not apply to inbound NIC traffic
into the port, but remember that these ports run in full duplex and
thus QoS for outgoing traffic remains in effect even with high levels
of inbound NIC traffic.

SR-IOV is considered simple virtualization and does not include
higher-level virtualization functions provided by VIOS such as Live
Partition Mobility. SRIOV can optionally be combined with VIOS to
leverage VIOS's higher level of functionality, but continues to
provide bandwidth QoS. The portion of the Ethernet port's bandwidth
controlled by VIOS would use software virtualization.

The expanded POWER8 SR-IOV NIC capability with a planned
availability of June 2015 requires:

AIX 6.1 TL9 SP5 and APAR IV68443, or later

AIX 7.1 TL3 SP5 and APAR IV68444, or later

IBM i 7.1 TR10, or later

IBM i 7.2 TR2, or later

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise Server 11 SP3, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu 15.04, or later

VIOS 2.2.3.51, or later

Firmware level 8.3, or later

This expanded capability is on the Power scale-out servers, is on
the PCIe Gen3 I/O drawer, and with the new 4-port 10Gb Ethernet
adapter on the enterprise system node.

IBM Manufacturing is not announcing the capability to preconfigure
SR-IOV-capable I/O using SR-IOV. Such hardware will be shipped from
IBM when ordered with a server in the same way adapters have been
shipped previously. The client will need to configure the SR-IOV
capability.

The following POWER8 PCIe slots are SR-IOV capable:

All Power E870/E880 system node slots.

Slots C1 and C4 of the 6-slot Fan-out Module in a PCIe Gen3 I/O
drawer.

Slots C2, C3, C4, C5, C6, C7, C10, and C12 of a S824 or S824L
server (2-socket, 4U) with both sockets populated. If only one socket
is populated, then C6, C7, C10, and C12.

Slots C2, C3, C5, C6, C7, C10, and C12 of a S822 or S822L server
(2-socket, 2U) with both sockets populated. If only one socket is
populated, then C6, C7, C10, and C12.

An HMC is required to configure SR-IOV.

eConfig recognizes VIOS is not a prerequisite for Power E870 and
E880 system lower-profile feature EN0J and EN0L adapters because of
the SR-IOV capability. But eConfig will not be updated until June 2,
2015, with the same insight for the full-high feature EN0H/EN0K/EN0M
adapters for IBM i configurations. These full-high adapters can be
ordered if VIOS is present on the configuration prior to eConfig's
planned enhancement. eConfig recognizes that features EN15, EN16,
EN17, and EN18 do not require VIOS for IBM i configurations.

POWER8 processor single chip modules (SCM) are provided in each
system node. Processors are interconnected by two sets of system
buses. Each SCM contains two memory controllers per processor module.
Four 4.35 GHz 8-core SCMs are used in each system node, providing 32
cores (#EPBB). As few as eight cores in the system can be activated
or up to 100% of the cores in the system can be activated.
Incrementing one core at a time is available through built-in capacity
on demand (CoD) functions to the full capacity of the system.

The Power E880 can have up to four system nodes per system. With
32-core nodes, the maximum is a 128-core system. In 2014, up to two
system nodes per server are supported or a maximum of 64 cores.

The system control unit provides redundant system master clocks and
redundant system master service processors (FSPs). Additionally, it
contains the Operator Panel, the System VPD, and the Base DVD.

The 1600 MHz memory CDIMMs are available as 64 GB (#EM8J), 128 GB
(#EM8K), 256 GB (#EM8L), and 512 GB (#EM8M) memory features. Each
memory feature provides four CDIMMs. CDIMMS are custom DIMMs that
enhance memory performance and memory reliability. Each system node
has 32 CDIMM slots that support a maximum of eight memory features.
Using 512 GB CDIMM features yields a maximum of 4 TB per node. A
two-node system has a maximum of 16 memory features and 8 TB. A
four-node
system has a maximum of 16 TB Capacity. Memory activations of 50% of
the installed capacity are required.

The 19-inch PCIe Gen3 4U I/O Expansion Drawer (#EMX0) provides
slots to hold PCIe adapters that cannot be placed into a system node.
In 2014, either zero or two PCIe I/O expansion drawers can be attached
per system node. Two PCIe I/O drawers provide 24 PCIe Gen3 adapter
slots. Thus a two-node system has a maximum of 48 PCIe Gen3 slots in
I/O drawers plus PCIe slots in the system node.

As part of the IBM commitment to delivering the most flexible and
resilient Power high-end systems, all initial shipments of Power E880
systems come with a specific number of no-charge elastic processor and
memory days. The number of days depends on the configuration of the
system. For every new Power E880, 15 elastic CoD processor days for
every processor core (inactive or active) installed on the system and
240 GB of elastic CoD memory days for every processor core (inactive
or active) are included. These elastic CoD processor and memory days
must be used in accordance with the temporary capacity on demand terms
and conditions. Also available are 90-day temporary elastic CoD
processor and memory enablement features. These features enable a
system to temporarily activate all inactive processor and memory CoD
resources for a maximum of 90 days before you must order another
temporary elastic enablement feature number.

With Power Enterprise Pools, IBM continues to enhance the ability
to freely move processor and memory activations from one system to
another system in the same pool, without the need for IBM involvement.
This capability now allows the movement of resources not only between
like systems, but also between generations of Power Systems, and thus
delivers unsurpassed flexibility for workload balancing and system
maintenance. Power Enterprise Pools deliver the support to meet
clients' business goals when it comes to the following:

Improving the flexibility, load balancing, and disaster recovery
planning and operations of your Power Systems infrastructure

Enhanced reliability, availability, and serviceability (RAS) to
handle the requirements to accommodate a global economy

Power Enterprise Pools mobile activations are available for use on
the Power 770, 780, and 795 systems and now on the Power E870 and E880
systems. Two types of Power Enterprise Pools can be created: a
mid-range pool that can consist of Power 770s (9117-MMD) and Power E870s
(9119-MME) and High End Pools that consist of Power 780s (9179-MHD),
Power 795s (9119-FHB) and Power E880s (9119-MHE).

Summary of features

The following features are available on the Power E880:

One or two 5U 19-inch rack-mount system nodes in 2014 and one to
four 5U system nodes in 2015

One 2U 19-inch rack-mount system control unit drawer

Only 12U for a system with two system nodes or 22U for a four-node
system

One processor feature per system node:

4.35 GHz, (4 x 0/8W) 32-core POWER8 processor (#EPBB)

Static or mobile processor activation features available on a per
core basis

Eight PCIe Gen3 x16 I/O low profile expansion slots per system
node (maximum 16 with 2-node system or 32 in a 4-node system)

One slim-line, SATA media bay per system control unit enclosure
(DVD drive defaulted on order, option to de-select)

Redundant hot-swap ac power supplies in each system node drawer

Two Hardware Management Console (HMC) ports per service processor
(FSP) in system control unit enclosure service processor (FSP)
(maximum of one #5550 and one #5557)

Dynamic logical partition (LPAR) support and processor and memory
CUoD

PowerVM Virtualization built:

Micro-Partitioning

Dynamic logical partitioning

Shared processor pools

Shared storage pools

Live Partition Mobility

Active Memory Sharing

Active Memory Deduplication

NPIV support

PowerVP Performance Monitor

Optional PowerHA for AIX, IBM i, and Linux

Optional PCIe Gen3 I/O Expansion Drawer with PCIe Gen3 slots:

Zero to four drawers per system node drawer (#EMX0)

Zero or two or four PCIe Gen3 drawers per 2-node system

Each Gen3 I/O drawer holds two 6-slot PCIe3 Fan-out Modules
(#EMXF)

Each Gen3 I/O drawer attaches to the system node through two PCIe3
Optical Cable Adapters (#EJ07)

Processor cores and memory

Each system must have a minimum of eight active processor cores.
Each processor feature (#EPBB) will deliver a set of four identical
single chip modules (SCMs). All processor features in the system must
be identical.

Cable features are required to connect system node drawers to the
system control unit and to other system nodes.

For a single system node configuration, feature ECCA is required.

For a dual system node configuration, features ECCA and ECCB are
required.

Each system node drawer has 32 memory CDIMM slots holding up to
eight DDR3 memory features.

Each system node drawer must have a minimum of four memory
features or 16 DDR3 CDIMMs. Select from features EM8J (64 GB), EM8K
(128 GB), EM8L (256 GB), or EM8M (512 GB) (four CDIMMs per feature).

The minimum activations ordered with all initial orders of memory
features EM8J, EM8K, and EM8L must be 50% of their installed capacity.

The minimum activations ordered with MES orders of memory features
EM8J, EM8K, EM8L, and EM8M will depend on the total installed capacity
of features EM8J, EM8K, EM8L, and EM8M. This enables you to purchase
newly ordered memory with less than 50% activations when the currently
installed capacity exceeds 50% of the existing features EM8J, EM8K,
EM8L, and EM8M capacity.

The minimum activations installed for all memory must be 50% of
their installed capacity.

DDR3 memory features EM8J, EM8K, and EM8L can be mixed on the same
POWER8 system node drawer. If placing two memory features on the same
SCM they must be identical.

It is recommended that memory be installed evenly across all
system node drawers and all SCMs in the system. Balancing memory
across the installed system planar cards allows memory access in a
consistent manner and typically results in better performance for your
configuration.

Though maximum memory bandwidth is achieved by filling up all the
memory slots, plans for future memory additions should be taken into
account when deciding which memory feature size to use at the time of
initial system order.

These PCIe slots can be used for either low-profile PCIe adapters
or for attaching a PCIe I/O drawer.

A new form factor blind swap cassette (BSC) is used to house the
low-profile adapters that go into these slots. The server is shipped
with a full set of BSC, even if the BSC is empty. A feature code to
order additional low-profile BSC is not required or announced.

If additional Gen3 PCIe slots beyond the system node slots are
required, a system node x16 slot is used to attach a six-slot
expansion module in the I/O drawer. An I/O drawer holds two expansion
modules that are attached to any two x16 PCIe slots in the same system
node or in different system nodes.

PCIe Gen1, Gen2, and Gen3 adapter cards are supported in these
Gen3 slots. The set of PCIe adapters that are supported is found in
the Sales Manual, identified by feature code number.

Concurrent repair and add/removal of PCIe adapter cards is done by
HMC guided menus or by operating system support utilities.

The system nodes sense which IBM PCIe adapters are installed in
their PCIe slots; and if an adapter requires higher levels of cooling,
they automatically speed up the fans to increase airflow across the
PCIe adapters.

Each system node supports up to four CAPI adapters, which can be
located in slots C2, C4, C6, or C8.

PCIe Gen1, Gen2, and Gen3 full-high adapter cards are supported.
The set of full-high PCIe adapters that are supported is found in the
Sales Manual, identified by feature code number. See the PCI Adapter
Placement manual for the 9119-MHE or 9119-MME for details and rules
associated with specific adapters supported and their supported
placement in x8 or x16 slots.

A PCIe X16 to Optical CXP converter adapter (#EJ07) and 2.0 M
(#ECC6) or 10.0 M (#ECC8) CXP 16X Active Optical cables (AOC) connect
the system node to a PCIe FanOut module in the I/O expansion drawer.
One feature ECC6 or one ECC8 ships two AOC cables from IBM.

The two AOC cables connect to two CXP ports on the fan-out module
and to two CXP ports on the EJ07 adapter. The top port of the fan-out
module must be cabled to the top port of the EJ07 port. Likewise, the
bottom two ports must be cabled together.

It is recommended but not required that one I/O drawer be attached
to two different system nodes in the same server (one drawer module
attached to one system node and the other drawer module attached to a
different system node). This can help provide cabling for higher
availability configurations.

It is generally recommended that any attached PCIe Gen3 I/O Expansion
Drawer be located in the same rack as the POWER8 server for ease of
service, but they can be installed in separate racks if the
application or other rack content requires it. If attaching a large
number of cables such as SAS cables or CAT5/CAT6 Ethernet cables to a
PCIe Gen3 I/O drawer, then it is generally better to place that EMX0
in a separate rack for better cable management. Limitation: When
this cable is ordered with a system in a rack specifying IBM Plant
integration, IBM Manufacturing will ship SAS cables longer than 3
meters in a separate box and not attempt to place the cable in the
rack. This is because the longer SAS cable is probably used to attach
to an EXP24S drawer in a different rack.

Concurrent repair and add/removal of PCIe adapter cards is done by
HMC guided menus or by operating system support utilities.

A blind swap cassette (BSC) is used to house the full-high
adapters that go into these slots. The BSC is the same BSC as used
with the previous generation server's 12X attached I/O drawers (#5802,
#5803, #5877, #5873). The drawer is shipped with a full set of BSC,
even if the BSC is empty. A feature code to order additional
full-high BSC is not required or announced.

There is a maximum of 16 EXP24s Drawers per PCIe Gen3 Drawer (#EMX0)
to enable SAS cables to be properly handled by the EMX0 cable
management bracket.

EXP24S Disk/SSD Drawer

The EXP24S SFF Gen2-bay Drawer (#5887) is an expansion drawer with
twenty-four 2.5-inch form-factor SAS bays. Slot filler panels are
included for empty bays when initially shipped. The EXP24S supports
up to 24 hot-swap SFF-2 SAS hard disk drives (HDDs) or solid-state
drives (SSDs). It uses only 2 EIA of space in a 19-inch rack. The
EXP24S includes redundant ac power supplies and uses two power cords.

With AIX, Linux, and VIOS, you can order the EXP24S with four sets
of six bays, two sets of 12 bays, or one set of 24 bays (mode 4, 2, or
1). With IBM i, you can order the EXP24S as one set of 24 bays (mode
1). Mode setting is done by IBM Manufacturing and there is no option
provided to change the mode after it is shipped from IBM.

The EXP24S SAS ports are attached to a SAS PCIe adapter or pair of
adapters using SAS YO or X cables.

To maximize configuration flexibility and space utilization, the
system node does not have integrated SAS bays or integrated SAS
controllers. PCIe adapters and the EXP24S can be used to provide
direct access storage.

To further reduce possible single points of failure, EXP24S
configuration rules consistent with previous Power Systems are used.
IBM i configurations require the drives to be protected (RAID or
mirroring). Protecting the drives is highly recommended, but not
required for other operating systems. All Power operating system
environments that are using SAS adapters with write cache require the
cache to be protected by using pairs of adapters.

It is recommended for SAS cabling ease that the EXP24S drawer be
located in the same rack in which the PCIe adapter is located. Note,
however, it is often a good availability practice to split a SAS
adapter pair across two PCIe drawers/nodes for availability and that
may make the SAS cabling ease recommendation difficult or impossible
to implement.

HDDs and SSDs that were previously located in POWER7 system units
or in feature 5802 or 5803 12X-attached I/O drawers (SFF-1 bays) can
be "re-trayed" and placed in EXP24S drawers. See feature conversions
previously announced on the POWER7 servers. Ordering a conversion
ships an SFF-2 tray or carriage onto which the client can place their
existing drive after removing it from the existing SFF-1
tray/carriage. The order also changes the feature code so that IBM
configuration tools can better understand what is required.

There is a maximum of 16 EXP24s Drawers per PCIe Gen3 Drawer (#EMX0)
to enable SAS cables to be properly handled by the EMX0 cable
management bracket.

DVD and boot devices

A device capable of reading a DVD must be attached to the system
and available to perform operating system installation, maintenance,
problem determination, and service actions such as maintaining system
firmware and I/O microcode at their latest levels. Alternatively, the
system must be attached to a network with software such as AIX NIM
server or Linux Install Manager configured to perform these functions.

System boot is supported through three options:

Disk or SSD located in an EXP24S drawer attached to a PCIe adapter

A network through LAN adapters

A SAN attached to Fibre Channel or FCoE adapters and indicated to
the server by the 0837 specify code

Assuming option 1 above, the minimum system configuration requires
at least one SAS disk drive in the system for AIX and Linux and two
for IBM i. If you are using option 3 above, a disk or SSD drive is not
required.

Each system control unit enclosure can have one slim-line bay that
can support one DVD drive (#EU13). The feature EU13 DVD is cabled to
a USB PCIe adapter located in either the system node or in a PCIe Gen3
I/O drawer. A USB to SATA converter is included in the configuration
without a separate feature code. The 1.6M USB cable #EBK4 is only
used to attach the #EU13 DVD drive.

For IBM i, a DVD drive must be available on the server when
required. The DVD can be in the system control unit or it can be
located in an external enclosure such as a 7226-1U3 Multimedia drawer.

Racks

The Power E880 is designed to fit a standard 19-inch rack. IBM
Development has tested and certified the system in the IBM Enterprise
rack (7014-T42, 7014-T00, #0551, or #0553). The client can choose to
place the server in other racks if they are confident those racks have
the strength, rigidity, depth, and hole pattern characteristics that
are needed. The client should work with IBM Service to determine
other racks' appropriateness. The Power E880 rails can adjust their
depth to fit a rack that is 23.25 inches - 30.4375 inches in depth.

It is highly recommended that the Power E880 be ordered with an IBM
42U enterprise rack (7014-T42 or #0553). An initial system order is
placed in a 7014-T42 rack. A same serial-number model upgrade MES is
placed in a feature 0553 rack. This is done to ease and speed client
installation, provide a more complete and higher quality environment
for IBM Manufacturing system assembly and testing, and provide a more
complete shipping package.

The 7014-T42 or feature 0553 is a 2-meter enterprise rack
providing 42U or 42 EIA of space. Clients who don't want this rack can
remove it from the order, and IBM Manufacturing will then remove the
server from the rack after testing and ship the server in separate
packages without a rack. Use the factory-deracking code ER21 on the
order to do this.

Additional E880 PCIe Gen3 I/O drawers (#EMX0) for an already
installed server can be MES ordered with or without a rack. When
clients want IBM Manufacturing to place these MES I/O drawers into a
rack and ship them together (factory integration), then the racks
should be ordered as feature codes on the same order as the I/O
drawers. Use feature 0553 (42U enterprise rack) for this order.

Three rack front door options are supported with Power E880 system
nodes for the 42U enterprise rack (7014-T42 or #0553), the acoustic
door (#6249), the attractive geometrically accented door (#ERG7) and
the cost-effective plain front door (#6069). The front trim kit is
also supported (#6272). The Power 880 logo rack door (#6250) is not
supported.

It is strongly recommended that the bottom 2U of the rack be left
open for cable management when below-floor cabling is used. Likewise,
if overhead cabling is used, it is strongly recommended the top 2U be
left open for cable management. If clients are using both overhead
and below-floor cabling, leaving 2U open on both the top and bottom of
the rack is a good practice. Rack configurations placing equipment in
these 2U locations can be more difficult to service if there are a lot
of cables running by them in the rack.

The system node and system control unit must be immediately
physically adjacent to each other in a contiguous space. The cables
connecting the system control unit and the system node are built to
very specific lengths. In a two-node configuration, system node 1 is
on top, and then the system control unit in the middle and system node
2 is on the bottom. Use specify feature ER16 to reserve 5U space in
the rack for a future system node and avoid the work of shifting
equipment in the rack in the future. On a four-node configuration
system node 4 is on the top, then node 1 is below it, then the system
control unit, then node 2 and finally node 3 is on the bottom.

With the 2 meter 7014-T42 or feature 0553, a rear rack extension of
8 inches or 20.3 cm (#ERG0) provides space to hold cables on the side
of the rack and keep the center area clear for cooling and service
access. Including this extension is very, very strongly recommended
where large numbers of thicker I/O cables are present or may be added
in the future. The definition of a "large number" depends on the type
of I/O cables used. Probably around 64 short-length SAS cables per
side of a rack or around 50 longer-length (thicker) SAS cables per
side of a rack is a good rule of thumb. Generally, other I/O cables
are thinner and easier to fit in the sides of the rack and the number
of cables can be higher. SAS cables are most commonly found with
multiple EXP24S SAS drawers (#5887) driven by multiple PCIe SAS
adapters. For this reason, it can be a very good practice to keep
multiple EXP24S drawers in the same rack as the PCIe Gen3 I/O drawer
or in a separate rack close to the PCIe Gen3 I/O drawer, using
shorter, thinner SAS cables. The feature ERG0 extension can be good
to use even with a smaller numbers of cables as it enhances the ease
of cable management with the extra space it provides.

Multiple service personnel are required to manually remove or
insert a system node drawer into a rack, given its dimensions and
weight and content. To avoid any delay in service it is very strongly
recommended that the client obtain an optional lift tool (#EB2Z). One
EB2Z lift tool can be shared among many servers and I/O drawers. The
EB2Z lift tool provides a hand crank to lift and position up to 159 kg
(350 lb). The EB2Z lift tool is 1.12 meters x 0.62 meters (44 in. x
24.5 in.). Note that a single system node can weigh up to 75.7 kg
(167 lb).

System node power

Four ac power supplies provide 2 + 2 redundant power for enhanced
system availability. A system node is designed to continue
functioning with just two working power supplies. A failed power
supply can be hot swapped but must remain in the system until the
replacement power supply is available for exchange.

Four ac power cords are used for each system node (one per power
supply) and are ordered using the AC Power Chunnel feature (#EBAA).
The chunnel carries power from the rear of the system node to the hot
swap power supplies located in the front of the system node where they
are more accessible for service.

System control unit power

The system control unit is powered from the system nodes. UPIC
cables provide redundant power to the system control unit. Two UPIC
cables attach to system node drawer 1 and two UPIC cables attach to
system node drawer 2. They are ordered with features ECCA and ECCB.
Just one UPIC cord is enough to power the system control unit and the
rest are in place for redundancy.

Power distribution units (PDU)

The Power E880 server factory integrated into an IBM rack uses
horizontal PDUs located in the EIA drawer space of the rack instead of
the typical vertical PDUs found in the side pockets of a rack. This
is done to aid cable routing. Each horizontal PDU occupies 1U.
Vertically mounting the PDUs to save rack space can cause cable
routing challenges and interfere with optimal service access.

When mounting the horizontal PDUs, it is a good practice to place
them almost at the top or almost at the bottom of the rack, leaving 2U
or more of space at the very top or very bottom of the open for cable
management. Mounting a horizontal PDU in the middle of the rack is
generally not optimal for cable management.

Two possible PDU ratings are supported: 60A/63A (orderable in
most countries) and 30A/32A.

The 60A/63A PDU supports four system node power supplies and one
I/O expansion drawer or eight I/O expansion drawers.

The 30A/32A PDU supports two system node power supplies and one
I/O expansion drawer or four I/O expansion drawers.

Rack-integrated system orders require two of either feature 7109
or 7188.

Feature 7109 -- Intelligent PDU with Universal UTG0247 Connector
is for an intelligent ac power distribution unit (PDU+) that will
allow the user to monitor the amount of power being used by the
devices that are plugged in to this PDU+. This ac power distribution
unit provides twelve C13 power outlets. It receives power through a
UTG0247 connector. It can be used for many different countries and
applications by varying the PDU to Wall Power Cord, which must be
ordered separately. Each PDU requires one PDU to Wall Power Cord.
Supported power cords include the following features: 6489, 6491,
6492, 6653, 6654, 6655, 6656, 6657, and 6658.

Feature 7188 -- Power Distribution Unit mounts in a 19-inch rack
and provides twelve C13 power outlets. Feature 7188 has six 16A
circuit breakers, with two power outlets per circuit breaker. System
units and expansion units must use a power cord with a C14 plug to
connect to the feature 7188. One of the following line cords must be
used to distribute power from a wall outlet to the feature 7188:
feature 6489, 6491, 6492, 6653, 6654, 6655, 6656, 6657, or 6658.

Hot-plug options

The following options are hot-plug capable:

PCIe I/O adapters.

System node ac power supplies: Two functional power supplies must
remain installed at all times while the system is operating.

System node fans.

System control unit fans.

System control unit Op Panel.

System control unit DVD drive.

UPIC power cables from system node to system control unit.

If the system boot device or system console is attached using an
I/O adapter feature, that adapter may not be hot-plugged if a
nonredundant topology has been implemented.

PowerVM Enterprise virtualization is built into the Power E880
system and provides the complete set of PowerVM virtualization
functionality needed for Power enterprise servers with POWER8
technology. This enables efficient resource sharing through
virtualization, which allows workload consolidation and secure
workload isolation as well as the flexibility to redeploy resources
dynamically.

Other PowerVM technologies include the following:

Migrate from older generation Power servers to the Power E880
system.

System Planning Tool simplifies the process of planning and
deploying Power Systems LPARs and virtual I/O.

Virtual I/O Server (VIOS) is a single-function appliance that
resides in an IBM Power processor-based partition. It facilitates the
sharing of physical I/O resources between client partitions AIX,
Linux, or IBM i within the server.

With Live Partition Mobility, you can move a running AIX, Linux,
or IBM i LPAR from one physical server to another with no downtime.
Use this capability to do the following:

Evacuate workloads from a system before performing scheduled
maintenance.

Move workloads across a pool of different physical resources as
business needs shift.

Move workloads away from under-utilized machines so that they can
be powered off to save on energy and cooling costs. Active Memory
Sharing allows memory to be dynamically moved between running
partitions for optimal resource usage.

PowerVP Virtualization Performance monitor provides real-time
monitoring of a virtualized system showing the mapping of VMs to
physical hardware.

Note: Alternative configuration options are available on a special bid
basis from your IBM representative or Business Partner.

Active Memory Expansion

Active Memory Expansion is an innovative POWER7, POWER7+, and
POWER8 technology supporting the AIX operating system that allows the
effective maximum memory capacity to be much larger than the true
physical memory maximum. Sophisticated compression/decompression of
memory content can allow memory expansion up to 100% or more. This
can allow a partition to do significantly more work or support more
users with the same physical amount of memory. Similarly, it can
allow a server to run more partitions and do more work for the same
physical amount of memory.

Active Memory Expansion uses CPU resource to compress/decompress
the memory contents. The trade-off of memory capacity for processor
cycles can be an excellent choice, but the degree of expansion varies
on how compressible the memory content is. It also depends on having
adequate spare CPU capacity available for this
compression/decompression. Tests in IBM laboratories using sample
workloads showed excellent results for many workloads in terms of
memory expansion per additional CPU utilized. Other test workloads
had more modest results. Feedback from many POWER7 and POWER7+
clients using the function has been very positive.

POWER7+ and POWER8 chips include a hardware accelerator designed to
boost Active Memory Expansion efficiency and use less POWER core
resource. The POWER8 accelerator includes some minor enhancements and
also leverages POWER8 higher bandwidth and lower latency
characteristics.

You have a great deal of control over Active Memory Expansion
usage. Each individual AIX partition can turn on or turn off Active
Memory Expansion. Control parameters set the amount of expansion
desired in each partition to help control the amount of CPU used by
the Active Memory Expansion function. An IPL is required for the
specific partition that is turning memory expansion. Once turned on,
monitoring capabilities are available in standard AIX performance
tools such as lparstat, vmstat, topas, and svmon.

A planning tool is included with AIX, allowing you to sample actual
workloads and estimate both how expandable the partition's memory is
and how much CPU resource is needed. Any Power Systems model can run
the planning tool. In addition, a one-time, 60-day trial of Active
Memory Expansion is available to enable more exact memory expansion
and CPU measurements. You can request the trial using the Capacity on
Demand web page

Active Memory Expansion is enabled by chargeable hardware feature
EM82, which can be ordered with the initial order of the system node
or as an MES order. A software key is provided when the enablement
feature is ordered, which is applied to the system node. An IPL is
not required to enable the system node. The key is specific to an
individual system node and is permanent. It cannot be moved to a
different server.

The additional CPU resource used to expand memory is part of the
CPU resource assigned to the AIX partition running Active Memory
Expansion. Normal licensing requirements apply.

IBM i operating system

For clients loading the IBM i operating system, the four-digit
numeric QPRCFEAT value used on the 9119-MME or 9119-MHE is the same as
the four-digit numeric feature number for the processors used in the
system. For example, if the processor feature number in a system is
EPBA, the QPRCFEAT value for the system would be EPBA.

The QPRCFEAT value does not change with the addition of system
node enclosures.

Capacity backup offering (applies to IBM i only)

The Power System 880 Capacity Backup (CBU) designation can help
meet your requirements for a second system to use for backup, high
availability, and disaster recovery. It enables you to temporarily
transfer IBM i processor license entitlements and 5250 Enterprise
Enablement entitlements purchased for a primary machine to a secondary
CBU-designated system. Temporarily transferring these resources
instead of purchasing them for your secondary system may result in
significant savings. Processor activations cannot be transferred as
part of this CBU offering, but programs such as Power Enterprise Pools
are available for the function.

The CBU specify feature number 4891 is available only as part of a
new server purchase or during an MES upgrade from an existing system
to a 9119-MHE. Certain system prerequisites must be met, and system
registration and approval are required before the CBU specify feature
can be applied on a new server. A used system that has an existing
CBU feature cannot be registered. The only way to attain a CBU
feature that can be registered is with a plant order.

Standard IBM i terms and conditions do not allow either IBM i
processor license entitlements or 5250 OLTP (Enterprise Enablement)
entitlements to be transferred permanently or temporarily. These
entitlements remain with the machine they were ordered for. When you
register the association between your primary and on-order CBU system,
you must agree to certain terms and conditions regarding the temporary
transfer.

After a CBU system designation is approved and the system is
installed, you can temporarily move your optional IBM i processor
license entitlement and 5250 Enterprise Enablement entitlements from
the primary system to the CBU system when the primary system is down
or while the primary system processor cores are inactive. The CBU
system can then better support fail-over and role swapping for a full
range of test, disaster recovery, and high availability scenarios.
Temporary entitlement transfer means that the entitlement is a
property transferred from the primary system to the CBU system and may
remain in use on the CBU system as long as the registered primary and
CBU system are in deployment for the high availability or disaster
recovery operation.

The primary system for a E880 server can be any of the following:

9119-FHB

9119-MHE

9117-MMB

9117-MMC

9117-MMD

9179-MHB

9179-MHC

9179-MHD

These systems have IBM i software licenses with an IBM i P30
software tier, or higher. The primary machine must be in the same
enterprise as the CBU system.

Before you can temporarily transfer IBM i processor license
entitlements from the registered primary system, you must have more
than one IBM i processor license on the primary machine and at least
one IBM i processor license on the CBU server. An activated processor
must be available on the CBU server to use the transferred
entitlement. You may then transfer any IBM i processor entitlements
above the minimum one, assuming the total IBM i workload on the
primary system does not require the IBM i entitlement you would like
to transfer during the time of the transfer. During this temporary
transfer, the CBU system's internal records of its total number of IBM
i processor license entitlements are not updated, and you may see IBM
i license noncompliance warning messages from the CBU system. Such
messages that arise in this situation do not mean you are not in
compliance.

Before you can temporarily transfer 5250 entitlements, you must
have more than one 5250 Enterprise Enablement entitlement on the
primary server and at least one 5250 Enterprise Enablement entitlement
on the CBU system. You may then transfer the entitlements that are
not required on the primary server during the time of transfer and
that are above the minimum of one entitlement. Note that if you are
using software replication (versus PowerHA), you may well need more
than a minimum of one entitlement on the CBU to support the
replication workload.

For example, if you have a 64-core Power 780 as your primary system
with twenty IBM i processor license entitlements (nineteen above the
minimum) and two 5250 Enterprise Enablement entitlements (one above
the minimum), you can temporarily transfer up to nineteen IBM i
entitlements and one 5250 Enterprise Enablement entitlement. During
the temporary transfer, the CBU system's internal records of its total
number of IBM i processor entitlements is not updated, and you may see
IBM i license noncompliance warning messages from the CBU system.

If your primary or CBU machine is sold or discontinued from use,
any temporary entitlement transfers must be returned to the machine on
which they were originally acquired.

Power Enterprise Pools

Power Enterprise Pools provide a level of flexibility and value for
systems that operate together as a pool of resources. Power
Enterprise Pools mobile activations are available for use on the Power
770, 780, and 795 systems and now on the new Power E870 and E880
systems. They can be assigned to any system in a predefined pool by
the user with simple HMC commands. IBM does not need to be notified
when these resources are reassigned within a pool. The simplicity of
operations offers new flexibility when managing large workloads in a
pool of systems. This capability is especially appealing to aid in
providing continuous application availability during maintenance
windows. Not only can workloads easily move to alternate systems, but
now the activations can move as well.

Now with the availability of the new Power E870 and E880 systems,
IBM continues to enhance the ability to freely move processor and
memory activations from one system to another system in the same pool,
without the need for IBM involvement. This capability now allows the
movement of resources not only between like systems but also between
generations of Power Systems, and thus delivering unsurpassed
flexibility for workload balancing and system maintenance.

Now, more than ever, Power Enterprise Pools delivers the support to
meet client's business goals when it comes to the following:

Improving the flexibility, load balancing, and disaster recovery
planning and operations of your Power Systems infrastructure

Enhanced reliability, availability, and serviceability (RAS) to
handle the requirements to accommodate a global economy

Prerequisites for Power Enterprise Pools:

All systems in a pool must be attached to the same HMC (or
redundant set of HMCs).

Systems must be one of the following models: 9117-MMD, 9179-MHD,
9119-FHB, 9119-MME, or 9119-MHE.

Systems must be within one country.

Two types of pools are available. One that enables Power 770
(9117-MMD)
or Power E870 (9119-MME) class systems to run in the same pool and the
other that enables Power 780 (9179-MHD), Power 795 (9119-FHB), and Power
E880 (9119-MHE) class systems to run in the same pool.
Both pools allow both processor and memory activations to move
between servers within the pool. Systems with different clock speeds
are supported, co-existing within the same pool. All systems in the
pool must be attached to the same HMC.

Power Systems users now have the satisfaction of knowing that as
their requirements change, so can their systems. A simple movement of
activations from one system to another helps users rebalance resources
and respond to business needs. Maintenance windows now open up more
easily as both workloads and activations move transparently across
systems. Even disaster recovery planning becomes more manageable with
the ability to move activations where and when they are needed. Power
Enterprise Pools are just one more reason why enterprise class servers
from Power Systems deliver value for your ever-changing business.

Mobile and static activations

A new, more flexible activation type is employed for Power
Enterprise Pools. Historically, only "static" activations that could
not move from server to server were available. These static
activations remain available on the Power 770, 780, and 795, and are
announced on the E870 and E880 and a certain number are required per
server. But mobile activation features can be moved in the Power
Enterprise Pool. Mobile activations apply to both processor core
activations and memory activations.

The new Power E870 and E880 must have at least eight cores
activated in static capability. All remaining processor core
activations on these systems can optionally be mobile activations, be
static activations, or be a mixture. Static and mobile core
activations can co-reside in the same system and in the same
partition.

A minimum of 50% of the memory on the system must be active. A
maximum of 75% of all physically installed memory can have mobile
activations. A minimum of 25% of all memory activations on a server
must have static activations. Static and mobile memory activations
can co-reside in the same system and in the same partition. Mobile
activation feature codes are in 100 GB increments.

Existing static activation features can be converted to mobile
activations for memory and cores. To provide administrative and
pricing advantages, there are "regular" static core activations and
"mobile-enabled" static core activations. The price of a
mobile-enabled core activation is priced the same as a mobile core
activation. However, there is no-charge for converting a
mobile-enabled activation to a mobile activation. This allows the client
to avoid the additional conversion premium charge when converting an
existing static core activation to a mobile core activation.

The mobile activation features are as follows:

For the Power E870 and E880 100 GB Mobile Memory Activation (#EMA7)
For the Power E870 and E880 100 GB Mobile Enabled Memory Activation
(#EMA9)
For the Power E870 1-core Mobile Activation for EPBA (#EPBJ)
For the Power E870 1-core Mobile Activation for EPBC (#EPBL)
For the Power E870 1-core Mobile-Enabled Activation for EPBA
(#EPBN)
For the Power E870 1-core Mobile-Enabled Activation for EPBC
(#EPBQ)
For the Power E880 1-core Mobile Activation for EPBC (#EPBK)
For the Power E880 1-core Mobile-Enabled Activation for EPBB
(#EPBP)

The Power Enterprise Pools mobile activation feature codes continue
to exist for the Power 770, 780, and 795 servers and can co-exist in
the same pool as the new Power E870 and E880 feature codes. The Power
770, 780, and 795 mobile activations feature codes are as follows:

Each Power Enterprise Pool is managed by a single master HMC (physical
hardware HMC or a virtual appliance (vHMC)). The HMC that was used to
create a Power Enterprise Pool is set as the master HMC of that pool.
After a Power Enterprise Pool is created, a redundant HMC can be
configured as a backup. All Power Enterprise Pool resource assignments
must be performed by the master HMC. When powering on or restarting a
server, ensure that the server is connected to the master HMC. This
ensures that the required Mobile CoD resources are assigned to the
server.

The maximum number of systems in a Power Enterprise Pool is 32
high-end or 48 mid-range systems. An HMC can manage multiple Power
Enterprise Pools but is limited to 1000 total partitions. The HMC can
also manage systems that are not part of the Power Enterprise Pool.
Powering down an HMC does not limit the assigned resources of
participating systems in a pool but does limit the ability to perform
pool change operations.

After a Power Enterprise Pool is created, the HMC can be used to
perform the following functions:

Mobile CoD processor and memory resources can be assigned to
systems with inactive resources. Mobile CoD resources remain on the
system to which they are assigned until they are removed from the
system.

New systems can be added to the pool and existing systems can be
removed from the pool.

New resources can be added to the pool or existing resources can
be removed from the pool.

Pool information can be viewed, including pool resource
assignments, compliance, and history logs.

Power Enterprise Pools qualifying machines

To qualify for use of the Power Enterprise Pool offering, a
participating system must be one of the following:

IBM Power E880 with POWER8 processors, designated as 9119-MHE

IBM Power E870 with POWER8 processors, designated as 9119-MME

IBM Power 795 with POWER7 processors, designated as 9119-FHB

IBM Power 780 with POWER7+ processors, designated as 9179-MHD

IBM Power 770 with POWER7+ processors, designated as 9117-MMD

Each system must have installed Machine Code release level 7.8.0,
or later, and be configured with at least the minimum amount of
permanently active processor cores (listed below). Processor and
memory activations that are enabled for movement within the pool will
be in addition to these base minimum configurations.

Ordering Power Enterprise Pools

Ordering and enabling mobile activations for enterprise class
systems is accomplished by following these steps:

Complete and submit the Power Enterprise Pools contract and
addendum (Z126-6228 and Z126-6229), specifying all system serial
numbers to be included in the pool. To generate a pool ID number,
send a copy to the Power Systems CoD Project Office at
pcod@us.ibm.com. This IBM License Supplement for Power Enterprise
Pools (Z126-6228) is required prior to ordering mobile resources but
is only required once per client. The IBM License Supplement for
Power Enterprise Pools Addendum (Z126-6229) is used to assign or
remove systems to or from a pool.

Order mobile enablement, processor, and memory activation features
for participating systems. Every system in the pool must have feature
EB35 as an identifier.

Ensure all participating systems and controlling HMCs have the
proper level of supporting software (eFW 7.8, or later, for systems;
V7.8, or later, for HMCs)

When the order is fulfilled, a configuration file will be
generated that contains a Power Enterprise Pool membership activation
code for each of the systems in the pool along with the mobile
processor and memory activations. This file will be made available on
the IBM COD website at

Download the client-specific configuration file with mobile
activations to the controlling HMC for the pool. The file will work
only for the specified system serial numbers. A new file will be
generated when systems or mobile resources are added or removed from
the pool.

Adding or removing systems from Power Enterprise Pools

Adding or removing a system from an established Power Enterprise
Pool requires notification to IBM. An updated addendum must be
submitted to the Power Systems CoD Project Office (pcod@us.ibm.com) to
make this change. When the update is processed, a new pool
configuration file will be posted on the CoD website and must be
downloaded to the controlling HMC.

Before removal from a pool, all assets (including mobile resources)
that were originally purchased with the system must be returned to
that same system serial number. Mobile assets belonging to a system
may qualify for transfer to another system serial number, depending on
specific qualifying guidelines, and will require additional
administrative action.

Systems removed from a pool can join another pool and contribute
mobile activation resources to the new pool or use another system's
mobile activation resources. Mobile activations require a pool ID to
be recognized.

Capacity on demand

Several types of capacity on demand (CoD) processors are optionally
available for the Power E880 system node. They help meet changing
resource requirements in an on demand environment by using resources
installed on the system but not activated.

Capacity upgrade on demand (CUOD) enables you to
purchase additional permanent processor or memory capacity and
dynamically activate it when needed.

Elastic capacity on demand (elastic CoD) enables
processors or memory to be temporarily activated in full-day
increments as needed. Charges are based on usage reporting collected
monthly. Processors and memory can be activated and turned off an
unlimited number of times, whenever you want additional processing
resources. With this offering, system administrators have an
interface at the HMC to manage the activation and deactivation of
resources. A monitor that resides on the server logs the usage
activity. You must send this usage data to IBM monthly. A bill is
then generated based on the total amount of processor and memory
resources utilized, in increments of processor and memory (1 GB) days.
Before using temporary capacity on your server, you must enable your
server. To do this, order an enablement feature (MES only) and sign
the required contracts.

If a Power E880 system node uses the IBM i operating system and the
temporarily activated cores were used for IBM i partitions, you must
inform the sales team placing the billing feature order which
operating system caused the temporary elastic CoD processor use so
that the correct feature can be used for billing.

Use the following features to order enablement features and support
billing charges on the Power E880:

Note: Feature EMA5 is for 1 GB Memory activation and feature EMA6 is
for 100 of feature EMA5 Memory activations.

All initial shipments of Power E880 system nodes come with a
specific number of no-charge elastic processor and memory days. The
number of days depends on the configuration of the system. For every
new Power E880, 15 elastic CoD processor days for every processor core
(inactive or active) installed on the system and 240 GB of elastic CoD
memory days for every processor core (inactive or active) are
included. MES orders do not include the no-charge elastic processor
and memory days.

The following features provide no-charge elastic processor and
memory days:

48 Proc-Days of Elastic CoD Temporary Processor Resources (#EPJ3)

72 Proc-Days of Elastic CoD Temporary Processor Resources (#EPJ5)

384 GB-Days of Elastic CoD Memory Resources (#EMJ8)

576 GB-Days of On/Off CoD Temporary Memory Resources (#EMJ9)

The Elastic CoD process consists of three steps: enablement,
activation, and billing

Elastic CoD enablement: Description

Before requesting temporary capacity on a server, you must "enable"
it for elastic CoD. To do this, order a no-charge enablement feature
(MES only) and sign the required contracts. IBM will generate an
enablement code, mail it to you, and post it on the web for you to
retrieve and enter on your server. A processor enablement code lets
you request up to 90 processor days of temporary unused CoD processor
capacity for all your processor cores that have not been permanently
activated. For example, if you have 20 processor cores that are not
permanently activated, the processor enablement code allows up to
1,800 processor days (20 x 90). If you have reached or are about to
reach the limit of 90 processor days per unactivated processor core,
place an order for another processor enablement code to reset the
number of days you can request. Similarly, a memory enablement code
lets you request up to 90 days of temporary unused CoD memory capacity
for all your gigabytes of memory that have not been permanently
activated. For example, if you had 100 GB of memory that was not
permanently activated, the memory enablement code allows up to 9000 GB
memory days (100 x 90). If you have reached the limit of 90 memory
days per unactivated memory, place an order for another memory
enablement code to reset the number of days you can request. Before
ordering a new enablement code for either memory or processor
temporary CoD, you must first process an MES delete order, deleting
the current enablement code installed in the server configuration
file.

Elastic CoD enablement: Step-by-step

Prerequisite 1: The sales channel (IBM Business Partner) must sign
one of the following contracts, if applicable:

Step 1: The client initiates the request for elastic CoD use by
asking the sales channel to enable the machine for temporary capacity.

Step 2: The client must complete and sign the following
contracts. It is the sales channel's responsibility to return the
signed contract to the responsible CSO organization and fax a copy to
IBM at 507-253-4553 or email a copy to tcod@us.ibm.com.

Note: The order for an enablement feature will not be fulfilled until
this step is completed.

Step 5: IBM generates an enablement code, mails it, and posts it.

Step 6: The client retrieves the enablement code and applies it
to the system node.

Elastic activation requests: Description

When Elastic CoD temporary capacity is needed, simply use the HMC
menu for elastic CoD and specify how many of the inactive processors
or how many gigabytes of memory you would like temporarily activated
for some number of days. You will be billed for the days requested,
whether the capacity is assigned to partitions or left in the shared
processor pool. At the end of the temporary period (days you
requested), you must ensure the temporarily activated capacity is
available to be reclaimed by the server (not assigned to partitions)
or you will be billed for any processor days not returned (per the
contract you signed).

Elastic CoD activation requests: Step-by-step

When you need temporary capacity, use the Elastic CoD temporary
capacity HMC menu for the server and specify how many of the inactive
processors or how many gigabytes of memory you would like temporarily
activated for some number of days. The user must assign the temporary
capacity to a partition (whether or not the machine is configured for
LPAR) to begin using temporary capacity.

Elastic CoD billing: Description

The contract, signed by the client before receiving the enablement
feature, requires the elastic CoD user to report billing data at least
once a month (whether there is activity or not). This data is used to
determine the proper amount to bill at the end of each billing period
(calendar quarter). Failure to report billing data for use of
temporary processor or memory capacity during a billing quarter will
result in default billing equivalent to 90 processor days of temporary
capacity. The sales channel will be notified of client requests for
temporary capacity. As a result, the sales channel must order a
quantity of billing features (using the appropriate billing features
EPJ3, EPJ5, EPJC, EPJD, EPJE, EPJF, EPJQ, EPJR, EPJS, or EPJT for each
billable processor and memory day reported less any outstanding credit
balance of procesor and memory days).

Elastic CoD billing: Step-by-step

The client must report billing data (requested and unreturned
processor and memory days) at a minimum of once per month either
electronically or by fax (stated requirement in the signed contract).
At the end of each billing period (calendar quarter), IBM will process
the accumulated data reported and notify the sales channel for proper
billing. The sales channel places an order for the appropriate
quantity of billing features (one processor billing feature ordered
for each processor day used, or one memory day for each memory day
utilized). IBM will ship a billing notice (notifies client of billing
actions) to the ship-to address on the order as part of the fulfilment
process. The client pays the sales channel and the sales channel pays
IBM for the fulfillment of the billing features.

For more information regarding registration, enablement, and usage
of elastic CoD, visit

Utility COD

Utility CoD provides additional processor performance on a
temporary basis within the shared processor pool. Utility CoD enables
you to place a quantity of inactive processors into the system node's
shared processor pool, which then becomes available to the pool's
resource manager. When the system node recognizes that the combined
processor utilization within the shared pool exceeds 100% of the level
of base (purchased/active) processors assigned across uncapped
partitions, then a Utility CoD Processor Minute is charged and this
level of performance is available for the next minute of use. If
additional workload requires a higher level of performance, the system
will automatically enable the additional Utility CoD processors to be
used. The system continuously monitors and charges for the
performance needed above the base (permanent) level. Registration and
usage reporting for Utility CoD is made using a public website and
payment is based on reported usage. Utility CoD requires PowerVM
Enterprise Edition to be active on the 9119-MHE.

If a Power E880 system node uses the IBM i operating system and the
temporarily activated cores were used for IBM i partitions, the client
must inform the sales team placing the billing feature order which
operating system caused the temporary Utility CoD processor use so
that the correct feature can be used for billing.

The reliability of systems starts with components, devices, and
subsystems that are designed to be highly reliable. During the design
and development process, subsystems go through rigorous verification
and integration testing processes. During system manufacturing,
systems go through a thorough testing process to help ensure the
highest level of product quality.

Redundant infrastructure

Considerable redundancy in the infrastructure of these systems is
included so as to avoid failing components leading to system outages.

Such components include power supplies, fans, processor and memory
voltage regulation outputs, global service processors, and processor
clocks.

All of these redundant elements are present, even in single-system
node systems.

Processor and memory availability functions

The Power Systems family continues to offer and introduce
significant enhancements designed to increase system availability.

POWER8 processor functions

As previously provided in POWER7 and POWER7+, the POWER8 processor
has the ability to do processor instruction retry for some transient
errors and alternate processor recovery for a number of core-related
faults. This significantly reduces exposure to both hard (logic) and
soft (transient) errors in the processor core. Soft failures in the
processor core are transient (intermittent) errors, often due to
cosmic rays or other sources of radiation, and generally are not
repeatable. When such an error is encountered in the core, the POWER8
processor will first automatically retry the instruction. If the
source of the error was truly transient, the instruction will succeed
and the system will continue as before.

Hard failures are more difficult, being true logical errors that
will be replicated each time the instruction is repeated. Retrying
the instruction will not help in this situation. As POWER7/POWER7+
technology, processors have the ability to extract the failing
instruction from the faulty core and retry it elsewhere in the system
for a number of faults, after which the failing core is dynamically
deconfigured and called out for replacement in the PowerVM
environment. These features are designed to avoid a full system
outage.

As in POWER7/POWER7+, the POWER8 processor includes single
processor check stopping for certain faults that cannot be handled by
the availability enhancements described in the preceding section.
This significantly reduces the probability of any one processor
affecting total system availability.

Partition availability priority

Also available is the ability to assign availability priorities to
partitions. In the PowerVM environment, if an alternate processor
recovery event requires spare processor resources in order to protect
a workload, when no other means of obtaining the spare resources is
available, the system will determine which partition has the lowest
priority and attempt to claim the needed resource. On a properly
configured POWER8 processor-based server, this allows that capacity to
be first obtained from, for example, a test partition instead of a
financial accounting system.

Cache availability

The L2 and L3 caches in the POWER8 processor and L4 cache in the
memory buffer chip are protected with double-bit detect, single-bit
correct error detection code (ECC). In addition, a threshold of
correctable errors detected on cache lines can result in the data in
the cache lines being purged and the cache lines removed from further
operation without requiring a reboot in the PowerVM environment. In
addition, the L2 and L3 caches have the ability to dynamically
substitute a spare bit-line for a faulty bit-lane, allowing an entire
faulty "column" of cache, impacting multiple cache lines, to be
repaired. An ECC uncorrectable error detected in these caches can
also trigger a purge and delete of cache lines. This results in no
loss of operation if the cache lines contained data unmodified from
what was stored in system memory.

Modified data would be handled through Special Uncorrectable Error
handling. L1 data and instruction caches also have a retry capability
for intermittent errors and a cache set delete mechanism for handling
solid failures.

Special Uncorrectable Error handling

Special Uncorrectable Error (SUE) handling is designed to prevent
an uncorrectable error in memory or cache from immediately causing the
system to terminate. Rather, the system tags the data and determines
whether it will ever be used again. If the error is irrelevant, it
will not force a check stop. If the data is used, termination may be
limited to the program/kernel or hypervisor owning the data; or the
I/O adapters controlled by an I/O hub controller would freeze if data
were transferred to an I/O device.

Memory error correction and recovery

The memory has error detection and correction circuitry designed
such that the failure of any one specific memory module within an ECC
word by itself can be corrected absent any other fault.

In addition, a spare DRAM per rank on each memory port provides for
dynamic DRAM device replacement during runtime operation. Also,
dynamic lane sparing on the DMI link allows for repair of a faulty
data lane.

Other memory protection features include retry capabilities for
certain faults detected at both the memory controller and the memory
buffer. Memory is also periodically scrubbed to allow for soft errors
to be corrected and for solid single-cell errors reported to the
hypervisor, which supports operating system deallocation of a page
associated with a hard single-cell fault.

Active memory mirroring for the hypervisor

The POWER8 memory subsystem is capable of mirroring sections of
memory by writing to two different memory locations, and when an error
is detected when reading from one location, taking data from the
alternate location. This is used by the POWER hypervisor in these
systems to mirror critical memory within the hypervisor so that a
fault, even a solid uncorrectable error in the data, can be tolerated
using the mirrored memory.

Dynamic processor and memory deallocation

When correctable solid faults occur in components of the processor
and memory subsystem, the system will attempt to correct the problem
by using spare capacity in the failing component, using a spare column
in an L2 or L3 cache, for example, a spare data line in a memory or
processor bus, or a spare DRAM in memory. Use of such spare capacity
restores the system to full functionality without the need to take a
repair action.

When such spare capacity is not available, the service processor
and POWER hypervisor may request deallocation of the component
experiencing the fault. When there are sufficient resources to
continue running partitions at requested capacity, the system will
continue to do so. This includes taking advantage of unlicensed
capacity update on demand processor and memory resources as well as
licensed but unallocated resources.

When such unlicensed or unused capacity is used in this manner, a
request for service will be made.

PCI extended error handling

PCI extended error handling (EEH)-enabled adapters respond to a
special data packet generated from the affected PCI slot hardware by
calling system firmware, which will examine the affected bus, allow
the device driver to reset it, and continue without a system reboot.
For Linux, EEH support extends to the majority of frequently used
devices, although some third-party PCI devices may not provide native
EEH support.

Mutual surveillance

The service processor monitors the operation of the firmware during
the boot process and also monitors the hypervisor for termination.
The hypervisor monitors the service processor and reports the service
reference code when it detects surveillance loss. In the PowerVM
environment, it will perform a reset/reload if it detects the loss of
the service processor.

Environmental monitoring functions

The Power Systems family does ambient and over temperature
monitoring and reporting.

Uncorrectable error recovery

When the auto-restart option is enabled, the system can
automatically restart following an unrecoverable software error,
hardware failure, or environmentally induced (ac power) failure.

Serviceability

The purpose of serviceability is to efficiently repair the system
while attempting to minimize or eliminate impact to system operation.
Serviceability includes system installation, MES (system
upgrades/downgrades), and system maintenance/repair. Depending upon
the system and warranty contract, service may be performed by the
customer, an IBM representative, or an authorized warranty service
provider.

The serviceability features delivered in this system provide a
highly efficient service environment by incorporating the following
attributes:

Design for SSR Set Up and Customer Installed Features (CIF).

Detection and Fault Isolation (ED/FI).

First Failure Data Capture (FFDC).

Guiding Light service indicator architecture is used to control a
system of integrated LEDs that lead the individual servicing the
machine to the correct part as quickly as possible.

Service labels, service cards, and service diagrams available on
the system and delivered through the HMC.

Step-by-step service procedures available through the HMC.

Service environment

The POWER8 processor-based system requires attachment to one or
more HMCs.

The HMC is a dedicated server that provides functions for
configuring and managing servers for either partitioned or full-system
partition using a GUI or command-line interface (CLI). An HMC
attached to the system allows support personnel (with client
authorization) to remotely log in to review error logs and perform
remote maintenance if required.

The I/O device and adapter diagnostics consists of stand-alone
diagnostics, which are loaded from the DVD-RAM drive and online
diagnostics. Online diagnostics, when installed, are resident with
the AIX operating system on the disk or system. They can be booted in
single-user mode (service mode), run in maintenance mode, or run
concurrently (concurrent mode) with other applications. They have
access to the AIX error log and the AIX configuration data.

Service mode enables checking of system devices and features.

Concurrent mode allows the normal system functions to continue
while selected resources are being checked.

Maintenance mode enables checking of devices and adapters.

Note: Because the 9119-MME and 9119-MHE systems have an optional
DVD-RAM, alternative methods for maintaining and servicing the system
need to be available if the DVD-RAM is not ordered. An external
Internet connection must be available to maintain or update system
firmware to the latest required level.

Concurrent maintenance guided service procedures will continue to
be supported by the Repair and Verify (R&V) component of the Service
Focal Point application running on the HMC. Repair procedures that
are not covered by the guided R&V component will be documented and
available for display on any web browser-enabled system as well as on
the HMC. These procedures are available through IBM Knowledge Center.

Service interface

The service interface allows support personnel to communicate with
the service support applications in a server using a console, an
interface, or a terminal. Delivering a clear, concise view of
available service applications, the service interface allows the
support team to manage system resources and service information in an
efficient and effective way. Applications available through the
service interface are carefully configured and placed to give service
providers access to important service functions.

Different service interfaces are used, depending on the state of
the system, hypervisor, and operating environment. The primary
service interfaces are:

Service Indicators

Operator Panel

Service Processor menu

Operating system service menu

Service Focal Point on the HMC

In the Guiding Light service indicator implementation, when a fault
condition is detected on the POWER8 processor-based server, a blue
Enclosure Fault LED will illuminate on the enclosure containing the
failing part. The Guiding Light system pinpoints the exact part by
blinking the amber FRU identify LED associated with the part to be
replaced when selected by the servicer as part of the repair
procedure. This action will roll up to the blue enclosure locate LED
and down to the individual component to be serviced.

First Failure Data Capture and error data analysis

First Failure Data Capture (FFDC) is a technique that helps ensure
that when a fault is detected in a system, the root cause of the fault
will be captured without the need to re-create the problem or run any
sort of extending tracing or diagnostics program. For the vast
majority of faults, a good FFDC design means that the root cause can
also be detected automatically without servicer intervention.

FFDC information, error data analysis, and fault isolation are
necessary to implement the advanced serviceability techniques that
enable efficient service of the systems and to help determine the
failing items.

In the rare absence of FFDC and Error Data Analysis, diagnostics
are required to re-create the failure and determine the failing items.

Diagnostics

General diagnostic objectives are to detect and identify problems
so they can be resolved quickly. Elements of IBM's diagnostics
strategy include:

Provide a common error code format equivalent to a system
reference code with PowerVM, system reference number, checkpoint, or
firmware error code.

Provide fault detection and problem isolation procedures. Support
remote connection ability to be used by the IBM Remote Support Center
or IBM Designated Service.

Because of the FFDC technology designed into IBM servers, it is not
necessary to perform re-create diagnostics for failures or require
user intervention. Solid and intermittent errors are designed to be
correctly detected and isolated at the time the failure occurs.
Runtime and boot-time diagnostics fall into this category.

Stand-alone diagnostics with PowerVM

As the name implies, stand-alone or user-initiated diagnostics
requires user intervention. The user must perform manual steps,
including:

Booting from the diagnostics CD, DVD, USB, or network

Interactively selecting steps from a list of choices

Concurrent maintenance

It is expected that the majority of the components that will fail
will be able to be replaced using concurrent maintenance. These
include power supplies, power cable, fans, op panel, real-time clock
battery, and PCIe adapters.

The system also supports updating firmware with service packages,
typically concurrently. The determination of whether a firmware
release can be updated concurrently is identified in the readme
information file that is released with the firmware.

Service labels

Service providers use these labels to assist them in performing
maintenance actions. Service labels are found in various formats and
positions and are intended to transmit readily available information
to the servicer during the repair process. Following are some of
these service labels and their purpose:

Location diagrams: Location diagrams are located on the system
hardware, relating information regarding the placement of hardware
components.

Location diagrams may include location codes, drawings of physical
locations, concurrent maintenance status, or other data pertinent to a
repair. Location diagrams are especially useful when multiple
components such as DIMMs, CPUs, processor books, fans, adapter cards,
LEDs, and power supplies are installed.

Remove/replace procedures: Service labels that contain
remove/replace procedures are often found on a cover of the system or
in other spots accessible to the servicer. These labels provide
systematic procedures, including diagrams, detailing how to remove or
replace certain serviceable hardware components.

Arrows: Numbered arrows are used to indicate the order of
operation and the serviceability direction of components. Some
serviceable parts such as latches, levers, and touch points need to be
pulled or pushed in a certain direction and in a certain order for the
mechanical mechanisms to engage or disengage. Arrows generally
improve the ease of serviceability.

Packing for service

The following service enhancements are included in the physical
packaging of the systems to facilitate service:

Color coding (touch points): Terracotta-colored touch points
indicate that the system may not be required to be powered off to
perform service to the FRU. This is dependent on system configuration
and preparatory steps may be required before the service action is
taken. Blue-colored touch points indicate that the procedure may
require that the unit or system be shut down before servicing. This
is dependent on system configuration, and preparatory steps may be
required before the service action is taken.

Tool-less design: Most FRUs support tool-less or simple tool
designs. These designs require no tools or simple tools such as
flathead screw drivers to service the hardware components.

A lift tool may be required for certain installation and service
situations (system backplane). The service and installation scenario
should be reviewed during system installation planning.

Positive retention: Positive retention mechanisms help to assure
proper connections between hardware components such as cables to
connectors, and between two cards that attach to each other. Without
positive retention, hardware components run the risk of becoming loose
during shipping or installation, preventing a good electrical
connection. Positive retention mechanisms like latches, levers,
thumb-screws, pop Nylatches (U-clips), and cables are included to help
prevent loose connections and aid in installing (seating) parts
correctly. These positive retention items do not require tools.

Error handling and reporting

In the unlikely event of system hardware or environmentally induced
failure, the system runtime error capture capability systematically
analyzes the hardware error signature to determine the cause of
failure. The analysis result will be stored in system NVRAM. When
the system can be successfully restarted either manually or
automatically, or if the system continues to operate, the error will
be reported to the operating system. Hardware and software failures
are recorded in the system log. An Error Log Analysis (ELA) routine
analyzes the error, forwards the event to the Service Focal Point
(SFP) application running on the HMC, and notifies the system
administrator that it has isolated a likely cause of the system
problem. The Service Processor event log also records unrecoverable
checkstop conditions, forwards them to the SFP application, and
notifies the system administrator. The system has the ability to call
home through OS to report platform recoverable errors and errors
associated with PCI adapters/devices. When the information is logged
in the SFP application, if the system is properly configured, a call
home service request will be initiated and the pertinent failure data
with service parts information and part locations will be sent to an
IBM service organization. Customer contact information and specific
system-related data such as the machine type, model, and serial
number, along with the error log data related to the failure, are sent
to IBM Service.

Live Partition Mobility

With Live Partition Mobility, users can migrate an AIX, Linux, or
IBM i partition running on one POWER partition system to another POWER
system without disrupting services. The migration transfers the
entire system environment, including processor state, memory, attached
virtual devices, and connected users. It provides continuous
operating system and application availability during planned partition
outages for repair of hardware and firmware faults.

When initially purchasing a system, the client needs to consider
the effect on the workloads that servicing the system will make. It
is expected that most of the system FRUs that fail will be replaced
while the system is still powered on, using Concurrent Maintenance.
For the remainder of the FRUs, the entire system must be powered off
to service the FRU. For customers with critical workloads, which
cannot afford to be stopped, the use of Live Partition Mobility needs
to be planned for when the initial purchase of the system is made.

Service processor

The service processor provides the capability to diagnose, check
the status of, and sense the operational conditions of a system. It
runs on its own power boundary and does not require resources from a
system processor to be operational to perform its tasks.

Under PowerVM the service processor supports surveillance of the
connection to the HMC and to the system firmware (hypervisor). It
also provides several remote power control options, environmental
monitoring, reset, restart, remote maintenance, and diagnostic
functions, including console mirroring. The service processors menus
(ASMI) can be accessed concurrently with system operation, allowing
nondisruptive abilities to change system default parameters.

Call home

Call home refers to an automatic or manual call from a customer
location to the IBM support structure with error log data, server
status, or other service-related information. Call home invokes the
service organization in order for the appropriate service action to
begin. Call home can be done through the HMC. While configuring call
home is optional, clients are encouraged to implement this feature in
order to obtain service enhancements such as reduced problem
determination and faster and potentially more accurate transmittal of
error information. In general, using the call home feature can result
in increased system availability. The Electronic Service Agent
application can be configured for automated call home. Refer to the
next section for specific details on this application.

IBM Electronic Services

Electronic Service Agent and the IBM Electronic Services web portal
comprise the IBM Electronic Services solution, which is dedicated to
providing fast, exceptional support to IBM customers. IBM Electronic
Service Agent is a no-charge tool that proactively monitors and
reports hardware events such as system errors, performance issues, and
inventory. Electronic Service Agent can help focus on the customer's
company business initiatives, save time, and spend less effort
managing day-to-day IT maintenance issues.

System configuration and inventory information collected by
Electronic Service Agent also can be viewed on the secure Electronic
Services web portal and used to improve problem determination and
resolution between the customer and the IBM support team. As part of
an increased focus to provide even better service to IBM customers,
Electronic Service Agent tool configuration and activation comes
standard with the system. In support of this effort, a new HMC
External Connectivity security whitepaper has been published, which
describes data exchanges between the HMC and the IBM Service Delivery
Center (SDC) and the methods and protocols for this exchange. To read
the whitepaper and prepare for Electronic Service Agent installation,
go to the "Security" section at

Electronic Service Agent is designed to enhance the warranty and
maintenance service by providing faster hardware error reporting and
uploading system information to IBM Support. This can optimize the
time monitoring the symptoms, diagnosing the error, and manually
calling IBM Support to open a problem record. And 24x7 monitoring and
reporting means no more dependency on human intervention or off-hours
customer personnel when errors are encountered in the middle of the
night.

Security: The Electronic Service Agent tool is designed to be
secure in monitoring, reporting, and storing the data at IBM. The
Electronic Service Agent tool is designed to securely transmit either
through the Internet (HTTPS or VPN) or modem to provide customers a
single point of exit from their site. Communication is one way.
Activating Electronic Service Agent does not enable IBM to call into a
customer's system.

Because system information and error logs are automatically
uploaded to the IBM Support Center in conjunction with the service
request, customers are not required to find and send system
information, decreasing the risk of misreported or misdiagnosed
errors. Once inside IBM, problem error data is run through a data
knowledge management system and knowledge articles are appended to the
problem record.

Customized support

Using the IBM ID entered during activation, customers can view
system and support information in the "My Systems" and "Premium
Search" sections of the Electronic Services website.

The Electronic Services web portal is a single Internet entry point
that replaces the multiple entry points traditionally used to access
IBM Internet services and support. This web portal enables you to
gain easier access to IBM resources for assistance in resolving
technical problems. The newly improved My Systems and Premium Search
functions make it even easier for Electronic Service Agent-enabled
customers to track system inventory and find pertinent fixes.

My Systems provides valuable reports of installed hardware and
software using information collected from the systems by IBM
Electronic Service Agent. Reports are available for any system
associated with the customer's IBM ID. Premium Search combines the
function of search and the value of Electronic Service Agent
information, providing advanced search of the technical support
knowledgebase. Using Premium Search and the Service Agent information
that has been collected from the system, customers are able to see
search results that apply specifically to their systems.

For more information on how to utilize the power of IBM Electronic
Services, visit the following website or contact an IBM Systems
Services Representative

PowerCare service

Included with Power E880 system is a PowerCare services option,
which entitles you to choose one of several high-value technical
service offerings from IBM to complement and assist in the deployment
of a new Power E880 system. This option is provided at no additional
charge and requires no additional approvals. By leveraging the
skills, experiences, and proven methodology of IBM Systems Lab
Services professionals, you can potentially increase the efficiency
and quality of your complex data center operations.

The PowerCare option is included with new Power E880 systems and
MES upgrades into a Power E880. The Lab Services PowerCare team will
contact the account team/customer after the shipment to help with the
selection of a PowerCare service offer.

Optionally, the account team or the client may contact the WW
PowerCare team at

pwrcare@us.ibm.com

The customer has up to 90 days from the installation date of the
Power E880 to select a PowerCare offering. Delivery of the selected
service must be completed within nine months of the installation date
of the Power E880 system.

Regardless the rack-integrated system to which the PCIe Gen3 I/O
Expansion Drawer is attached to, if the Expansion Drawer is ordered as
factory integrated, the PDUs in the rack will be defaulted to be placed
horizontally to enhance cable management.

Expansion Drawers complicate the access to vertical PDUs if located at
the same height. IBM recommends accommodating PDUs horizontally on racks
containing one or more PCIe Gen3 I/O Expansion Drawers.

Once the rack with Expansion Drawers is delivered to the customer, the
customer is allowed to rearrange the PDUs from horizontal to vertical.
However, the configurator will remain considering the PDUs are placed
horizontally for the matter of calculating the free space still available
in the rack.

Vertical PDUs can be used only if CSRP (#0469) is on the order. When
specifying CSRP, the customer will provide the locations where the PCIe
Gen3 I/O Expansion Drawers must be placed, avoiding locating those
adjacent to Vertical PDU locations, EIA 6 through 16 and 21 through 31.
Back to top

Devices supported

Model conversions

The existing components being replaced during a model or feature
conversion become the property of IBM and must be returned.

Feature conversions are always implemented on a "quantity of one
for quantity of one" basis. Multiple existing features may not be
converted to a single new feature. Single existing features may not
be converted to multiple new features.

The existing components being replaced during a model or feature
conversion become the property of IBM and must be returned.

Feature conversions are always implemented on a "quantity of one
for quantity of one" basis. Multiple existing features may not be
converted to a single new feature. Single existing features may not
be converted to multiple new features.

The Power E880 must be installed in a rack with a rear door and
side panels for EMC compliance. The native HMC Ethernet ports must
use shielded Ethernet cables.

Note: Government regulations (such as those prescribed by OSHA or
European Community Directives) may govern noise level exposure in the
workplace and may apply to you and your server installation. This IBM
system is available with an optional acoustical door feature that can
help reduce the noise emitted from this system. The actual sound
pressure levels in your installation depend upon a variety of factors,
including the number of racks in the installation; the size,
materials, and configuration of the room where you designate the racks
to be installed; the noise levels from other equipment; the room
ambient temperature, and employees' location in relation to the
equipment. Further, compliance with such government regulations also
depends upon a variety of additional factors, including the duration
of employees' exposure and whether employees wear hearing protection.
IBM recommends that you consult with qualified experts in this field
to determine whether you are in compliance with the applicable
regulations.

EMC conformance classification

This equipment is subject to FCC rules and shall comply with the
appropriate FCC rules before final delivery to the buyer or centers of
distribution.

US: FCC CFR, Title 47, Part 15, EMI Class A

EEA, Turkey: EU Council Directive 2004/108/EC, EMI Class A

Japan: VCCI Council, EMI Class A

Korea: KCC, EMI Class A

China (PRC): CPCS, EMI Class A

Taiwan: Taiwan BSMI, EMI Class A

Australia\New Zealand: ACMA, EMI Class A

Canada: ICES-003, EMI Class A

Russia: GOST R, EMI Class A

Saudi Arabia: MoCI, EMI Class A

Vietnam: MPT, EMI Class A

Homologation -- Telecom Type Approval

Homologation approval for specific countries has been initiated
with the IBM Homologation and Type Approval (HT&A) organization in
LaGaude, France.

The Power E880 system nodes or system control unit or PCIe Gen3
Expansion units are not certified for connection to interfaces of
public telecommunications networks. Certification may be required by
law prior to making any such connection. Contact an IBM
representative or reseller for any questions and for information on
PCIe adapters that can be used in the system and which are certified.

Product safety/Country testing/Certification

UL 60950-1:2007 Underwriters Laboratory

CAN/CSA22.2 No. 60950-1-07

EN60950-1:2006 European Norm

IEC 60950-1 2nd Edition + all National Differences

General requirements

The product is in compliance with IBM Corporate Bulletin C-B
0-2594-000 Statement of Conformity of IBM Product to External Standard
(Suppliers Declaration).

Systems

Product category: D.

Weighted Teraflops (WT) is based on the number of floating point
operations the processor can perform in a cycle.

The WT for a Power E880 0/32 core, 4.365 GHz, single-core
processor is 0.01089 WT.

Note:

Multiply 0.01089 WT by the number of 4.365 GHz processor cores to
find the system WT number.

For example, a 4.365 GHz 64-core system is 0.01089 x 64 = 0.697
WT.

Homologation

The Power E880 system nodes or system control unit or PCIe Gen3
Expansion units are not certified for direct connection to interfaces
of public telecommunications networks. Certification may be required
by law prior to making any such connection. Contact an IBM
representative or reseller for any questions and for information on
PCIe adapters which can be used in the system and which are
certified.

The first (in plug order) 16 memory DIMM slots of each system node
must always be populated to ensure each SCM has at least one memory
feature. Using the same same memory capacity feature (equal memory
CDIMM sizes) can typically provide optimal memory performance, but is
not required.

For same-serial-number upgrades, all features that are not
supported on the 9119-MHE must be removed.

Static memory activations can be converted during a D-model,
same-serial-number upgrade, and mobile activations can be moved to POWER8
servers in a Power Systems Enterprise Pool.

The system node or system control unit or PCIe3 I/O drawer does
not provide a serial port to which a UPS communication cable can be
attached for IBM i such as is available on smaller Power servers.
Clients should use standard data center electrical power backup
options used by many larger data centers.

The PCIe Gen3 I/O Expansion Drawer has a few adapter plugging
considerations

The following adapters are only supported in slot c6 of either
fanout module (max 2 adapters per drawer)

4-Port Async EIA-232 PCIe Adapter (#5785)

PCIe 2-Line WAN w/Modem (#2893/2894)

POWER GXT145 PCI Express Graphics Accelerator (#5748)

The following 4-port SAS adapters are not supported in slots C2 or
C5 of either fanout module (max 8 adapters per drawer).

PCIe3 RAID SAS Adapter Quad-port 6Gb x8 (#EJ0J)

PCIe3 SAS Tape/DVD Adapter Quad-port 6Gb x8 (#EJ10)

PCIe3 12GB Cache RAID SAS Adapter Quad-port 6Gb x8 (#EJ0L)

The following hardware features supported on the 9117-MMD and
9179-MHD are not supported on the 9119-MHE. Newer technology must be
used.

Additional optional features can be added, as desired. IBM i
systems require a DVD to be available to the system. This DVD can be
located in the system control unit (DVD feature #EU13) or it can be
located externally in an enclosure like the 7226-1U3. A USB PCIe
adapter such as #EC45 is required for #EU13. A SAS PCIe adapter such
as #EJ11 is required to attach a SATA DVD in the 7226

Feature-coded
racks are allowed for I/O expansion only.

A machine type/model rack, if desired, should be ordered as the
primary rack.

A minimum number of eight processor activations must be ordered
per system.

A minimum of four memory features per system node are required.

At least 50% of available memory must be activated through a
combination of feature EMA5, EMA6, and EMA9.

Power Integrated Facility for Linux Package feature (#ELJG) is a
chargeable feature code which will cause IBM configuration tools
to add the proper quantity of no-charge features associated with
processor activations (ELJ6), and memory activations (ELJH). This
feature delivers a quantity of 4 processor activations, a quantity
of 32 1GB memory activations and a quantity of 4 PowerVM for Linux
licenses for 4 processor cores. When the number of #ELJG
features ordered are equal to defined system minimum order
requirement for static processor and memory activations, the
activation requirement is satisfied.

Memory sizes can differ across the four SCMs of the system node,
but the eight CDIMM slots connected to the same SCM must be filled
with identical CDIMMs (one or two identical memory features per SCM).

If SAN Load Source Specify (#0837) is ordered, features 0040,
0041, 0043, 0047, 0308 are not supported.

The language group is auto-selected based on geographic rules.

No feature codes are assigned for the following:

Four ac power supplies are delivered as part of the system node.
No features are assigned to power supplies. Four line cords are
auto-selected according to geographic rules.

Two default ac PDU to wall cables are included. No features are
assigned. Cables are auto-selected according to geographic rules.

There must be one system control unit on each system. The system
control unit is considered the system with the system serial number.

One HMC is required for every 9119-MHE (physical hardware HMC or a
virtual appliance (vHMC)); however, a communal HMC is acceptable. HMCs
supported on POWER8 hardware are 7042-CR5 through 7042-CR9.

A single 7014-T42 rack can be ordered containing one or two Power
E880 systems. Note that two four-node Power E880 will not fit in a
single 42U rack and that care must be taken anytime ordering a pair of
Power 880 servers in a rack to ensure there is space for horizontal
PDUs, future system node additions and anything else the rack needs to
contain. A customer can order one factory-integrated system node and
later order an additional system node to be installed in the same rack
at the client establishment. IBM factory-integrated orders of the
7014-T42 can be with standard rack doors (#ERG7 or #6069) or OEM
doors. On initial order, if the rack contains a system node, the
7014-T42 machine/type must* be ordered. If a MES same-serial-number
upgrade the #0553 feature must* be ordered. ( * See also deracking
feature #ER21.)

Only horizontal PDUs are used by IBM Manufacturing in racks hosting
the 9119-MHE systems. Each PDU will occupy one EIA to aid cable
routing.

If attaching an HMC to a new server or adding function to an
existing server that requires a firmware update, the HMC machine code
may need to be updated. To determine the HMC machine code level
required for the firmware level on any server, go to the following
website to access the Fix Level Recommendation Tool (FLRT) on or after
the planned availability date for this product. FLRT will identify
the correct HMC machine code for the selected system firmware level

If a single HMC is attached to multiple servers, the HMC machine
code level must be updated to the server with the most recent firmware
level. All prior levels of server firmware are supported with the
latest HMC machine code level.

An HMC is required to manage POWER8 processor-based system nodes
implementing partitioning. Multiple POWER8 processor-based system
nodes can be supported by a single HMC.

If an HMC is used to manage any POWER8 processor-based system node,
the HMC must be 7042 -- CR5 through CR8 rack models or 7042 -- CR8
deskside model, or later.

When PowerVC is enabled, 4 GB of RAM is recommended. HMC 7042-CR5
ships with a default of 2 GB RAM.

SUSE Linux Enterprise Server 11 Service Pack 3 and later Service
Packs

Red Hat Enterprise Linux 7 or later

Red Hat Enterprise Linux 7.1 (le) or later

SUSE Linux Enterprise Server 12 or later

Ubuntu Server 14.04.3 or later

If installing VIOS:

VIOS 2.2.3.4 with interim fix IV63331, or later

Java is supported on POWER8 servers. For best exploitation of the
outstanding performance capabilities and most recent improvements of
POWER8 technology, IBM recommends upgrading Java-based applications to
Java 7 or Java 6, whenever possible.

The IBM Systems Information Center provides you with a single
information center where you can access product documentation for IBM
systems hardware, operating systems, and server software. Through a
consistent framework, you can efficiently find information and
personalize your access. Visit the IBM Systems Information Center, at

The following feature availability matrix for MT 9119 uses the
letter "A" to indicate features that are available and orderable on
the specified models. "S" indicates a feature that is supported on
the new model during a model conversion; these features will work on
the new model, but additional quantities of these features cannot be
ordered on the new model; they can only be removed. "N" indicates
that the feature is not supported on the new model and must be removed
during the model conversion. As additional features are announced,
supported, or withdrawn, this list will be updated. Contact your
Marketing Representative for additional information.

The following is a list of all feature codes in numeric order for
the IBM Power Systems 9119 machine type.

Attributes, as defined in the following feature descriptions, state
the interaction of requirements among features.

Minimums and maximums are the absolute limits for a single feature
without regard to interaction with other features. The maximum valid
quantity for MES orders may be different than for initial orders. The
maximums listed below refer to the largest quantity of these two
possibilities.

The order type defines if a feature is orderable only on initial
orders, only on MES orders, on both initial and MES orders, or if a
feature is supported on a model due to a model conversion. Supported
features cannot be ordered on the converted model, only left on or
removed from the converted model.

(#0010) - One CSC Billing Unit

One Billing Unit used by the Customer Solution Center.

Attributes provided: One CSC Billing Unit

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0011) - Ten CSC Billing Units

Ten Billing Units used by the Customer Solutions Center.

Attributes provided: Ten CSC Billing Units

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0040) - Mirrored System Disk Level, Specify Code

This code indicates the level of disk protection desired and helps
ensure that adequate hardware is in the final configuration.

Attributes provided: Device-level mirrored protection

Attributes required: Minimum of two (2) disk units

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#0041) - Device Parity Protection-All, Specify Code

This code indicates the level of disk protection desired and helps
ensure that adequate hardware is in the final configuration.

Attributes provided: RAID Data Protection

Attributes required: RAID-capable disk unit controller

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#0043) - Mirrored System Bus Level, Specify Code

This code indicates the level of disk protection desired and helps
ensure that adequate hardware is in the final configuration.

For new systems: The marketing configurator will error
on an order if sufficient disk units, and expansion units are not
included on the order to support bus-level mirrored
protection for all disk units. New, preloaded systems will be shipped
with bus-level mirroring enabled.

Attributes provided: Bus-level mirrored protection

Attributes required: Minimum of 2 (two) disk units

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: N/A

Return parts MES: Does not apply

(#0047) - Device Parity RAID-6 All, Specify Code

This code indicates the level of disk protection desired and helps
ensure that adequate hardware is in the final configuration.

Attributes provided: RAID-6 Data Protection

Attributes required: RAID-6 capable disk unit controller

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#0098) - Special Manufacturing Operations Indicator

This feature is a no-charge feature that indicates special
manufacturing operations are to be performed.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 999 (Initial order maximum: 250)

OS level required: n/a

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#0205) - RISC-to-RISC Data Migration

#0205 is used on initial orders to designate that the new server
should only receive a partial load of IBM i in IBM Mfg.

When #0205 is on the order, manufacturing will only load SLIC and
up through QSYS of IBM i. Ensure that enough storage is ordered to
contain the additional OS code that will be loaded following
installation of the system at the Client location. Specify code #0205
is mutually exclusive with #5000, SW Preload Required

The migration process requires that the installed model be at the
same version and release level of IBM i and other licensed programs as
the new server.

More information, and an updated IBM i Upgrade and Data Migration
Road Map (RISC-RISC) are available at

(#0265) - AIX Partition Specify

This feature indicates customers intend to create a partition on
the system that will use the AIX operating system. This feature
should be included once for each intended AIX partition. This feature
is an indicator and does not deliver parts, software, or services.

Attributes provided: None

Attributes required: Customers intend to create a partition on
the system that will run the AIX operating system.

Minimum required: 0

Maximum allowed: 4096 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0266) - Linux Partition Specify

This feature indicates customers intend to create a partition on
the system that will use the Linux operating system. This feature
should be included once for each intended Linux partition. This
feature is an indicator and does not deliver parts, software, or
services.

Attributes provided: None

Attributes required: Customers intend to create a partition on
the system that will run the Linux operating system.

Minimum required: 0

Maximum allowed: 4096 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0267) - IBM i Operating System Partition Specify

This feature indicates customers intend to create a partition on
the system that will use the IBM i operating system. This feature
should be included once for each intended IBM i partition. This
feature is an indicator and does not deliver parts, software, or
services.

Attributes provided: None

Attributes required: Customers intend to create a partition on
the system that will run the IBM i operating system.

Minimum required: 0

Maximum allowed: 4096 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0296) - Specify Custom Data Protection

#0296 specifies that a system has multiple IBM i partitions and
that data protection schemes should be considered separately for each
partition instead of only for an overall system level. Each
partition's data protection scheme can be different or the same.

Attributes provided: N/A

Attributes required: N/A

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

IBM i 7.1 TR9 or later

IBM i 7.2 TR1 or later

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0308) - Mirrored Level System Specify Code

This specify code indicates the level of disk protection desired
and helps ensure that adequate hardware is in the final configuration.

For new systems, the marketing configurator will show an error if
sufficient disk units and disk controllers are not included on the
order to support IOA-level mirroring protection. #0308 causes all
disk units to be placed into configurations capable of IOA-level
mirroring. Each disk unit and its mirrored pair must be on a
different disk controller.

Note that the load source disk unit in a new, preloaded system will
be device-level mirrored (same protection as provided with feature
#0040). This means that the load source is controlled by the first
disk unit controller on the first system bus, and will be mirrored
with a like disk unit that is also attached to the same first disk
controller on the first system bus.

For upgrade orders, #0308 will cause the marketing configurator to
show an error if sufficient disk units and disk controllers are not
available to provide the capability to enable IOA-level mirrored
protection for all DASD.

It is the client's responsibility to start mirroring on their
system.

Attributes provided: IOA level system mirroring

Attributes required: A minimum of two disk controllers and an
even number of disk units (with a minimum of four disk units on a
system).

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

IBM i 7.1 TR9 or later

IBM i 7.2 TR1 or later

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0347) - RAID Hot Spare Specify

#0347 is a specify code that indicates to IBM configuration tools
and to IBM Manufacturing that RAID-5 or RAID-6 disk arrays should be
further protected using the IBM i function of RAID hot spare. If
specified, IBM will ship a configuration which has at least one
stand-by disk drive for each disk controller in the system or designated
partition. The customer may alter the hot spare configuration
selecting different options once the system is installed.

Attributes provided: N/A

Attributes required: Existence of #0041 or #0047

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

IBM i 7.1 TR9 or later

IBM i 7.2 TR1 or later

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0348) - V.24/EIA232 6.1m (20-Ft) PCI Cable

This feature provides a 20-foot WAN PCI cable that supports a V.24
or a EIA232 electrical connection interface.

(#0373) - UPS Factory Integration Specify

Used in manufacturing to ensure that a UPS ordered from IBM under a
separate machine type/model, is associated with the system order, and
is shipped concurrently.

Attributes provided: N/A

Attributes required: Ordered machine type-model UPS.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

Note: Available only when ordered with 19 inch feature code rack on an
MES order. Max. is 36 with rack feature 0551.

(#0374) - HMC Factory Integration Specify

Used in manufacturing to ensure that an HMC ordered from IBM under
a separate machine type/model is associated with the system order and
is shipped concurrently.

Attributes provided: N/A

Attributes required: Ordered machine type-model HMC.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

Note: Available only when ordered with 19 inch feature code rack on an
MES order. Max. is 36 with rack feature 0551.

(#0375) - Display Factory Integration Specify

Used in manufacturing to ensure that a display ordered from IBM
under a separate machine type/model, is associated with the system
order, and is shipped concurrently.

Attributes provided: N/A

Attributes required: Ordered machine type-model display.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

Note: Available only when ordered with 19 inch feature code rack on an
MES order. Max. is 36 with rack feature 0551.

(#0376) - Reserve Rack Space for UPS

Used in manufacturing to reserve 1 EIA of rack space in the bottom
of the rack for later client installation of a UPS.

Attributes provided: 1 EIA rack space reserved.

Attributes required: Ordered rack feature.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

Note: Available only when ordered with 19 inch feature code rack on an
MES order. Max. is 36 with rack feature 0551.

(#0377) - Reserve Rack Space for HMC

Used in manufacturing to reserve 1 EIA of rack space in the middle
of the rack for later client installation of a rack-mounted HMC.

Attributes provided: 1 EIA rack space reserved.

Attributes required: Ordered rack feature.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

Note: Available only when ordered with 19 inch feature code rack on an
MES order. Max. is 36 with rack feature 0551.

(#0378) - Reserve Rack Space for Display

Used in manufacturing to reserve 1 EIA of rack space in the middle
of the rack for later client installation of an HMC rack-mounted
display such as the 7316.

Attributes provided: 1 EIA rack space reserved.

Attributes required: Ordered rack feature.

Minimum required: 0

Maximum allowed: 42 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

(#0397) - MMA/MMB/MHB upgrade indicator

(No Longer Available as of January 24, 2017)

This feature indicates that this 9117-MMA system was created as a
model conversion from a 9406-MMA or the 9117-MMB/9179-MHB was created
as a model conversion from a 9117-MMA. This feature may only be added
to MES orders when placing an order for one of these model
conversions.

Attributes provided: MMA function carryover indicator

Attributes required: Model conversion.

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

Note: When migrating a 9117-MMA with an existing #0397 feature, note
that PowerVM needs to be present on the server in order to enable
micro-partitioning.

(#0465) - SSD Placement Indicator - 5887, EL1S

This is an IBM internal automatic generated SSD specify indicator
for placement and it is not selectable.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#0551) - 19 inch, 1.8 meter high rack

Provides a 19 inch, 1.8 meter high rack with 36 EIA units of total
space for installing rack mounted CECs and/or expansion units. One of
the following features are required on the #0551:

#6068 - Optional Front Door for 1.8m Rack

#6263 - 1.8m Rack Trim Kit

#6248 - 1.8m Rack Acoustic Doors

The following features are also orderable on the #0551:

#0599 - Rack Filler Panel Kit

#6580 - Optional Rack Security Kit

#7840 - Side-By-Side for 1.8m Racks

#7841 - Ruggedize Rack Kit

The #0551 can support up to eight PDUs, four mounted vertically and
four mounted horizontally. Each PDU mounted horizontally takes up 1
EIA of rack space. The following PDUs are supported:

#7188 - Power Distribution Unit (12, C-13 sockets)

#7109 - Power Distribution Unit (12, C-13 sockets)

Attributes provided: 19 inch, 1.8M, 36 EIA Rack

Attributes required: #6068 or #6246 or #6248 or #6263.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#0553) - 19 inch, 2.0 meter high rack

Provides a 19-inch, 2.0 meter high rack with 42 EIA units of total
space for installing rack mounted CECs and/or expansion units. One of
the following features are required on the #0553:

#6069 - Optional Front Door for 2.0m Rack

#6247 - 2.0m Rack Trim Kit (w/d 5/28/2010)

#6272 - 2.0m Rack Trim Kit

#6249 - 2.0m Rack Acoustic Doors

The #0553 can support up to nine power distribution units (PDU),
four mounted vertically and five mounted horizontally. Each PDU
mounted horizontally takes up 1 EIA of rack space. The following PDUs
are supported:

#7188 - Power Distribution Unit (12, C13 sockets)

#7109 - Power Distribution Unit (12, C13 sockets)

The manufacturing practice and recommended configuration of the rack
is:

Reserve 2U Rack Space at Bottom of Rack

Mount PDUs horizontally using 1U (1 EIA) for 9119-MME/MHE

Attributes provided: 19 inch, 2.0M, 42 EIA Rack

Attributes required: #6069 or #6247 or #6272 or #6249

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#0567) - IBM i 7.1 Specify Code

This feature is used to indicate the correct level of code when IBM
i is specified.

This specify code indicates that a #1947 Disk Unit is being used as
the Load Source.

Attributes provided: None

Attributes required: Feature #1947

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0872) - #1948 Load Source Specify (283GB 15k RPM SAS SFF-2 Disk)

This specify code indicates that a #1948 Disk Unit is being used as
the Load Source.

Attributes provided: None

Attributes required: Feature #1948

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#0874) - #1956 Load Source Specify (283GB 10k RPM SAS SFF-2 Disk)

This specify code indicates that a #1956 Disk Unit is being used as
the Load Source.

Attributes provided: None

Attributes required: Feature #1956

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0875) - #1962 Load Source Specify (571GB 10k RPM SAS SFF-2 Disk)

This specify code indicates that a #1962 Disk Unit is being used as
the Load Source.

Attributes provided: None

Attributes required: Feature #1962

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0880) - #1738 Load Source Specify (856GB 10k RPM SAS SFF-2 Disk)

(No Longer Available as of January 20, 2017)

This specify code indicates that a #1738 Disk Unit is being used as
the Load Source.

Attributes provided: None

Attributes required: Feature #1738

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#0894) - #ES0D Load Source Specify (387GB SFF-2 SSD for IBM i)

This specify code indicates that a #ES0D Solid State Drive is being
used as the Load Source.

Attributes provided: None

Attributes required: Feature #ES0D

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#0911) - #ESD2 Load Source Specify (1.1TB 10k SFF-2)

Indicates that a #ESD2 (1.1 TB 10k rpm SFF-2 disk drive) is being
used as the Load Source.

Attributes provided: None

Attributes required: Feature #ESD2

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - not supported

IBM i 7.1 TR9 and 7.2 TR1 or later

Linux - not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1019) - Modem Cable - Australia

Australia modem cable, use with #2894, 6809 and 6834 or similar
modem adapters. Maximum of two per adapter.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1020) - Modem Cable - HK/NZ

China(Hong Kong S.A.R.)/New Zealand modem cable, use with #2894,
6809 and 6834 or similar modem adapters for New Zealand and with
#2893, 6808 and 6833 or similar modem adapters for Hong Kong. Maximum
of two per adapter.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1025) - Modem Cable - US/Canada and General Use

Modem cable, use with #2893, 6808 and 6833 or similar modem
adapters. Maximum of two per adapter. Select this cable for use with
your modem if there is not another cable feature that is identified as
specific to your country.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1106) - USB 160 GB Removable Disk Drive

Provides a RDX disk drive in a rugged cartridge to be used in an
RDX docking station such as #EU03, #EU04, #EU23, #1123, #1103, #1104
or #EU07. 160 GB is uncompressed.Compression/decompression is
provided by the operating system, not the drive itself. Feature 1106
is not entitled under the IBM Maintenance Agreement, if one is
purchased.

Attributes provided: 160 GB RDX rugged disk/cartridge

Attributes required: RDX docking station

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: See RDX Docking Station

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#1107) - USB 500 GB Removable Disk Drive

Provides an RDX disk drive in a rugged cartridge to be used in an
RDX Internal and External docking station such as the #1103, #1104,
#1123, #EU03, #EU04, #EU23 or #EU07. 500 GB is uncompressed. With
typical 2X compression, capacity would be 1000 GB. Compression/
decompression is provided by the operating system, not the drive
itself. Feature 1107 is not entitled under the IBM Maintenance
Agreement, if one is purchased.

Attributes provided: 500 GB RDX rugged disk/cartridge

Attributes required: None.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: See RDX Docking Station

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1120) - Decline Electronic Service Agent Install Indicator

Electronic Service Agent is a basic component of the Power Systems
server and is an important tool in maintaining the high availability
expected from your IBM system and the service response time you expect
from your IBM service team. You should select this feature only if
you do not intend to have the IBM SSR install ESA at the time your
system is being installed. Before selecting this feature,
discuss this decision with your IBM Account Team or IBM Business
Partner.

Electronic Service Agent installation by an IBM SSR is available to
you for no additional charge during initial server installation only.

Attributes provided: Documents Customers intention to not allow
the IBM SSR to install Electronic Service Agent during system
installation

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#1140) - Custom Service Specify, Rochester, MN, USA

Having #1140 on the order, will cause the order to be routed to
Rochester and the machine to be internally routed to the CSC build
area in Building 114 (Rochester).

The drive can be reformatted to 528 byte sectors and then used by
IBM i/AIX/Linux/VIOS

Limitation: Cannot be used in #5802/5803 I/O drawers or in CEC
SFF bays due to physical difference in Gen1 and Gen2 carriers.

Attributes provided: 900GB of Disk Storage mounted in a Gen2
carrier

Attributes required: one SFF (SFF-2) SAS GEN2 drive bay

Minimum required: 0

Maximum allowed: 4032 (Initial order maximum: 250)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1817) - Quantity 150 of #1962

This feature ships a quantity of 150 #1962 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1962

Attributes required: See feature #1962

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1818) - Quantity 150 of #1964

This feature ships a quantity of 150 #1964 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1964

Attributes required: See feature #1964

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1844) - Quantity 150 of #1956

(No Longer Available as of July 8, 2016)

This feature ships a quantity of 150 #1956 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1956

Attributes required: See feature #1956

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1866) - Quantity 150 of #1917

(No Longer Available as of July 8, 2016)

This feature ships a quantity of 150 #1917 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1917

Attributes required: See feature #1917

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1868) - Quantity 150 of #1947

(No Longer Available as of July 8, 2016)

This feature ships a quantity of 150 #1947 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1947

Attributes required: See feature #1947

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1869) - Quantity 150 of #1925

(No Longer Available as of July 8, 2016)

This feature ships a quantity of 150 #1925 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Limitation: Cannot be used in #5802/5803 I/O drawers or in CEC
due to physical difference in Gen1 and Gen2 carriers.

Attributes provided: 300GB SFF disk drive mounted in a carrier

Attributes required: One SFF Gen-2 bay

Minimum required: 0

Maximum allowed: 4032 (Initial order maximum: 250)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1927) - Quantity 150 of #1948

This feature ships a quantity of 150 #1948 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1948

Attributes required: See feature #1948

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1929) - Quantity 150 of #1953

This feature ships a quantity of 150 #1953 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

(#1953) - 300GB 15k RPM SAS SFF-2 Disk Drive (AIX/Linux)

300 GB SFF 15k RPM SAS disk drive mounted in a Gen-2 carrier and
supported in SFF SAS bay such as used in the #5887 EXP24S I/O drawer
or #ESLS EXP24SX I/O drawer. Disk can be formatted for either 512
bytes or 528 byte sectors and is thus called having 5xx drive sectors.
528 byte sectors provide additional protection. IBM Manufacturing
will ship pre-formatted with 528 or with 512 byte sectors. Selection
of the formatting is selected by IBM Manufacturing based on
manufacturing rules, but the client may change at their location. In
the EPX24SX IBM Manufacturing will ship using 528 byte sectors.
Capacity is 300 GB with 512 byte formatting and is 283 GB with 528
byte sector. CCIN is 19B1.

Limitations: physical difference in carriers prevent this drive
from being used in SFF-1 bays such as used in the #5802/5803 I/O
drawer or in SFF-3 bays such as used in POWER8 system units.

(#1962) - 571GB 10k RPM SAS SFF-2 Disk Drive (IBM i)

Limitation: Cannot be used in #5802/5803 I/O drawers or in CEC
bays due to physical difference in Gen1 and Gen2 carriers.

Attributes provided: 571GB of Disk Storage mounted in a Gen2
carrier

Attributes required: one SFF SAS disk drive bay

Minimum required: 0

Maximum allowed: 4032 (Initial order maximum: 250)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#1964) - 600GB 10k RPM SAS SFF-2 Disk Drive (AIX/Linux)

600 GB SFF 10k RPM SAS disk drive mounted in a Gen-2 carrier and
supported in SFF SAS bay such as used in the #5887 EXP24S I/O drawer
or #ESLS EXP24SX I/O drawer. Disk can be formatted for either 512
bytes or 528 byte sectors and is thus called having 5xx drive sectors.
528 byte sectors provide additional protection. IBM Manufacturing
will ship pre-formatted with 528 or with 512 byte sectors. Selection
of the formatting is selected by IBM Manufacturing based on
manufacturing rules, but the client may change at their location. In
the EPX24SX IBM Manufacturing will ship using 528 byte sectors.
Capacity is 600 GB with 512 byte formatting and is 571 GB with 528
byte sector.

Limitations: physical difference in carriers prevent this drive
from being used in SFF-1 bays such as used in the #5802/5803 I/O
drawer or in SFF-3 bays such as used in POWER8 system units.

Multiple #0367 cables can be ordered but only one per #2893) to
serve as consoles for secondary partitions when Logical Partitioning
is utilized. ECS is supported from both the modem port, and the RVX
port.

The following cable is required to support ECS from the RVX port:

#0348 - V.24/EIA232 20-Ft PCI Cable

The #2893 does not support the remote ring indicate function.

Attributes provided: One PCIe slot

Attributes required: Modem

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 0)

OS level required:

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#2894) - PCIe 2-Line WAN w/Modem CIM

The #2894 is a 2-line/port WAN w/modem PCIe adapter. This feature
is the CIM (Complex Impedance Matching) version which is offered only
in Australia and New Zealand.

Multiple #0367 cables can be ordered but only one per #2894) to
serve as consoles for secondary partitions when Logical Partitioning
is utilized. ECS is supported from both the modem port, and the RVX
port. The following cable is required to support ECS from the RVX
port:

#0348 - V.24/EIA232 20-Ft PCI Cable

The #2894 does not support the remote ring indicate function.

Attributes provided: One PCIe slot

Attributes required: Modem

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 0)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#2934) - 3M Asynchronous Terminal/Printer Cable EIA-232

The Asynchronous Printer/Terminal Cable is used for attaching
printers, plotters, and terminals that support the EIA-232 standard to
any asynchronous adapter. This cable is the equivalent of the
combination of FC 2936 (modem cable) and FC 2937 (printer/terminal
interposer) and replaces this method of printer/terminal attachment.

This cable is 3m (9.8 feet) long, uses DB25 connectors and is
supported on all RS/6000 systems using any asynchronous ports.

Attributes provided: EIA232 device attachment capability

Attributes required: Any Asynchronous port

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#2936) - Asynchronous Cable EIA-232/V.24 3M

Used to attach a modem to the standard I/O ports with the 10-pin to
25-pin converter cable (#3925), 8-port Cable Assembly, 16-Port Cable
Assembly. The cable is 3 meters (9.8 feet) in length.

Attributes provided: Modem attachment to async or serial port

Attributes required: Async or serial port

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3124) - Serial-to-Serial Port Cable for Drawer/Drawer- 3.7M

This 3.7 meter cable is available to provide a null-modem
connection between the serial ports of two system drawers that are
mounted within the same rack. The cable provides a DB25 female
connector at each end.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3125) - Serial-to-Serial Port Cable for Rack/Rack- 8M

This 8 meter cable is available to provide a null-modem connection
between the serial ports of two system drawers that are mounted in
separate racks. The cable provides a DB25 female connector at each
end.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3287) - 1m, (3.3-ft) IB 40G Copper Cable QSFP/QSFP

This 1 meter (3.3 foot) copper cable provides a high bandwidth
connection up to 40 Gb/s (Quad Data Rate). Both ends of the cable
have a QFSP connector. The cable connectors can be described as
QSFP+/QSFP+ as well as QSFP/QSFP. Usage examples include a)
connecting a PCIe2 InfiniBand QDR Adapter to a QDR IB switch or b)
connecting two IBM Rack Switches together using QSFP+ ports.

(#3288) - 3m, (9.8-ft.) IB 40G Copper Cable QSFP/QSFP

This 3 meter (9.8 foot) copper cable provides a high bandwidth
connection up to 40 Gb/s (Quad Data Rate). Both ends of the cable
have a QFSP connector. The cable connectors can be described as
QSFP+/QSFP+ as well as QSFP/QSFP. Usage examples include a)
connecting a PCIe2 InfiniBand QDR Adapter to a QDR IB switch or b)
connecting two IBM Rack Switches together using QSFP+ ports.

(#3290) - 10 meter Quad Data Rate InfiniBand Optical Cable, QSFP/QSFP

This 10 meter Quad Data Rate (40Gb/s) InfiniBand optical cable is
used to connect the QDR Host Channel Adapter to QDR InfiniBand
Switches. It can also be used for QDR InfiniBand switch-to-switch
applications.

(#3293) - 30 meter Quad Data Rate InfiniBand Optical Cable, QSFP/QSFP

This 30 meter Quad Data Rate (40Gb/s) InfiniBand optical cable is
used to connect the QDR Host Channel Adapter to QDR InfiniBand
Switches. It can also be used for QDR InfiniBand switch-to-switch
applications.

(#3450) - SAS YO Cable 1.5m - HD 6Gb Adapter to Enclosure

This 1.5 meter SAS cable connects a SAS adapter to a SAS I/O
enclosure. This YO cable has three connectors, one Mini-SAS HD (High
Density)connector and two Mini SAS connectors. The Mini-SAS HD
connector attaches to a SAS adapter such as the #5913 1.8GB RAID SAS
Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5887 EXP24S SAS I/O drawer. This cable can
support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#3451) - SAS YO Cable 3m - HD 6Gb Adapter to Enclosure

This 6 meter SAS cable connects a SAS adapter to a SAS I/O
enclosure. This YO cable has three connectors, one Mini-SAS HD (High
Density)connector and two Mini SAS connectors. The Mini-SAS HD
connector attaches to a SAS adapter such as the #5913 1.8GB RAID SAS
Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5887 EXP24S SAS I/O drawer. This cable can
support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#3452) - SAS YO Cable 6m - HD 6Gb Adapter to Enclosure

This 6 meter SAS cable connects a SAS adapter to a SAS I/O
enclosure. This YO cable has three connectors, one Mini-SAS HD (High
Density)connector and two Mini SAS connectors. The Mini-SAS HD
connector attaches to a SAS adapter such as the #5913 1.8GB RAID SAS
Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5887 EXP24S SAS I/O drawer. This cable can
support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#3453) - SAS YO Cable 10m - HD 6Gb Adapter to Enclosure

This 10 meter SAS cable connects a SAS adapter to a SAS I/O
enclosure. This YO cable has three connectors, one Mini-SAS HD (High
Density)connector and two Mini SAS connectors. The Mini-SAS HD
connector attaches to a SAS adapter such as the #5913 1.8GB RAID SAS
Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5887 EXP24S SAS I/O drawer. This cable can
support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#3454) - SAS X Cable 3m - HD 6Gb 2-Adapter to Enclosure

This 3 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four connectors, two Mini-SAS HD (High
Density) connector and two Mini SAS connectors. The Mini-SAS HD
connectors attaches to two SAS adapters such as the #5913 1.8GB RAID
SAS Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5886 EXP12S or #5887 EXP24S SAS I/O drawer.
This cable can support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapters can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Attributes provided: connection between two SAS adapter with
Mini-SAS HD connectors and a SAS I/O drawer with Mini SAS connectors

(#3455) - SAS X Cable 6m - HD 6Gb 2-Adapter to Enclosure

This 6 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four connectors, two Mini-SAS HD (High
Density) connector and two Mini SAS connectors. The Mini-SAS HD
connectors attaches to two SAS adapters such as the #5913 1.8GB RAID
SAS Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5886 EXP12S or #5887 EXP24S SAS I/O drawer.
This cable can support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapters can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Attributes provided: connection between two SAS adapter with
Mini-SAS HD connectors and a SAS I/O drawer with Mini SAS connectors

(#3456) - SAS X Cable 10m - HD 6Gb 2-Adapter to Enclosure

This 10 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four connectors, two Mini-SAS HD (High
Density) connector and two Mini SAS connectors. The Mini-SAS HD
connectors attaches to two SAS adapters such as the #5913 1.8GB RAID
SAS Adapter. Both Mini SAS connectors attach to the same I/O drawer
(enclosure) such as a #5886 EXP12S or #5887 EXP24S SAS I/O drawer.
This cable can support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapters can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Attributes provided: connection between two SAS adapter with
Mini-SAS HD connectors and a SAS I/O drawer with Mini SAS connectors

(#3457) - SAS YO Cable 15m - HD 3Gb Adapter to Enclosure

This 15 meter SAS cable connects a SAS adapter to a SAS I/O
enclosure. This YO cable has three connectors, one Mini-SAS HD (High
Density)connector and two Mini SAS connectors. The Mini-SAS HD
connector attaches to a SAS adapter such as the #5913 1.8GB RAID SAS
Adapter or #ESA1/ESA2 PCIe2 RAID SAS Adapter. Both Mini SAS
connectors attach to the same I/O drawer (enclosure) such as a #5887
EXP24S SAS I/O drawer. This cable can support up to 3Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#3458) - SAS X Cable 15m - HD 3Gb 2-Adapter to Enclosure

This 15 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four connectors, two Mini-SAS HD (High
Density) connector and two Mini SAS connectors. The Mini-SAS HD
connectors attaches to two SAS adapters such as the #5913 1.8GB RAID
SAS Adapter or #ESA1/ESA2 PCIe2 RAID SAS Adapter. Both Mini SAS
connectors attach to the same I/O drawer (enclosure) such as a #5886
EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support up to
3Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapters can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Attributes provided: connection between two SAS adapter with
Mini-SAS HD connectors and a SAS I/O drawer with Mini SAS connectors

This SAS cable (X) connects a SAS disk drawer to two SAS controller
adapters. This cable supports dual controller/dual path attach
between two SAS controller adapters and the SAS disk drawer. The SAS
controller adapters can be in the same or in different host systems.
This cable has four Mini SAS 4x plug connectors. Two of the Mini SAS
4x plug connectors attach to the adapters and are keyed as END
DEVICES. Two of the Mini SAS 4x plug connectors attach to the SAS
disk drawer and are keyed for ENCLOSURE DOWN Arrow. All of the
connectors are wired in 2x mode. Follow the directions on the
connector labels when attaching the connectors on this cable. This
cable is 3 meters long, choose the SAS (X) cable length that matches
the distance between the adapters and the SAS disk drawer. The
adapter legs of this cable are each 2.5 meters long.

Attributes provided: Connection between two SAS controller
adapters and a SAS disk drawer

This SAS cable (X) connects a SAS disk drawer to two SAS controller
adapters. This cable supports dual controller/dual path attach
between two SAS controller adapters and the SAS disk drawer. The SAS
controller adapters can be in the same or in different host systems.
This cable has four Mini SAS 4x plug connectors. Two of the Mini SAS
4x plug connectors attach to the adapters and are keyed as END
DEVICES. Two of the Mini SAS 4x plug connectors attach to the SAS
disk drawer and are keyed for ENCLOSURE DOWN Arrow. All of the
connectors are wired in 2x mode. Follow the directions on the
connector labels when attaching the connectors on this cable. This
cable is 6 meters long, choose the SAS (X) cable length that matches
the distance between the adapters and the SAS disk drawer. The
adapter legs of this cable are each 5.5 meters long.

Attributes provided: Connection between two SAS controller
adapters and a SAS disk drawer

This SAS cable (X) connects a SAS disk drawer to two SAS controller
adapters. This cable supports dual controller/dual path attach
between two SAS controller adapters and the SAS disk drawer. The SAS
controller adapters can be in the same or in different host systems.
This cable has four Mini SAS 4x plug connectors. Two of the Mini SAS
4x plug connectors attach to the adapters and are keyed as END
DEVICES. Two of the Mini SAS 4x plug connectors attach to the SAS
disk drawer and are keyed for ENCLOSURE DOWN Arrow. All of the
connectors are wired in 2x mode. Follow the directions on the
connector labels when attaching the connectors on this cable. This
cable is 15 meters long, choose the SAS (X) cable length that matches
the distance between the adapters and the SAS disk drawer. The
adapter legs of this cable are each 14.5 meters long.

Attributes provided: Connection between two SAS controller
adapters and a SAS disk drawer

This SAS cable (YO) connects a remote SAS drawer to a SAS
controller adapter. This cable supports single controller/dual path
attach between the SAS controller adapter and the SAS disk drawer.
This cable has one Mini SAS 4X plug connector on the adapter end keyed
for an END DEVICE, wired in 4x mode and two Mini SAS 4X plug
connectors on the drawer end keyed for ENCLOSURE DOWN Arrow, both are
wired in 2x mode. Follow the directions on the connector labels when
attaching the connectors on this cable. This cable is 1.5 meters
long, choose the SAS (YO) cable length that matches the distance
between the adapter and the SAS disk drawer.

This SAS cable (YO) connects a remote SAS drawer to a SAS
controller adapter. This cable supports single controller/dual path
attach between the SAS controller adapter and the SAS disk drawer.
This cable has one Mini SAS 4X plug connector on the adapter end keyed
for an END DEVICE, wired in 4x mode and two Mini SAS 4X plug
connectors on the drawer end keyed for ENCLOSURE DOWN Arrow, both are
wired in 2x mode. Follow the directions on the connector labels when
attaching the connectors on this cable. This cable is 3 meters long,
choose the SAS (YO) cable length that matches the distance between the
adapter and the SAS disk drawer.

This SAS cable (YO) connects a remote SAS drawer to a SAS
controller adapter. This cable supports single controller/dual path
attach between the SAS controller adapter and the SAS disk drawer.
This cable has one Mini SAS 4X plug connector on the adapter end keyed
for an END DEVICE, wired in 4x mode and two Mini SAS 4X plug
connectors on the drawer end keyed for ENCLOSURE DOWN Arrow, both are
wired in 2x mode. Follow the directions on the connector labels when
attaching the connectors on this cable. This cable is 6 meters long,
choose the SAS (YO) cable length that matches the distance between the
adapter and the SAS disk drawer.

This SAS cable (YO) connects a remote SAS drawer to a SAS
controller adapter. This cable supports single controller/dual path
attach between the SAS controller adapter and the SAS disk drawer.
This cable has one Mini SAS 4X plug connector on the adapter end keyed
for an END DEVICE, wired in 4x mode and two Mini SAS 4X plug
connectors on the drawer end keyed for ENCLOSURE DOWN Arrow, both are
wired in 2x mode. Follow the directions on the connector labels when
attaching the connectors on this cable. This cable is 15 meters long,
choose the SAS (YO) cable length that matches the distance between the
adapter and the SAS disk drawer.

(#3925) - 0.3M Serial Port Converter Cable, 9-Pin to 25-Pin

This cable converts the 9-pin serial port on the system to a 25-pin
serial port which allows the user to attach 25-pin serial devices to
the system.

Attributes provided: 9-Pin to 25-Pin connectivity

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3927) - Serial Port Null Modem Cable, 9-pin to 9-pin, 3.7M

This 3.7 meter 9 pin to 9 pin Null modem Serial cable allows two
EIA-232 communications ports to exchange data with one another without
going through a modem.

Attributes provided: 9 pin female connector at each end of the
cable

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3928) - Serial Port Null Modem Cable, 9-pin to 9-pin, 10M

This 10 meter 9 pin to 9 pin Null Modem Serial cable allows two
EIA-232 communications ports to exchange data with one another without
going through a modem.

Attributes provided: 9 pin female connector at each end of the
cable

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#3930) - System Serial Port Converter Cable

This cable is used to connect the System port/UPS Conversion Cable
(#1827) to the CEC serial port. Also used to connect an ASCII
terminal or modem to the serial port which is physically an RJ45
connection on the Power 710/720/730/740 and Power S824/S822/S814/
S822L/S812L systems unit.

Attributes provided: Attachment of #1827 to CEC serial port;
attachment of ASCII terminal or modem to the serial port.

This cable is required to connect displays with a 15-pin "D" shell
connector to the appropriate accelerator connector when it is farther
away than the attached monitor cable can reach. Rack mounted systems
are likely candidates for this extender cable.

(#4256) - Extender Cable - USB Keyboards, 1.8M

This feature provides a 1.8M extension cable for use with USB
keyboards.

Attributes provided: 1.8M Extension Cable

Attributes required: USB Keyboard

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#4276) - VGA to DVI Connection Converter

This feature is a plug converter that will allow a Video device
with a 15 pin D-shell VGA cable plug (such as a KVM switch) to connect
to a graphics adapter with a 28 pin D-shell DVI receptacle connector.
This device has both a 28 pin D-Shell DVI plug and a 15 pin D-shell
VGA receptacle.

One and only one rack indicator feature is required onall orders (#4650 to #4666).

(#4650) - Rack Indicator- Not Factory Integrated

This indicator is used to specify that the rack mountable device in
this initial order should not be merged into a rack within IBM
Manufacturing.

Note: This "no additional charge" feature will be placed on an initial
order for a rack mountable device by the Configuration Tool when the
order does not include a 19" Rack.

A rack integration indicator is required on all 19" Rack mountable
device initial orders. One feature code from the group 4650 to 4666
must be listed on the order. More than one feature code from this
group is not allowed.

Attributes provided: System will not be shipped in a rack.

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4651) - Rack Indicator, Rack #1

When added to an initial rack order, this indicator is used to
specify the first rack for a multi rack order, or the only rack for a
single rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #1.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack Integration/ Rack Specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4652) - Rack Indicator, Rack #2

When added to an initial rack order, this indicator is used to
specify the second rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #2 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack Integration/Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4653) - Rack Indicator, Rack #3

When added to an initial rack order, this indicator is used to
specify the third rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #3 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4654) - Rack Indicator, Rack #4

When added to an initial rack order, this indicator is used to
specify the fourth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #4 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4655) - Rack Indicator, Rack #5

When added to an initial rack order, this indicator is used to
specify the fifth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #5 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4656) - Rack Indicator, Rack #6

When added to an initial rack order, this indicator is used to
specify the sixth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #6 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4657) - Rack Indicator, Rack #7

When added to an initial rack order, this indicator is used to
specify the seventh rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #7 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code
from the group 4650 to 4666 must be listed on the order. More than
one feature code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4658) - Rack Indicator, Rack #8

When added to an initial rack order, this indicator is used to
specify the eighth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #8 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4659) - Rack Indicator, Rack #9

When added to an initial rack order, this indicator is used to
specify the ninth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #9 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4660) - Rack Indicator, Rack #10

When added to an initial rack order, this indicator is used to
specify the tenth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #10 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4661) - Rack Indicator, Rack #11

When added to an initial rack order, this indicator is used to
specify the eleventh rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #11 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4662) - Rack Indicator, Rack #12

When added to an initial rack order, this indicator is used to
specify the twelfth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #12 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4663) - Rack Indicator, Rack #13

When added to an initial rack order, this indicator is used to
specify the thirteenth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #13 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4664) - Rack Indicator, Rack #14

When added to an initial rack order, this indicator is used to
specify the fourteenth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #14 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4665) - Rack Indicator, Rack #15

When added to an initial rack order, this indicator is used to
specify the fifteenth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #15 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg.is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4666) - Rack Indicator, Rack #16

When added to an initial rack order, this indicator is used to
specify the sixteenth rack for a multi rack order.

When added to an initial rack mountable device order, this
indicator is used to specify that the rack mountable device (such as a
system or I/O drawer) is to be mounted in rack #16 of a multi rack
order.

Note: For 19" rack mountable device orders: One feature code from the
group 4650 to 4666 must be listed on the order. More than one feature
code from this group is not allowed.

For 19" rack orders: If IBM Mfg. is to assemble a rack mountable
device into the rack, one feature code selection from the group 4651
to 4666 must be listed on the order. More than one feature code
selection from this group is not allowed. The quantity of this
selected feature code on the 19" rack order must equal the number of
rack mountable devices to be installed in the rack by IBM Mfg.

Attributes provided: Rack specify

Attributes required: Rack

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#4891) - CBU SPECIFY

This specify code indicates this system has been properly registered
as a Capacity BackUp system and has, through that registration been
authorized to temporarily receive IBM i Operating System License
Entitlements and 5250 Processor Enablement entitlements, from a
primary system under the conditions specified at the time the system
was registered. This feature is an indicator only, authorization to
use this system as a backup is obtained only by registering the system
with IBM on the CBU website at: www.ibm.com/systems/power/
hardware/cbu

Attributes provided: Indicates the system has been registered for
use as a CBU system for IBM i License entitlement purposes.

Attributes required: # 2145 Primary OS - IBM i

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#5000) - Software Preload Required

Indicates that preloaded software and/or consolidated I/O is
shipped with the initial order. A maximum of one (#5000) is
supported. This feature has country-specific usage.

Attributes provided: Software Pre-load

Attributes required: N/A

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

Linux not supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: No

(#5228) - PowerVM Enterprise Edition

This feature allows the customer to create partitions that are in
units of less than 1 CPU (sub-CPU LPARs) and allows the same system
I/O to be virtually allocated to these partitions. When PowerVM is
installed in the system, all activated processors must have the
PowerVM feature. A fully activated 4-core system requires that four
of this feature be ordered. An encrypted key is supplied to the
customer and is installed on the system, authorizing the partitioning
at the sub-processor level. PowerVM Enterprise Edition also includes
Live Partition Mobility, which allows for the movement of a logical
partition from one Power6, Power7, Power7+ or Power8 server to another
Power6, Power7, Power7+ or Power8 with no application downtime. Note:
If feature 5228 is ordered, the quantity ordered must be equal to the
number of active processors.

Attributes provided: Capability to partition processor

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5260) - PCIe2 LP 4-port 1GbE Adapter

This short PCIe Gen2 adapter provides four 1Gb Ethernet ports that
can be configured to run at 1000, 100 or 10 Mbps. 4-pair CAT-5
Unshielded Twisted Pair (UTP) cables up to 100 meters in length are
attached to the copper RJ45 connectors. Each port is independent of
one another and supports full-duplex or half-duplex. 1000 Mbps speed
is not supported in Half Duplex (HDX) mode.

Feature #5260 and #5899 are electronically identical and have the
same CCIN of 576F. #5260 indicates a low profile tail stock while
#5899 indicates a full high tail stock.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four-port 1 Gb Ethernet

Attributes required: 1 Low Profile (LP) PCIe slot (Gen1 or Gen2)

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5269) - PCIe LP POWER GXT145 Graphics Accelerator

This feature is the Low Profile (LP) equivalent of feature number
5748. The POWER GXT145 is a versatile, low-priced 2D graphics
accelerator. It can be configured to operate in either 8-bit or
24-bit color modes. This adapter supports both analog and digital
monitors. The adapter requires a PCI Express slot. If attaching a
device that requires a 15 pin D-Shell receptacle for a VGA connection
(eg. when the graphic adapter output is routed directly to a 7316-TF3
display or indirectly through a KVM switch), order a VGA to DVI
Connection Converter, feature number 4276 to accommodate the attaching
device. A short converter cable is included which adapts the 2nd
28-pin port on the adapter to 15-pin D-Shell receptacle.

Limitation: Placement of this PCIe adapter is not supported in the
PCIe Gen3 I/O Drawer.

Recommended performance Max assuming high utilization, One adapter
per four active processors cores.

Recommended connectivity Max assuming high utilization, two
adapters per one physical processor core.

Attributes provided: 10Gb FCoE PCIe Dual Port Adapter (LP)

Attributes required: 1 Low Profile (LP) slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i supported only with VIOS

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

For 9119-MME: (#5270)

Minimum required: 0

Maximum allowed: 16 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i supported only with VIOS

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5271) - PCIe LP 4-Port 10/100/1000 Base-TX Ethernet Adapter

This is the low-profile (LP) equivalent of feature number 5717.
The 4-Port 10/100/1000 Base-TX PCI Express Adapter is a short/ low-
profile, full duplex, four-ported Gigabit Ethernet adapter that can be
configured to run any of the ports at 1000,100 or 10 Mbps data rate.
This adapter interfaces to the system via a PCIe bus and connects to a
network using a 4-pair CAT-5 Unshielded Twisted Pair (UTP) cable for
distances of up to 100m. Each port is independent of one another and
is boot capable under AIX Network Installation Management (NIM). The
adapter conforms to the IEEE 802.3ab 1000Base-T standard. The #5717
supports jumbo frames when running at the 1000 Mbps speed.

The 4-Port 10/100/1000 Base-TX PCI Express adapter (#5717) should
be considered where maximum port density is required per I/O card
slot. For a suggested maximum number of adapters considering
performance, see the IBM System p PCI placement guide (SA76-0090) for
information about the PCIe slots on your system unit.

Limitations: The 1000 Mbps speed is not supported in Half Duplex
(HDX) mode.

Attributes provided: Four-port Gigabit Ethernet

Attributes required: 1 low-profile (LP) slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i supported with VIOS

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

For 9119-MME: (#5271)

Minimum required: 0

Maximum allowed: 16 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i supported with VIOS

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5273) - PCIe LP 8Gb 2-Port Fibre Channel Adapter

This feature is the Low Profile (LP) equivalent of feature number
5735.

The 8 Gigabit PCI Express Dual Port Fibre Channel Adapter is a
high-performance 8x short form adapter based on the Emulex LPe12002
PCIe Host Bus Adapter (HBA). Each port provides single initiator
capability over a fibre link. The ports have LC type connectors and
utilize shortwave laser optics. The adapter operates at link speeds
of 2, 4, and 8 Gbps and will automatically negotiates to the highest
speed possible. LEDs on each port provide information on the status
and link speed of the port.

The adapter connects to a Fibre Channel switch (AIX, IBM i,
Linux,VIOS). If in an IBM i environment, devices can also be directly
attached.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables.

The following table shows the supported distances for the three
different cable types at the three different link speeds.

See also feature #EN0F or #EN0G for a 2-port 8Gb Fibre Channel
adapter based on a QLogic adapter.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) In some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: Dual port fibre channel

Attributes required: 1 Low Profile (LP) slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5274) - PCIe LP 2-Port 1GbE SX Adapter

(No Longer Available as of May 11, 2015)

This feature is the Low Profile (LP) equivalent of feature number
5768. The IBM 2-Port Gigabit Ethernet-SX PCI Express (PCIe) Adapter
provides two 1 Gbps (1000 Base-SX) full-duplex Ethernet LAN
connections. The adapter interfaces to the system via a PCIe bus. It
is PCIe x4 capable and conforms to the PCIe 1.0a standard. The
adapter connects to a network using a 50/62.5 micron shortwave (850
nm) multimode optical cable that conforms to the IEEE 802.3z standard.
The adapter supports distances of 260m for 62.5 micron Multi Mode
Fiber (MMF) and 550m for 50.0 micron MMF. AIX Network Install Manager
(NIM) boot capability is supported with this adapter.

A function called 'Large Send' or sometimes known as TCP
Segmentation is also provided by this adapter. This function offloads
the TCP segmentation operation from the AIX IP layer to the adapter
for outgoing (transmit side) TCP segments. Another function known as
"Checksum Offload" which offloads the TCP Checksum Operation or
workload from the CPU to the adapter is also provided.

For a suggested maximum number of adapters taking performance into
consideration, see the IBM System p PCI placement guide (SA76-0090)
for information about the PCIe slots on your system unit.

Note: The 2-Port IBM Gigabit Ethernet-SX PCIe Adapter incorporates an
LC type connector on the card. This new, smaller form factor
connector is being used by the industry for the next generation of
fiber optic networks. If connecting into an older, existing SC type
connector network, an LC-SC 62.5 Micron Fiber Converter Cable (#2459)
or LC-SC 50 Micron Fiber Converter Cable (#2456) is required.

(#5276) - PCIe LP 4Gb 2-Port Fibre Channel Adapter

This feature is the Low Profile (LP) equivalent of feature number
5774. The 4 Gigabit Dual Port Fibre Channel Adapter is a 64-bit
address/data, short form factor PCIe adapter with an LC type external
fiber connector. With the use of appropriate optical fiber cabling,
this adapter provides the capability for a network of high-speed local
and remote located storage. The adapter will auto-negotiate for the
highest data rate between adapter and an attaching device at 1 Gbps, 2
Gbps or 4 Gbps of which the device or switch is capable. Between the
adapter and an attaching device or switch, the distances supported are
up to: 500 meters running at 1 Gbps data rate, 300 meters running at
2 Gbps data rate, and 150 meters running at 4 Gbps data rate. When
used with IBM Fibre Channel storage switches supporting long-wave
optics, distances of up to 10 kilometers are capable running at either
1 Gbps, 2 Gbps, or 4 Gbps data rates.

The 4 Gigabit PCIe Dual Port Fibre Channel Adapter can be used to
attach devices either directly, or by means of Fibre Channel Switches.
If attaching a device or switch with a SC type fiber connector(s), use
of an LC-SC 50 Micron Fiber Converter Cable (#2456) or a LC-SC 62.5
Micron Fiber Converter Cable (#2459) is required.

Consult with your IBM representative or Business Partner for
additional information relative to any third party attachment.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: Dual-port Fibre channel

Attributes required: 1 Low Profile (LP) slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5277) - PCIe LP 4-Port Async EIA-232 Adapter

This feature is the Low Profile (LP) equivalent of feature number
5785. One connector on the rear of the adapter provides attachment
for a fan-out cable (provided) which provides four EIA-232 ports.

Note #5785 and # 5277 are physically and electrically identical
adapters, except for the type of PCIe slot used (full-high or low
profile). Note also the 4-port #5277/5785 is functionally nearly
identical to the 2-port #5289/5290 except for the number and type of
connectors.

(#5281) - PCIe LP 2-Port 1GbE TX Adapter

The IBM PCIe LP 2-Port 1GbE TX Adapter is a Full Duplex, dual
ported, Gigabit Ethernet adapter designed with highly integrated
components. This adapter can be configured to run each port at 10,
100, or 1000 Mbps data rates. The adapter interfaces to the system
via a PCIe bus. It is PCIe x4 capable and conforms to the PCIe 1.0a
standard. The adapter connects to a network using a 4-pair CAT-5
Unshielded Twisted Pair (UTP) cable for distances of up to 100m. AIX
Network Install Manager (NIM) boot capability is supported with this
adapter. The adapter conforms to the IEEE 802.3ab 1000Base-T
standard. The adapter also supports jumbo frames when running at the
1000 Mbps speed.

A function called 'Large Send' or sometimes known as TCP
Segmentation is also provided by this adapter. This function offloads
the TCP segmentation operation from the AIX IP layer to the adapter
for outgoing (transmit side) TCP segments. Another function known as
"Checksum Offload" offloads the TCP Checksum Operation or workload
from the CPU to the adapter is also provided.

The IBM 2-Port 10/100/1000 Base-TX Ethernet PCIe Adapter (#5281)
should be considered where maximum port density is required per I/O
card slot. For a suggested maximum number of adapters taking
performance into consideration, see the IBM System p PCI placement
guide (SA76-0090) for information about the PCIe slots on your system
unit. The adapter is the same as 5767 and CINN.

Limitations: The 1000 Mbps speed is not supported in Half Duplex
(HDX) mode.

(#5283) - PCIe2 LP 2-Port 4X IB QDR Adapter 40Gb

(No Longer Available as of January 20, 2017)

The PCIe Gen-2 low profile 2-port 4X InfiniBand QDR adapter
provides high speed connectivity with other servers or IB switches.
Each port maximum of 40Gb assumes no other system or switch
bottlenecks are present.

Note: A PCIe Gen2 slot does not have the bandwidth to support two 40Gb
ports. The benefit of the second 40Gb port is additional attachment
redundancy, not performance. Also note: AIX supports multiple types
of adapter usages across different applications and adapter
performance should generally meet or exceed client needs. But when
using TCP/IP, some AIX clients may observe less than full
unidirectional link bandwidth and less than linear scaling for duplex
bandwidth with the #5283.

Limitation: Not supported by VIOS

Attributes provided: PCIe2 LP 2-Port 4X IB QDR Adapter 40Gb

Attributes required: PCIe Gen2 low profile slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX supported

IBM i not supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

SUSE Linux Enterprise 12, or later

Ubuntu 15.04, or later

Red Hat Enterprise Linux 7.1, little endian, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5284) - PCIe2 LP 2-port 10GbE SR Adapter

The PCIe2 2-port 10GbE SR Adapter is a low-profile capable, high
performance adapter that uses a LC Duplex type connector and is
capable of transferring data a distance of 300m over MMF-850nm Fiber
cable. The product conforms to the IEEE, 802.3ae 10GBASE-SR
specification for Ethernet transmission.

(#5285) - PCIe2 2-Port 4X IB QDR Adapter 40Gb

(No Longer Available as of January 20, 2017)

The PCIe Gen-2 2-port 4X InfiniBand QDR adapter provides high speed
connectivity with other servers or IB switches. Each port maximum of
40Gb assuming no other system or switch bottlenecks are present. This
adapter performs the same function as #5283 but with a high profile
(HP) tail stock and cannot be plugged into a low profile Gen-2 slot.

Note: As of 3/12/2013 This feature is not available in the following EMEA
countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait, Lebanon, Libya,
Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi Arabia, Somalia,
Tunisia and United Arab Emirates which includes IBM Middle East -
Dubai and Abu Dhabi and Yemen.

Attributes provided: PCIe2 2-Port 4X IB QDR Adapter 40Gb

Attributes required: PCIe Gen2 full high slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5287) - PCIe2 2-port 10GbE SR Adapter

The PCIe2 2-port 10GbE SR Adapter is a full high, Gen2, high
performance adapter that uses a LC Duplex type connector and is
capable of transferring data a distance of 300m over MMF-850nm Fiber
cable. The product conforms to the IEEE, 802.3ae 10GBASE-SR
specification for Ethernet transmission.

Recommended performance Max assuming high utilization, One adapter
per four active processors cores.

Recommended connectivity Max assuming high utilization, two
adapters per one physical processor core.

Attributes provided: 10 Gb FCoE PCIe Dual Port Adapter

Attributes required: Open PCIe slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 0)

OS level required:

Linux supported

AIX supported

IBM i supported only with VIOS

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5717) - 4-Port 10/100/1000 Base-TX PCI Express Adapter

The 4-Port 10/100/1000 Base-TX PCI Express Adapter is short/
low-profile, full duplex, four-ported Gigabit Ethernet adapter that can
be configured to run any of the ports at 1000,100 or 10 Mbps data rate.
This adapter interfaces to the system via a PCIe bus and connects to a
network using a 4-pair CAT-5 Unshielded Twisted Pair (UTP) cable for
distance of up to 100m. Each port is independent of one another and
is boot capable under AIX Network install manager (NIM). The adapter
conforms to the IEEE 802.3ab 1000Base-T standard. The #5717 supports
jumbo frames when running at the 1000 Mbps speed.

The 4-Port 10/100/1000 Base-TX PCI Express adapter (#5717) should
be considered where maximum port density is required per I/O card
slot. For a suggested maximum number of adapters considering
performance, see the IBM System p PCI placement guide (SA76-0090) for
information about the PCIe slots on your system unit.

Limitations: The 1000 Mbps speed is not supported in Half Duplex
(HDX) mode.

Attributes provided: Four-ported Gigabit Ethernet

Attributes required: One available PCIe card slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 0)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5729) - PCIe2 8Gb 4-port Fibre Channel Adapter

PCIe Gen2 8 Gigabit quad port Fibre Channel Adapter is a high-
performance 8x short form adapter based on the Emulex LPe12004 PCIe
Host Bus Adapter (HBA). Each port provides single initiator
capability over a fibre link or with NPIV, multiple initiator
capability is provided. The ports have LC type connectors and utilize
shortwave laser optics. The adapter operates at link speeds of 2, 4,
and 8 Gbps and will automatically negotiate to the highest speed
possible. LEDs on each port provide information on the status and
link speed of the port.

The adapter connects to a Fibre Channel switch. Direct device
attachment has not been tested and is not supported.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables.

The following table shows the supported distances for the three
different cable types at the three different link speeds.

(#5735) - 8 Gigabit PCI Express Dual Port Fibre Channel Adapter

The 8 Gigabit PCI Express Dual Port Fibre Channel Adapter is a
high-performance 8x short form adapter based on the Emulex LPe12002
PCIe Host Bus Adapter (HBA). Each port provides single initiator
capability over a fibre link. The ports have LC type connectors and
utilize shortwave laser optics. The adapter operates at link speeds
of 2, 4, and 8 Gbps and will automatically negotiates to the highest
speed possible. LEDs on each port provide information on the status
and link speed of the port.

The adapter connects to a Fibre Channel switch (AIX, IBM i,
Linux,VIOS). If in an IBM i environment, devices can also be directly
attached.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables.

The following table shows the supported distances for the three
different cable types at the three different link speeds.

#5735 feature indicates a full high adapter. #5273 feature
indicates a low profile adapter which is electronically identical.
CCIN is 577D. Consult with your IBM representative or Business
Partner for additional information relative to any third party
attachment.

See also feature #EN0F or #EN0G for a 2-port 8Gb Fibre Channel
adapter based on a QLogic adapter.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) In some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: Dual Port Fibre Channel

Attributes required: 1 Empty PCIe slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#5744) - PCIe2 4-Port 10GbE&1GbE SR&RJ45 Adapter

(No Longer Available as of February 23, 2017)

This PCIe Gen2 Ethernet adapter provides two 10 Gb SFP+ SR optical
transceiver ports and two 1 Gb RJ45 ports. This is a full high
adapter.

For the 10Gb ports, LC Duplex type connectors and MMF-850nm fiber
cable are used. With 62.5 micron OM1, up to 33 meter length fiber
cables are supported. With 50 micron OM2, up to 82 meter fiber cable
lengths are supported. With 50 micron OM3, up to 300 meter fiber
cable lengths are supported.

For the 1Gb RJ45 ports, The RJ45 port, 4-pair CAT-5 Unshielded
Twisted Pair (UTP) cable or higher is supported for distances of up to
100 meters is supported. In addition to 1Gb (1000 Mb) networks, 100
Mb networks are also supported, but 10Mb networks are not supported.
Note #5744 and #5280 adapters are physically and electronically
identical except for their low profile or full high PCI slot
attribute.

(#5748) - POWER GXT145 PCI Express Graphics Accelerator

(No Longer Available as of November 5, 2015)

Note: Supported prior to 2015/05/11.

The POWER GXT145 is a versatile, low-priced 2D graphics
accelerator. It can be configured to operate in either 8-bit or
24-bit color modes. This adapter supports both analog and digital
monitors. The adapter requires a PCI Express slot. If attaching a
device that requires a 15 pin D-Shell receptacle for a VGA connection
(eg. when the graphic adapter output is routed directly to a 7316-TF3
display or indirectly through a KVM switch), order a VGA to DVI
Connection Converter, feature number 4276 to accommodate the attaching
device.

Limitation: Placement of this PCIe adapter is not supported in the
PCIe Gen3 I/O Drawer.

2nd monitor support in AIX is only in clone mode with an analog
connection

APIs Supported

X-Windows and Motif

Software requirements

The total number of Graphics Adapters in any one partition may not
exceed four.

Attributes provided: 2D Graphics Adapter

Attributes required: 1 PCI Express Slot

Minimum required: 0

Maximum allowed: 0 (Initial order maximum: 0)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5767) - 2-Port 10/100/1000 Base-TX Ethernet PCI Express Adapter

The IBM 2-Port 10/100/1000 Base-TX Ethernet PCI Express (PCIe)
Adapter is a Full Duplex, dual ported, Gigabit Ethernet adapter
designed with highly integrated components. This adapter can be
configured to run each port at 10, 100, or 1000 Mbps data rates. The
adapter interfaces to the system via a PCIe bus. It is PCIe x4
capable and conforms to the PCIe 1.0a standard. The adapter connects
to a network using a 4-pair CAT-5 Unshielded Twisted Pair (UTP) cable
for distances of up to 100m. AIX Network Install Manager (NIM) boot
capability is supported with this adapter. The adapter conforms to
the IEEE 802.3ab 1000Base-T standard. The adapter also supports jumbo
frames when running at the 1000 Mbps speed.

A function called 'Large Send' or sometimes known as TCP
Segmentation is also provided by this adapter. This function offloads
the TCP segmentation operation from the AIX IP layer to the adapter
for outgoing (transmit side) TCP segments. Another function known as
"Checksum Offload" which offloads the TCP Checksum Operation or
workload from the CPU to the adapter is also provided.

The IBM 2-Port 10/100/1000 Base-TX Ethernet PCIe Adapter (#5767)
should be considered where maximum port density is required per I/O
card slot. For a suggested maximum number of adapters taking
performance into consideration, see the IBM System p PCI placement
guide (SA76-0090) for information about the PCIe slots on your system
unit.

Limitations: The 1000 Mbps speed is not supported in Half Duplex
(HDX) mode.

A function called 'Large Send' or sometimes known as TCP
Segmentation is also provided by this adapter. This function offloads
the TCP segmentation operation from the AIX IP layer to the adapter
for outgoing (transmit side) TCP segments. Another function known as
"Checksum Offload" which offloads the TCP Checksum Operation or
workload from the CPU to the adapter is also provided.

The IBM 2-Port Gigabit Ethernet-SX PCIe Adapter (#5768) should be
considered where maximum port density is required per I/O card slot.
For a suggested maximum number of adapters taking performance into
consideration, see the IBM System p PCI placement guide (SA76-0090)
for information about the PCIe slots on your system unit.

Note: The 2-Port IBM Gigabit Ethernet-SX PCIe Adapter incorporates an
LC type connector on the card. This new, smaller form factor
connector is being used by the industry for the next generation of
fiber optic networks. If connecting into an older, existing SC type
connector network, an LC-SC 62.5 Micron Fiber Converter Cable (#2459)
or LC-SC 50 Micron Fiber Converter Cable (#2456) is required.

(#5769) - 10 Gigabit Ethernet-SR PCI Express Adapter

(No Longer Available as of July 14, 2017)

The 10 Gigabit Ethernet-SR PCI Express Adapter is a full-high, high
performance adapter that uses a LC Duplex type connector and is
capable of transferring data a distance of 300m over MMF-850nm Fiber
cable. The product conforms to the IEEE, 802.3ae 10GBASE-SR
specification for Ethernet transmission.

* Supported by SUSE Linux Enterprise Server 10 SP3 for POWER
Systems, or later and Red Hat Enterprise Linux for POWER, version 5.4,
or later.

Attributes provided: One 10 Gigabit Ethernet port

Attributes required: One x8 lane or x16 PCI Express slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#5774) - 4 Gigabit PCI Express Dual Port Fibre Channel Adapter

The 4 Gigabit Dual Port Fibre Channel Adapter is a 64-bit
address/data, short form factor PCIe adapter with an LC type external
fiber connector. With the use of appropriate optical fiber cabling,
this adapter provides the capability for a network of high-speed local
and remote located storage. The adapter will auto-negotiate for the
highest data rate between adapter and an attaching device at 1 Gbps, 2
Gbps or 4 Gbps of which the device or switch is capable. Between the
adapter and an attaching device or switch, the distances supported are
up to: 500 meters running at 1 Gbps data rate, 300 meters running at
2 Gbps data rate, and 150 meters running at 4 Gbps data rate. When
used with IBM Fibre Channel storage switches supporting long-wave
optics, distances of up to 10 kilometers are capable running at either
1 Gbps, 2 Gbps, or 4 Gbps data rates.

The 4 Gigabit PCIe Dual Port Fibre Channel Adapter can be used to
attach devices either directly, or by means of Fibre Channel Switches.
If attaching a device or switch with a SC type fiber connector(s), use
of an LC-SC 50 Micron Fiber Converter Cable (#2456) or a LC-SC 62.5
Micron Fiber Converter Cable (#2459) is required.

Consult with your IBM representative or Business Partner for
additional information relative to any third party attachment.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: 2 Fibre Channel

Attributes required: 1 Empty PCIe slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#5785) - 4 Port Async EIA-232 PCIe Adapter

Connection for 4 asynchronous EIA-232 devices. Ports are
programmable to support EIA-232 protocols, at a line speed of 128K
bps. One connector on the rear of the adapter provides attachment for
a fan-out cable (provided) which provides four EIA-232 ports.

Note #5785 and # 5277 are physically and electrically identical
adapters, except for the type of PCIe slot used (full-high or low
profile). Note also the 4-port #5277/5785 is functionally nearly
identical to the 2-port #5289/5290 except for the number and type of
connectors.

(#5805) - PCIe 380MB Cache Dual - x4 3Gb SAS RAID Adapter

The#5805 PCIe 380 MB Cache Dual - x4 3 Gb SAS RAID Adapter is a
short, full high form factor adapter that supports the attachment of
SAS disk and SAS solid state drives (SSD) using a pair of mini SAS 4x
connectors. Write cache can provide an I/O performance boost even if
RAID 5/6/10 is not used. From a high-level perspective, except for
having a larger write cache, it is very similar to the feature 5902
PCI-X SAS adapter and provides a high-performance connection to SAS
devices. Two of feature 5805 provides for mirrored write cache data
and mirrored RAID parity footprints between the adapters for superior
availability. Feature 5805 is installed in pairs, allowing for
redundancy of the write cache. If the feature 5805 pairing is broken,
then write cache is disabled. The #5805 can also be paired with a
#5903 adapter.

Feature 5805 supports SAS disk drives (HDD) and SAS-bay-based SSD
located in a PCIe 12X I/O Drawer or drives located in an EXP 12S Disk
Drawer or EXP24S Disk Drawer. AIX/Linux formatted SAS drives can be
attached, and RAID 0, RAID 5, RAID 6, and RAID 10 are supported.
Similarly, IBM i RAID formatted HDD and SSD can also also be attached,
and RAID 5 and RAID 6 is support. AIX, IBM i and Linux also support
The #5805 is electronically identical to the #5903 and both the 5903
and 5805 use CCIN number is 574E. The #5805 adapter is slightly more
narrow than the #5903 due to the placement of its cache batteries.

With proper cabling and configuration, multiple wide ports are used
to provide redundant paths to each dual port SAS disk or SSD. The
adapter manages SAS path redundancy and path switching should a SAS
drive failure occur. The pairing of #5805 provides a high
availability I/O configuration to protect against the failure of a SAS
adapter. SAS X cables attach SAS disk drives in an EXP 12S and/ or
EXP24S Disk Drawer. SAS #3688 cables attach SFF SAS drives in an PCIe
12X I/O Drawer. The high availability I/O configuration connection is
provided via the internal wiring within the PCIe 12X I/ O drawer
itself.

Highlights

Supports up to 48 SAS disks, when configured with four #5886 EXP
12S Disk Drawers or two #5887/EL1S EXP24S Disk Drawer

Attributes required: Configuration always requires even pairs of
#5805 (or #5903/5805). SAS Media devices are not supported. When
attaching #5886 EXP12S or #5887/EL1S EXP24S at least one of the
following SAS (X) cables #3661, #3662 or #3663 must be used.

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 0)

OS level required:

AIX supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#5887) - EXP24S SFF Gen2-bay Drawer

The EXP24S SFF Gen2-bay Drawer is an expansion drawer with
twenty-four 2.5-inch form factor SAS bays. #5887 supports up to 24
hot-swap SFF SAS Hard Disk Drives (HDD). It uses only 2 EIA of space in
a 19-inch rack. The EXP24S includes redundant AC power supplies and two
power cords. The EXP24S SFF bays use Gen-2 or SFF-2 SAS bays that are
not compatible with CEC SFF Gen-1 SAS bays or with #5802/ 5803 SFF SAS
bays.

With AIX/Linux/VIOS, the EXP24S can be ordered with four sets of 6
bays, two sets of 12 bays or one set of 24 bays (mode 4, 2 or 1).
With IBM i the EXP24S can be ordered as one set of 24 bays (mode 1).

The EXP24S SAS ports are attached to SAS controller(s) which can be
a SAS PCI-X or PCIe adapter or pair of adapters. The EXP24S can also
be attached to an imbedded SAS controller in a server with an imbedded
SAS port. Attachment between the SAS controller and the EXP24S SAS
ports is via the appropriate SAS Y or X cables.

Limitation: The mode is set at the IBM factory. The capability to
change modes after manufacture is not offered.

(#5899) - PCIe2 4-port 1GbE Adapter

This short PCIe Gen2 adapter provides four 1Gb Ethernet ports that
can be configured to run at 1000, 100 or 10 Mbps. 4-pair CAT-5
Unshielded Twisted Pair (UTP) cables up to 100 meters in length are
attached to the copper RJ45 connectors. Each port is independent of
one another and supports full-duplex or half-duplex. 1000 Mbps speed
is not supported in Half Duplex (HDX) mode.

Feature #5260 and #5899 are electronically identical and have the
same CCIN of 576F. #5260 indicates a low profile tail stock while
#5899 indicates a full high tail stock.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four-port 1 Gb Ethernet

Attributes required: 1 Full High Profile PCIe slot (Gen1 or Gen2)

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#5901) - PCIe Dual-x4 SAS Adapter

The #5901 PCIe Dual-4x SAS Adapter is a low-profile short form
factor adapter which supports the attachment of SAS disk, tape, and
DVD using a pair of mini SAS 4x connectors. From a high level
perspective, it is functionally equivalent to the #5912 PCI-X SAS
adapter and provides a high-performance connection to SAS devices.

The #5901 supports external SAS tape drives such as the DAT72, DAT160,
LTO-4, LTO-5, LTO-6, and LTO-7 found in the IBM tape units such as the
7226-1U3, 7214-1U2, TS2240, TS2340, TS3100, TS3200, and TS3310. Other
removable media devices supported include IBM SAS/SATA DVD-ROM/RAM
drives. SAS adapter- to-enclosure (AE) cables are used to attach these
drives. See also feature #EJ1N.

The #5901 supports SAS SFF disk drives located in a PCIe 12X I/O
Drawer or SAS disk drives located in an EXP 12S or EXP24S Disk Drawer
or drives in a Power6 system CEC (split DASD backplane). AIX/ Linux
formatted SAS drives are supported with RAID 0 (with mirroring) and
RAID 10. IBM i formatted SAS drives are supported and data spreading
and mirroring functions are provided by IBM i. RAID-5 or RAID-6 are
not supported on the #5901. #5901 has zero write cache. CCIN for
#5901 is 57B3.

With proper cabling and configuration, multiple wide ports are used
to provide redundant paths to each dual port SAS disk. The adapter
manages SAS path redundancy and path switching should a SAS drive
failure occur. SAS Y cables attach SAS disk drives in an EXP12S or
EXP24S Disk Drawers. SAS #3688 cables attach SFF SAS drives in an
PCIe 12X I/O Drawer. With the EXP12S or EXP24S Drawer, a high
availability I/O configuration can be created using a pair of #5901
adapters and SAS X cables to protect against the failure of a SAS
adapter. In the PCIe 12X I/O Drawer, this function is provided via
the internal wiring within the drawer itself.

Limitation: 4k byte sector HDD/SSD are not supported. Only 5xx
byte sector drives are supported on this generation of PCIe adapter.

Highlights:

Supports up to 48 SAS disks, when configured with four #5886
EXP12S Disk Drawers or two #5887/#EL1S EXP24S Disk Drawers

Supports up to 42 disk (18 SFF disk plus up to 24 3.5-inch SAS
disk) when configured with a #5802/#EL36 19-inch PCIe 12X I/O Drawer
and two #5886 EXP 12S Disk Drawers or configured with a #5802/#EL36
and one #5887/#EL1S EXP24S

Supports up to 50 disk (26 SFF disk plus up to 24 3.5-inch SAS
disk) when configured with a #5803 24-inch PCIe 12X I/O Drawer and two
#5886 EXP 12S Disk Drawers or configured with a #5803 and one
#5887/#EL1S EXP24S

(#5913) - PCIe2 1.8GB Cache RAID SAS Adapter Tri-port 6Gb

The PCIe2 1.8GB Cache RAID SAS Adapter Tri-port 6Gb (#5913)
provides high performance HDD or SSD controller function using PCIe
Gen2 technology. HDD and SSD can be either SFF or 3.5-inch drives (or
both). A pair of adapters are required to provide mirrored write
cache data and adapter redundancy. Integrated flash memory provides
protection of the write cache without batteries in case of power
failure.

The #5913 provides three Mini-SAS HD (high density) connectors for
the attachment of SAS drives located in the #5887/EL1S EXP24S, #5886
EPX12S, or #5802/5803/EL36 12X PCIe I/O drawers. X, YO or AT SAS
cables with HD connectors are used to attach to these drawers. A max
of 3 EXP24S or 6 EXP12S can be attached. If controlling drives in a
#5802/5803/EL36 the #5913 pairs must be located in that #5802/
5803/EL36. An AA SAS cable with HD connectors is attached to the
#5913 pair to communicate status and cache content information and is
required unless all three ports are being used to attach I/O drawers.

(#5915) - SAS AA Cable 3m - HD 6Gb Adapter to Adapter

This 3 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density)connectors
which connect the top connectors of two #5913 1.8GB RAID SAS Adapters.
The cable provides a high performance path of all the dual controller
communication including mirroring the write cache and status awareness
of each card. This cable can support up to 6Gb throughput.

An AA cable is always required between a pair of #5913 adapters
unless all six (three pair) of ports are used to attach I/O drawers.
If an AA cable fails or is disconnected, then the information it was
carrying is then sent over the cables attached to the I/O drawers,
sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

Attributes provided: high speed connection between two #5913 SAS
adapters

Attributes required: available top connectors on pair of #5913
SAS adapters

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5916) - SAS AA Cable 6m - HD 6Gb Adapter to Adapter

This 6 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density)connectors
which connect the top connectors of two #5913 1.8GB RAID SAS Adapters.
The cable provides a high performance path of all the dual controller
communication including mirroring the write cache and status awareness
of each card. This cable can support up to 6Gb throughput.

An AA cable is always required between a pair of #5913 adapters
unless all six (three pair) of ports are used to attach I/O drawers.
If an AA cable fails or is disconnected, then the information it was
carrying is then sent over the cables attached to the I/O drawers,
sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

Attributes provided: high speed connection between two #5913 SAS
adapters

Attributes required: available top connectors on pair of #5913
SAS adapters

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5917) - SAS AA Cable 1.5m - HD 6Gb Adapter to Adapter

This 1.5 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density)connectors
which connect the top connectors of two #5913 1.8GB RAID SAS Adapters.
The cable provides a high performance path of all the dual controller
communication including mirroring the write cache and status awareness
of each card. This cable can support up to 6Gb throughput.

An AA cable is always required between a pair of #5913 adapters
unless all six (three pair) of ports are used to attach I/O drawers.
If an AA cable fails or is disconnected, then the information it was
carrying is then sent over the cables attached to the I/O drawers,
sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

Attributes provided: high speed connection between two #5913 SAS
adapters

Attributes required: available top connectors on pair of #5913
SAS adapters

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5918) - SAS AA Cable 0.6m - HD 6Gb Adapter to Adapter

This 0.6 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density)connectors
which connect the top connectors of two #5913 1.8GB RAID SAS Adapters.
The cable provides a high performance path of all the dual controller
communication including mirroring the write cache and status awareness
of each card. This cable can support up to 6Gb throughput.

An AA cable is always required between a pair of #5913 adapters
unless all six (three pair) of ports are used to attach I/O drawers.
If an AA cable fails or is disconnected, then the information it was
carrying is then sent over the cables attached to the I/O drawers,
sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

Attributes provided: high speed connection between two #5913 SAS
adapters

Attributes required: available top connectors on pair of #5913
SAS adapters

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#5924) - Non-paired Indicator 5913 PCIe SAS RAID Adapter

Feature 5924 must be added for every instance of a non-paired SAS
RAID adapter #5913. It identifies a specific high availability
configuration supported by AIX or Linux which has one #5913 on one
system and the paired #5913 located on a second system. IBM i does
not support paired adapter on different servers.

Attributes required: Every #5924 requires a 6Gb/s SAS RAID
adapter (#5913) on both this server and on another server that will
pair up the SAS RAID adapter and enable the onboard caches to
function.

Minimum required: 0

Maximum allowed: 48 (Initial order maximum: 48)

OS level required: None

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#6068) - Opt Front Door for 1.8m Rack

#6068 provides an attractive black full height rack door on the
#0551 19 Inch 1.8m Rack. The door is steel, with a perforated flat
front surface. The perforation pattern extends from the bottom to the
top of the door to enhance ventilation and provide some visibility
into the rack.

Attributes provided: Front Door

Attributes required: #0551 19 inch 1.8m Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6069) - Opt Front Door for 2.0m Rack

#6069 provides an attractive black full height rack door on the
#0553 19 inch 2.0m Rack. The door is steel, with a perforated flat
front surface. The perforation pattern extends from the bottom to the
top of the door to enhance ventilation and provide some visibility
into the rack.

Attributes provided: Front Door

Attributes required: #0553 19 inch 2.0 meter Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6238) - HIGH-END APPEARANCE SIDE COVERS

This feature provides side covers designed to be used when the rack
will contain a 9179-MHB, 9179-MHC. 9179-MHD, 8412-EAD

Attributes provided: Side covers for a 2M rack

Attributes required: None

Minimum required: 0

Maximum allowed: 16 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6246) - 1.8m Rack Trim Kit

#6246 provides a decorative trim kit for the front of an #0551 19
inch 1.8m Rack.

Attributes provided: Decorative trim kit

Attributes required: #0551 19 inch 1.8m Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#6247) - 2.0m Rack Trim Kit

#6247 provides a decorative trim kit for the front of a #0553 19
inch 2.0m Rack.

Attributes provided: Decorative trim kit

Attributes required: #0553 19 inch 2.0 meter Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#6248) - 1.8m Rack Acoustic Doors

#6248 provides front and rear doors for use with the #0551 19 inch
1.8m Rack. This door kit provides additional acoustic dampening for
use where a quieter environment is desired. #6248 results in a larger
footprint and requires additional space.

Attributes provided: Acoustic Door Kit

Attributes required: #0551 19 inch 1.8m Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6249) - 2.0m Rack Acoustic Doors

#6249 provides front and rear doors for use with the #0553 19 inch
2.0m Rack. This door kit provides additional acoustic dampening for
use where a quieter environment is desired. #6249 results in a larger
footprint and requires additional space.

Attributes provided: Acoustic Door Kit

Attributes required: #0553 19 inch 2.0 meter Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6263) - 1.8m Rack Trim Kit

This feature provides a decorative trim kit for the front of
feature number 0551 (19 inch 1.8m Rack).

Attributes provided: Decorative trim kit

Attributes required: #0551 19 inch 1.8m Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6272) - 2.0m Rack Trim Kit

This feature provides a decorative trim kit for the front of
feature number 0553 (19 inch 2.0m Rack).

Attributes provided: Decorative trim kit

Attributes required: #0553 19 inch 2.0 meter Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6458) - Pwr Crd 4.3m 14ft to IBM PDU

Standard IBM rack power cable that goes from the system or I/O
drawer to the rack power distribution unit (PDU). Cable has C13 on
one end (for C14 power supply connector on system unit or I/O drawer)
and C14 on the other end (for IBM PDU C13 receptacle). Note for
different length C13/C14 cables see #6671 (2.7M) or #6672 (2.0M).

Attributes provided: Power jumper cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6460) - Power Cord 4.3m (14-ft), Drawer To OEM PDU (125V, 15A)

This power cord goes from the system or I/O drawer to the rack OEM
power distribution unit or wall socket outlet. Cable has C13 on one
end (for C14 power supply connector on system unit or I/O drawer) and
plug type #4 (NEMA 5-15) on the other end.

This power cord goes from the system or I/O drawer to the wall or
rack OEM power distribution unit. Cable has C13 on one end (for C14
power supply connector on system unit or I/O drawer) and Plug type #5
(NEMA 6-15) on the other end for wall or OEM PDU.

(#6470) - Power Cord 1.8m (6-ft), Drawer to Wall (125V/15A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #4 (NEMA 5-15). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. 6-foot
length.

The following countries/regions use the #6470 power cord to power
the system and/or peripheral features requiring a power cord:

(#6478) - Power Cord 2.7 M(9-foot), To Wall/OEM PDU, (250V, 16A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #25 (CEI 23-16). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. The
following countries/regions use the #6478 power cord to power the
system and/or peripheral features requiring a power cord: Chile Italy
Libya

This power cord goes from the system and/or peripheral features to
a wall-type outlet. 125V, 15A or 250V, 10A, Plug Type #2. Refer to
Corporate Bulletin C-B-2-4700-009 for a description of plug types.

The following countries/regions use the #6488 power cord to power
the system and/or peripheral features requiring a power cord:

Argentina, Paraguay, Uruguay.

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6489) - 4.3m (14-Ft) 3PH/24A 380-415V Power Cord

#6489 is a 14-FT/4.3m 3PH/24A power cable with a Type 46 plug which
distributes power from a power source to a Power Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6491) - 4.3m (14-Ft) 1PH/63A 200-240V Power Cord

#6491 is a 14-FT/4.3m 200-240V/63A power cord with a Type 46 plug
which distributes power from a power source to a Power Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6492) - 4.3m (14-Ft) 1PH/48-60A 200-240V Power Cord

#6492 is a 14-FT/4.3m 200-240V/48-60A power cord with a Type 46
plug which distributes power from a power source to a Power
Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6493) - Power Cord 2.7m (9-ft), Drawer to Wall/OEM PDU, (250V/10A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #62 (GB 1053). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types.

The following countries/regions use the #6493 power cord to power
the system and/or peripheral features requiring a power cord:

People's Republic of China.

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6494) - Power Cord 2.7m (9-ft), Drawer to Wall/OEM PDU, (250V/10A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #69 (IS 6538). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types.

The following countries/regions use the #6494 power cord to power
the system and/or peripheral features requiring a power cord:

India

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6496) - Power Cord 2.7M (9-foot), To Wall/OEM PDU, (250V, 10A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #66 (KETI). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. The
following countries/regions use the #6496 power cord to power the
system and/or peripheral features requiring a power cord: North Korea
South Korea

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6577) - Power Cable - Drawer to IBM PDU, 200-240V/10A

This feature permits manufacturing to select the optimum PDU power
jumper cord length (1.0M, 2.0M, 2.7M or 4.3M) for rack integration.
This feature is mandatory on initial order specifying factory
integration with IBM racks (such as with 7014-T00 or T42 racks).
Feature is not valid on initial order with non-factory integrated
feature 4650. Power jumper cord has C13 on one end (for C14 power
supply connector on system unit or I/O drawer) and C14 on the other
end (for C13 PDU receptacle).

Note: This feature is not used for MES orders except for bulk orders
by SDI clients only. See C13/C14 jumper cord features #6458 (4.3M),
#6671 (2.7M), #6672 (2.0M) when not using factory integration.

Attributes provided: One power jumper cord

Attributes required: At least one rack and the absence of #4650.

Minimum required: 0

Maximum allowed: 390 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: 1 or 2 per I/O drawer or CEC. MES orderable for SDI's only.
The MES order will ship the 14 foot cable equivalent to feature number
6458.

(#6580) - Optional Rack Security Kit

This feature provides hardware that can be added to a rack to
prevent unauthorized access. It includes keyed front and rear locks
for the #0551 and #0553 rack doors. It also includes two sliding bars
that mount inside the left and right rack side panels. The sliding
bars are accessible when the rack rear door is open. They can be
moved to a position that disables the external latches on the rack
side panels, and prevents removal of the side panels.

Attributes provided: Locking hardware for rack doors and
sidepanels

Attributes required: #0551 or #0553 19-Inch Rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6586) - Modem Tray for 19-Inch Rack

(No Longer Available as of December 29, 2017)

This feature provides hardware for installing one or two modems in
a 19-inch rack. The modem tray occupies 1U of rack space when it is
mounted in the front of the rack. It provides a secure location in
the rack for external modems such as the ones attached to the Hardware
Management Console.

Attributes provided: Hdw. to support two modems

Attributes required: 19-inch rack with 1U rack space available

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6651) - Power Cord 2.7M (9-foot), To Wall/OEM PDU, (125V, 15A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #75 (KETI). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. The
following countries/regions use the #6651 power cord to power the
system and/or peripheral features requiring a power cord: Taiwan

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6653) - 4.3m (14-Ft) 3PH/16A 380-415V Power Cord

#6653 is a 14-FT/4.3m 3PH/16A power cord with a Type 46 plug which
distributes power from a power source to a Power Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6654) - 4.3m (14-Ft) 1PH/24-30A Pwr Cord

#6654 is a 14-FT/4.3m 200-240V/24A locking power cord with a Type
12 plug which distributes power from a power source to a Power
Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6655) - 4.3m (14-Ft) 1PH/24-30A WR Pwr Cord

#6655 is a 14-FT/4.3m 200-240V/24A water-resistant power cord with
a Type 40 plug which distributes power from a power source to a Power
Distribution Unit.

Attributes provided: power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6656) - 4.3m (14-Ft)1PH/24A Power Cord

#6656 is a 14-FT/4.3m 200-240V/24A power cord with a Type 46 plug
which distributes power from a power source to a Power Distribution
Unit.

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6657) - 4.3m (14-Ft) 1PH/32A Power Cord

#6657 is a 14-FT/4.3m 1PH/32A power cord with a Type PDL plug which
distributes power from a power source to a Power Distribution Unit.

Attributes provided: PDU power cable

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6658) - 4.3m (14-Ft) 1PH/24A Pwr Cd-Korea

#6658 is a 14-FT/4.3m 200-240V/24A power cord with a Type KP plug
which distributes power from a power source to a Power Distribution
Unit.

Attributes provided: PDU power cable

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6659) - Power Cord 2.7M (9-foot), To Wall/OEM PDU, (250V, 15A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #76 (KETI). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. The
following countries/regions use the #6659 power cord to power the
system and/or peripheral features requiring a power cord: Taiwan

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6660) - Power Cord 4.3m (14-ft), Drawer to Wall/OEM PDU (125V/15A)

This power cord goes from the system and/or peripheral features to
a wall-type outlet. Plug type #59 (NEMA 5-15). Refer to Corporate
Bulletin C-B-2-4700-009 for a description of plug types. 14-foot
length.

This power cord meets the DENAN marking requirement in Japan.

Attributes provided: Power Cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6665) - Power Cord 2.8m (9.2-ft), Drawer to IBM PDU, (250V/10A)

Standard IBM rack power jumper cord that goes from the system or
I/O drawer to the rack power distribution unit (PDU). Cable has C13
on one end (for C14 power supply connector on system unit or I/O
drawer) and C20 on the other end (for IBM PDU C19 receptacle).

Note for power jumper cord which attach to PDUs with C13
receptacles, use features such as #6577, #6458, #6671, or #6672.

Attributes provided: Power jumper cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6667) - 4.3m (14-Ft) 3PH/32A 380-415V Power Cord-Australia

#6667 is a 14-FT/4.3m 380-45V/32A power cord with a Type PDL plug
which distributes power from a power source to a Power Distribution
Unit.

Attributes provided: PDU power cable

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#6669) - Power Cord 4.3M (14-foot), Drawer to OEM PDU, (250V, 15A)

This power cord goes from the system or I/O drawer to the rack
power distribution unit. Plug type #57 (NEMA 6-15). Refer to
Corporate Bulletin C-B-2-4700-009 for a description of plug types.
14-foot length. This power cord meets the DENAN marking requirement
in Japan.

Attributes provided: Power Cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6671) - Power Cord 2.7M (9-foot), Drawer to IBM PDU, 250V/10A

Standard IBM rack power cable that goes from the system or I/O
drawer to the rack power distribution unit (PDU). Cable has C13 on
one end (for C14 power supply connector on system unit or I/O drawer)
and C14 on the other end (for IBM PDU C13 receptacle). Note for
different length C13/C14 cables see #6458 (4.3M) or #6672 (2.0M).

Attributes provided: Power jumper cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6672) - Power Cord 2M (6.5-foot), Drawer to IBM PDU, 250V/10A

Standard rack power cable that goes from the system or I/O drawer
to the rack power distribution unit (PDU). Cable has C13 on one end
(for C14 power supply connector on system unit or I/O drawer) and C14
on the other end (for IBM PDU C13 receptacle). Note for different
length C13/C14 cables see #6458 (4.3M) or #6671 (2.7M).

Attributes provided: Power jumper cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#6680) - Power Cord 2.7m (9-ft), Drawer to Wall/OEM PDU, (250V/10A)

This insulated power cord goes from the system and/or peripheral
features to a wall-type outlet. Plug type #6 (AS 3112-1964 NZS 198).
Refer to Corporate Bulletin C-B-2-4700-009 for a description of plug
types.

The following countries/regions use the #6680 power cord to power
the system and/or peripheral features requiring a power cord:

(#7109) - Intelligent PDU+, 1 EIA Unit, Universal UTG0247 Connector

This feature is for an intelligent AC power distribution unit
(PDU+) that will allow the user to monitor the amount of power being
used by the devices that are plugged in to this PDU+. This AC power
distribution unit provides twelve C13 power outlets. It receives
power through a UTG0247 connector. It can be used for many different
countries and applications by varying the PDU to Wall Power Cord,
which must be ordered separately. Each PDU requires one PDU to Wall
Power Cord. Supported power cords include the following features:
#6489, #6491, #6492, #6653, #6654, #6655, #6656, #6657, and #6658.

Note: When purchased on an MES order with a feature code rack. This
PDU will be mounted in the rear side pockets until all 4 side pockets
on the rack have been filled. Any additional PDUs on the order will
be mounted in 1 unit of EIA rack space. When purchased as an MES
order for addition to a rack in the field. This PDU may not fit in
the side pockets of your rack due to a hardware interference with the
rack, and may require mounting in 1 unit of rack EIA space. Insure
rack space is available before placing the MES order for this PDU when
it is being ordered for field installation.

(#7118) - Environmental Monitoring Probe

The Environmental Monitoring Probe (EMP) enables you to remotely
monitor environmental conditions. Using a standard Web browser, you can
view the ambient temperature and humidity of the remote environment, as
well as the status of two additional contact devices, such as a smoke
detector or open-door sensor. The temperature/humidity probe plugs into a
RJ45 connector an a PDU+. The EMP can be used with any Powerware UPS
equipped with a 10/100 Mb ConnectUPS Web/SNMP Card (firmware v3.01 or
higher). The EMP can be located up to 20m (65.6 feet) away.

Attributes provided: Monitoring of temperature, humidity, and status
of two contacts/ sensors. A one meter cat5 Ethernet cable, double sided
hook and loop fabric, often called VELCRO(R) tape, two tie-wraps, and
screw with wall anchor for mounting.

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

Note: Maximum support is 8x 7118 per 0553 rack and 8x 7118 per 0551
rack. Maximum of one 7118 per 7109 is supported.

(#7188) - Power Distribution Unit

An AC Power Distribution Unit (PDU) which mounts in a 19" rack and
provides twelve C13 power outlets. The #7188 has six 16A circuit
breakers, with two power outlets per circuit breaker. System units
and/or expansion units must use a power cord with a C14 plug to
connect to the #7188.

One of the following line cords must be used to distribute power
from a wall outlet to the #7188:

#6489 - 14-Ft 3PH/24A Power Cord

#6491 - 14-Ft 1PH/63A Pwr Cord

#6492 - 14-Ft 1PH/48-60A Pwr Cord

#6653 - 14-Ft 3PH/16A Power Cord

#6654 - 14-Ft 1PH/24-30A Pwr Cord

#6655 - 14-Ft 1PH/24-30A WR Pwr Cord

#6656 - 14-Ft 1PH/32A Power Cord

#6657 - 14-Ft 1PH/24A Power Cord

#6658 - 14-Ft 1PH/24A Pwr Cd-Korea

Attributes provided: Power Distribution Unit with Twelve C13
power outlets.

(#7780) - 2.0m Rack Side Attach Kit

This feature allows a row of racks without side panels to be bolted
together in a continuous suite, using the provided side-to-side rack
connecting hardware. When multiple racks are joined in this way,
cables can be easily run between racks without exiting the continuous
rack enclosure. A small gap is maintained between the two adjacent
racks, which is filled by three matching steel trim pieces that snap
into place on the front, top, and rear, between each rack. The trim
pieces cover the space between each rack for an enhanced appearance
and for additional protection of the equipment inside the racks. Side
panels are needed only for the two end racks of the suite.

This feature provides a fifteen meter long Ethernet cable for
attachment of a Hardware Management Console to the system unit.

Attributes provided: 15M Ethernet Cable

Attributes required: Ethernet port on Hardware Management Console

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#7840) - Side-by-Side for 1.8m Racks

This feature allows a row of racks without side panels to be bolted
together in a continuous suite, using the provided side to side rack
connecting hardware. When multiple racks are joined in this way,
cables can be easily run between racks without having to exit the
continuous rack enclosure. A small gap is maintained between the two
adjacent racks, which is filled by three matching steel trim pieces
that snap into place on the front, top, and rear, between each rack.
The trim pieces cover the space between each rack for an enhanced
appearance and for additional protection of the equipment inside the
racks. Side panels are needed only for the two end racks of the
suite.

Attributes provided: Hardware and trim to attach two racks

Attributes required: #0551 rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#7841) - Ruggedize Rack Kit

For enhanced rigidity and stability of the rack, the optional
Ruggedized Rack Kit provides additional hardware that reinforces the
rack and anchors it to the floor. This hardware is designed primarily
for use in locations where earthquakes are a concern. The feature
includes a large steel brace or truss that bolts into the rear of the
rack. It is hinged on the left side so it can swing out of the way
for easy access to the rack drawers when necessary. The Ruggedize
Rack Kit also includes hardware for bolting the rack to a concrete
floor or a similar surface, and bolt-in steel filler panels for any
noccupied spaces in the rack.

(#8453) - Base Customer Spec Plcmnt

Requests that IBM deliver the system to the customer according to
the slot in drawer hardware placement defined by the account team.

Eliminates the need to have these parts relocated in the customers
environment as may happen if the order is placed without this feature
code.

Client placement specifications are collected using the System
Planning Tool (SPT) and processed through the marketing configurator.
(Use of the SPT is not required).

Requires account team to submit the output of the marketing
configurator into IBM manufacturing via the CSP website
http://www.ibm.com/eserver/power/csp (US Business Partners and
Distributors can bypass this step.)

Requires account team to assure that the marketing configurator
output submitted reflects the actual order placed.

Attributes provided: I/O component placement

Attributes required: Marketing Configurator output submitted to
the CSP website (US Business Partners and Distributors can bypass
this step.)

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: NA

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#8845) - USB Mouse

(No Longer Available as of October 31, 2017)

The optical LED USB Mouse has 2 buttons and a scroll wheel that
acts as a third button. Mouse cable is 1.8 meters long. OS does not
support scrolling with the wheel. Business black with red scroll
wheel.

Refer to the Software requirements section to find the supported
O/S levels for Linux.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#9169) - Order Routing Indicator- System Plant

This feature will be auto-selected by the Configurator Tool when
required. Use of this feature will affect the routing of the order.
Selection of this indicator will direct the order to a system plant
for fulfillment.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9300) - Language Group Specify - US English

English language group for nomenclature and standard publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9359) - specify mode-1 1(1)5901/5278 for EXP24S #5887/EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or EL1S)
will be configured by IBM Manufacturing in Mode 1 (One group of 24
drive bays) with one PCIe 3G SAS controller (#5901 or #5278),
utilizing one appropriate YO cable connecting to the I/O Adapter (IOA)
ports. IBM does not provide changes to the mode setting after #5887
or EL1S is shipped.

Attributes required: Two dual X cables (#3661/EL22 1.5M,
#3662/EL23 3M, #3663//EL21 15M) and four #5901/#5278s dedicated to the
single #5887 or #EL1S.

Minimum required: 0

Maximum allowed: 92 (Initial order maximum: 92)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#9366) - Specify mode-2 1(4)5901/5278 for EXP24S #5887/EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or EL1S
will be configured by IBM Manufacturing in Mode 2 (two sets of up to
12 drive bays) with four PCIe 3G SAS controllers (#5901 and/or #5278)
utilizing two appropriate dual X cables connecting to I/O Adapters
(IOA) ports. IBM does not provide changes to the mode setting after
#5887 or EL1S is shipped.

Supported by AIX and Linux.

Attributes provided: IOA configurations for Mode 2

Attributes required: Two dual X cables (##3661/EL22 1.5M,
#3662/EL23 3M, #3663//EL21 15M) and four #5901/#5278s dedicated to the
single #5887 or #EL1S.

Minimum required: 0

Maximum allowed: 92 (Initial order maximum: 92)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#9367) - Specify mode-1 1(2)5903/5805 for EXP24S #5887/EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or EL1S)
will be configured by IBM Manufacturing in Mode 1 (One group of 1-24
drive bays) with a pair of PCIe 3G SAS RAID controller (either pair
#5903 or 5805 or combination of both) utilizing two appropriate YO
cable connecting to the 3G I/O Adapter (IOA) ports. This specify only
impacts IBM's initial shipment of the EXP24S. IBM does not provide
changes to the mode setting after #5887 or EL1S is shipped.

Attributes provided: IOA configurations for Mode 1

Attributes required: Two YO cables (#3691/EL1T 1.5M, #3692/EL1V
3M, #3693/EL1W 6M, #3694/ EL1U 15M) and one port on each of two
#5903/#5805s

Minimum required: 0

Maximum allowed: 126 (Initial order maximum: 126)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

(#9368) - Specify mode-2 1(4) 5903/5805 for EXP24S #5887/EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer(#5887 or EL1S)
will be configured by IBM Manufacturing in Mode 2 (two sets of 12
drive bays) with two pair of PCIe 3G SAS RAID controller (either pair
#5903 or 5805 or combination of both) utilizing two appropriate X
cable connecting to the I/O Adapters (IOA). IBM does not provide
changes to the mode setting after #5887 or EL1S is shipped.

Supported by AIX and Linux.

Attributes provided: IOA configurations for Mode 2

Attributes required: Two dual X cables (#3661/EL22 1.5M,
#3662/EL23 3M, #3663//EL21 15M) and four #5903/#5805s dedicated to the
single #5887 or #EL1S".

Attributes required: Two YO cables (#3450/EL25 1.5M, #3451/EL29
3M, #3452/EL28 6M, #3453/ EL26 10M) and one port on each of two
#5913s. If fewer than 3 ports on each of the #5913s are used, an AA
cable (#5915/EL2C 3M, #5916/ EL2D 6M, #5917/EL2B 1.5M, #5918/EL2A
0.6M) is required to connect the pair of #5913s

(#9440) - New AIX License Core Counter

This feature is used to count the number of cores licensed to run
AIX.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9441) - New IBM i License Core Counter

This feature is used to count the number of cores licensed to run
IBM i.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9442) - New Red Hat License Core Counter

This feature is used to count the number of cores licensed to run
Red Hat Linux.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9443) - New SUSE License Core Counter

This feature is used to count the number of cores licensed to run
SUSE Linux.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9444) - Other AIX License Core Counter

This feature is used to count the number of existing AIX licenses
transferred from another server.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9445) - Other Linux License Core Counter

This feature is used to count the number of existing Linux licenses
transferred from another server.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9446) - 3rd Party Linux License Core Counter

This feature is used to count the number of cores licensed to run
3rd party Linux.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9447) - VIOS Core Counter

This feature is used to count the number of cores licensed to run
VIOS (Virtual I/O Server).

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9449) - Other License Core Counter

This feature is used to count the number of other cores licensed.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9450) - Ubuntu Linux License Core Counter

This feature is used to count the number of cores licensed to run
Ubuntu Linux.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9461) - Month Indicator

This month indicator is used to create a date stamp to enable CFR
splitting and rejoining in order to circumvent the AAS maximum
limitation of 30 systems entered on any one order. The quantity
ordered for this feature is generated by eConfig.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 12 (Initial order maximum: 12)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9462) - Day Indicator

This day indicator is used to create a date stamp to enable CFR
splitting and rejoining in order to circumvent the AAS maximum
limitation of 30 systems entered on any one order. The quantity
ordered for this feature is generated by eConfig.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 31 (Initial order maximum: 31)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9463) - Hour Indicator

This hour indicator is used to create a time stamp to enable CFR
splitting and rejoining in order to circumvent the AAS maximum
limitation of 30 systems entered on any one order. The quantity
ordered for this feature is generated by eConfig.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 24 (Initial order maximum: 24)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9464) - Minute Indicator

This hour indicator is used to create a time stamp to enable CFR
splitting and rejoining in order to circumvent the AAS maximum
limitation of 30 systems entered on any one order. The quantity
ordered for this feature is generatedd by eConfig.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 60 (Initial order maximum: 60)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9465) - Qty Indicator

This quantity indicator is used to specify the remaining, or N-1
quantity of CFR entities that need to be accumulated for rejoining.
The quantity ordered for this feature is generated by eConfig. and is
equal to N-1, where 'N' equals the total quantity of CFRs being
rejoined.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9466) - Countable Member Indicator

This administrative indicator used to identify each CFR associated
with a date/time stamp that is eligible for splitting and rejoining.
The quantity ordered for this feature is generated by eConfig.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9700) - Language Group Specify - Dutch

Dutch language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9703) - Language Group Specify - French

French language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9704) - Language Group Specify - German

German language group for Nomenclature and Standard Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9705) - Language Group Specify - Polish

Polish language group for Nomenclature and Standard Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9706) - Language Group Specify - Norwegian

Norwegian language group for Nomenclature and Standard
Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9707) - Language Group Specify - Portuguese

Portuguese language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9708) - Language Group Specify - Spanish

Spanish language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9711) - Language Group Specify - Italian

Italian language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9712) - Language Group Specify - Canadian French

Canadian French language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9714) - Language Group Specify - Japanese

Japanese language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9715) - Language Group Specify - Traditional Chinese (Taiwan)

Traditional Chinese language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9716) - Language Group Specify - Korean

Korean language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9718) - Language Group Specify - Turkish

Turkish language group for nomenclature and publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9719) - Language Group Specify - Hungarian

Hungarian language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9720) - Language Group Specify - Slovakian

Slovakian language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9721) - Language Group Specify - Russian

Russian language group for nomenclature and standard publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9722) - Language Group Specify - Simplified Chinese (PRC)

Simplified Chinese language group for nomenclature and standard
publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9724) - Language Group Specify - Czech

Czech language group for nomenclature and standard publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9725) - Language Group Specify -- Romanian

Romanian language group for Nomenclature and Standard Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9726) - Language Group Specify - Croatian

Croatian language group for Nomenclature and Standard
Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9727) - Language Group Specify -- Slovenian

Slovenian language group for Nomenclature and Standard
Publications.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9728) - Language Group Specify - Brazilian Portuguese

Brazilian Portuguese language group for Nomenclature and Standard
Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#9729) - Language Group Specify - Thai

Thai language group for Nomenclature and Standard Publications.

Attributes provided: Language specify

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#EBA5) - HVDC PDU Horizontal Mounting

(No Longer Available as of May 13, 2016)

Specify feature communicates to IBM Manufacturing that a HVDC DPU
such as feature #EPAA or feature #EPAE should be horizontally mounted
vs vertically mounted. PDU uses 1U rack space when horizontally
mounted, but makes accessing power cords much easier than in a side
pocket.

Attributes provided: Manufacturing communication information

Attributes required: HVDC PDU

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EBAD) - DC Power Chunnels

Left and right DC Power Chunnels and four cords to power supplies.

Attributes provided: DC Power Chunnels and four power supply cords.

Attributes required: None

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB27) - QSFP+ 40GBase-SR Transceiver

IBM QSFP+ optical tranceiver required for 40 Gbs ports which are
not using copper QSFP+ transceiver.

Attributes provided: QSFP+ tranceiver for 40 Gbs ports.

Attributes required: None

Minimum required: 0

Maximum allowed: 122 (Initial order maximum: 122)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

Supports adapters #EC3A and #EC3B

Optical cables for the transceiver are: #EB2J and #EB2K

Copper cables for the transceiver are: #EB2B, #EB2H and ECBN

(#EB2B) - 1m (3.3-ft), IBM Passive QSFP+ to QSFP+ Cable (DAC)

QSFP+ cable is used for 40Gb-to-40Gb Ethernet connectivity.
Clients can use this QSFP+ Direct Attach Cable for Ethenet
connectivity. This copper cable has QSFP+ transceivers already
attached to each end.

Attributes provided: 1m QSFP+ to QSFP+ Cable

Attributes required: QSFP/QSFP+ ports

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB2H) - 3m (9.8-ft), IBM Passive QSFP+ to QSFP+ Cable (DAC)

QSFP+ cable is used for 40Gb-to-40Gb Ethernet connectivity.
Clients can use this QSFP+ Direct Attach Cable for Ethenet
connectivity. This copper cable has QSFP+ transceivers already
attached to each end.

Attributes provided: 3m QSFP+ to QSFP+ Cable

Attributes required: QSFP/QSFP+ ports

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB2J) - 10m (30.3-ft), IBM Passive QSFP+ MTP Optical Cable

QSFP+ cable is used for 40Gb-to-40Gb Ethernet connectivity.
Clients can use this QSFP+ Direct Attach Cable for Ethenet
connectivity.

Attributes provided: 10m QSFP+ to QSFP+ Cable

Attributes required: QSFP/QSFP+ ports with optical transceivers

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB2K) - 30m (90.3-ft), IBM Passive QSFP+ MTP Optical Cable

QSFP+ cable is used for 40Gb-to-40Gb Ethernet connectivity.
Clients can use this QSFP+ Direct Attach Cable for Ethenet
connectivity.

Attributes provided: 30m QSFP+ to QSFP+ Cable

Attributes required: QSFP/QSFP+ ports with optical transceivers

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB2R) - Single 5250 Enterprise Enablement

This feature provides one processor core worth of 5250 OLTP capacity
which can be spread across multiple physical processor cores or
multiple partitions. A permanently activated processor core and
adequate IBM i processor license entitlements are prerequisites.

Attributes provided: One Processor of 5250 Enterprise Enablement

Attributes required: IBM i environment

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB2Z) - Lift Tool

A lift tool raises and lowers servers and I/O drawers so they can be
placed into or removed from standard 19-inch racks. It allows heavier
equipment to be handled more safely by fewer people. #EB27 has a hand
crank to lift and position up to 350 pounds or 159 kg. The #EB2Z
length and width are 44 inches x 24.5 inches or 1.12 meters x 0.62
meters. It has rollers which allow it to be moved to different racks
in the data center.

(#EB3J) - CBU for Power Enterprise Systems specify

This specify code indicates this system has been properly
registered as a Capacity BackUp system. and has, through that
registration been authorized to temporarily receive AIX and IBM i
Operating System License Entitlements and 5250 Processor Enablement
entitlements, from a primary system under the conditions specified at
the time the system was registered. This feature is an indicator
only. Authorization to use this system as a backup is obtained only
by registering the system with IBM on the
CBU website
at

This specify code is also associated with no-charge system nodes
and no-charge Elastic CoD processor and memory days.

Attributes provided: Indicates the system has been registered for
use as a CBU system

Attributes required: PowerHA for AIX and/or IBM i

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#EB3Z) - Lift tool based on GenieLift GL-8 (standard)

Feature #EB3Z is a feature that is available on multiple server
types (POWER S812L, S822L, S824L, S814, S824, S812, S822, E850C, E850,
E880C, E870C, E880, and E870, also the rack models 7965-S42, 7014-T00,
and 7014-T42). Failure to have at least one Lift tool available in a
location may result in delayed or prolonged maintenance times.

A lift tool raises and lowers servers and I/O drawers so they can
be placed into or removed from standard 19-inch racks. It allows
heavier equipment to be handled more safely by fewer people. Lift
tool feature EB3Z has a hand crank to lift and position up to 181 kg
(400 lbs). The lift tool feature EB3Z operating length and width are
88.3 cm x 62.9 cm (34 3/4 x 24 3/4 in). It has rollers which allow it
to be moved to different racks in the data center.

Attributes provided: Lift Tool

Attributes required: None

For 9119-MHE: (#EB3Z)

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB4Z) - Service wedge shelf tool kit for EB3Z

This feature provides a separate tool kit to replace the flat shelf
with a wedge/angle shelf at the client site.

Note: EB4Z wedge shelf is IBM SSR use only (due to safety labels/
instructions/certifications only for IBM and not filed for clients).
A client can order feature EB4Z to ensure the tool is conveniently
located on site in case an IBM SSR needed to use it and do not want to
wait for the SSR to locate and bring in an EB4Z or to schedule
additional personnel to manually handle server installation/removal
from the rack.

Client is free to use EB3Z (without EB4Z) for their normal work.

Attributes provided: Wedge/angle shelf

Attributes required: Feature EB3Z

For 9119-MHE: (#EB4Z)

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: No

Return parts MES: No

(#EB50) - 0.5M EDR IB Copper Cable QSFP28

0.5 meter length copper twinax cable, also called a DAC (Direct
Attached Copper) cable. The cable can be used for either 100Gb EDR
InfiniBand (IB) adapters or IB switches. Built onto each end of the
cable is a passive Quad (4-channel) Small Form-Factor Pluggable
(QSFP28) copper cable transceiver. Cable can also be used for FDR IB
(56Gb).

Cables are available in various lengths: 0.5M = #EB50, 1M = #EB51,
2M = #EB52, 1.5M = #EB54 See also optical fiber cables for longer
lengths such as #EB5A through #EB5H.

Attributes provided: Copper twinax cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB51) - 1.0M EDR IB Copper Cable QSFP28

1.0 meter length copper twinax cable, also called a DAC (Direct
Attached Copper) cable. The cable can be used for either 100Gb EDR
InfiniBand (IB) adapters or IB switches. Built onto each end of the
cable is a passive Quad (4-channel) Small Form-Factor Pluggable
(QSFP28) copper cable transceiver. Cable can also be used for FDR IB
(56Gb).

Cables are available in various lengths: 0.5M = #EB50, 1M = #EB51,
2M = #EB52, 1.5M = #EB54 See also optical fiber cables for longer
lengths such as #EB5A through #EB5H.

Attributes provided: Copper twinax cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB52) - 2.0M EDR IB Copper Cable QSFP28

2.0 meter length copper twinax cable, also called a DAC (Direct
Attached Copper) cable. The cable can be used for either 100Gb EDR
InfiniBand (IB) adapters or IB switches. Built onto each end of the
cable is a passive Quad (4-channel) Small Form-Factor Pluggable
(QSFP28) copper cable transceiver. Cable can also be used for FDR IB
(56Gb).

Cables are available in various lengths: 0.5M = #EB50, 1M = #EB51,
2M = #EB52, 1.5M = #EB54 See also optical fiber cables for longer
lengths such as #EB5A through #EB5H.

Attributes provided: Copper twinax cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB54) - 1.5M EDR IB Copper Cable QSFP28

1.5 meter length copper twinax cable, also called a DAC (Direct
Attached Copper) cable. The cable can be used for either 100Gb EDR
InfiniBand (IB) adapters or IB switches. Built onto each end of the
cable is a passive Quad (4-channel) Small Form-Factor Pluggable
(QSFP28) copper cable transceiver. Cable can also be used for FDR IB
(56Gb).

Cables are available in various lengths: 0.5M = #EB50, 1M = #EB51,
2M = #EB52, 1.5M = #EB54 See also optical fiber cables for longer
lengths such as #EB5A through #EB5H.

Attributes provided: Copper twinax cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB59) - 100Gb Optical Transceiver QSFP28

One optical transceiver for 100Gb Ethernet adapter such as #EC3L
or #EC3M using QSFP28. Does not include cable.

See also AOC fiber cables which include QSFP28 transceivers ER5R -
EB5Y.

Attributes provided: Optical Transceiver QSFP28 100Gb.

Attributes required: Port on adapter with QSFP28 socket.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5A) - 3M EDR IB Optical Cable QSFP28

3 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5B) - 5M EDR IB Optical Cable QSFP28

5 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5C) - 10M EDR IB Optical Cable QSFP28

10 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5D) - 15M EDR IB Optical Cable QSFP28

15 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5E) - 20M EDR IB Optical Cable QSFP28

20 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5F) - 30M EDR IB Optical Cable QSFP28

30 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5G) - 50M EDR IB Optical Cable QSFP28

50 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5H) - 100M EDR IB Optical Cable QSFP28

100 meter length optical fiber cable, also called an AOC (Active
Optical Cable). The cable can be used for either 100Gb EDR InfiniBand
(IB) adapters or IB switches. Built onto each end of the cable is a
active Quad (4-channel) Small Form-Factor Pluggable (QSFP28) optical
cable transceiver.

Limitation: Adapter and switch must have been manufactured by
Mellanox to use this cable. Cable is not supported for FDR IB (56Gb).

Attributes provided: Optical fiber cable with QSFP28 transceivers

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EB5J) - 0.5M 100GbE Copper Cable QSFP28

0.5 meter length passive copper cable with QSFP28 tranceivers on
each end. The cable used for 100Gb Ethernet adapters or switches.
Copper cable is also called "copper twinax" or "DAC" (Direct Attach
Copper).

(#EB5K) - 1.0M 100GbE Copper Cable QSFP28

1.0 meter length passive copper cable with QSFP28 tranceivers on
each end. The cable used for 100Gb Ethernet adapters or switches.
Copper cable is also called "copper twinax" or "DAC" (Direct Attach
Copper).

(#EB5L) - 1.5M 100GbE Copper Cable QSFP28

1.5 meter length passive copper cable with QSFP28 tranceivers on
each end. The cable used for 100Gb Ethernet adapters or switches.
Copper cable is also called "copper twinax" or "DAC" (Direct Attach
Copper).

(#EB5M) - 2.0M 100GbE Copper Cable QSFP28

2.0 meter length passive copper cable with QSFP28 tranceivers on
each end. The cable used for 100Gb Ethernet adapters or switches.
Copper cable is also called "copper twinax" or "DAC" (Direct Attach
Copper).

(#EBA3) - IBM Rack-mount Drawer Bezel and Hardware

Indicates that this order is for a rack-mount system requiring an
IBM bezel and hardware.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EBA4) - OEM Rack-mount Drawer Bezel and Hardware

Indicates that this order is for a rack-mount system requiring an
OEM bezel and hardware.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EBAA) - AC Power Chunnels

Left and right AC Power Chunnels and four cords to power supplies.

Attributes provided:

AC Power Chunnels and four power supply cords.

Attributes required: None

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EBK4) - 1.6M USB Cable

1.6M USB CABLE, MALE-TO-MALE -
This cable is only used to attach a #EU13 DVD located in the system
control unit.

Attributes provided: 1.6M USB CABLE

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EC01) - Rack Front Door (Black)

This feature provides a front door in flat black color with an IBM
logo for the 7953-94X rack. A front door such as #EC01 is required on
the 7953-94X. The door is the full width of the rack and the hinges
and lockplate can be moved from side to side allowing the door to be
opened on the left or on the right. IBM ships rack with the handle on
the right and hinges on the left viewed facing the front of the rack.
The door comes with a lock which is keyed the same as the rear door or
side panels. Uniquely keyed locks can be obtained by the client
directly from Southco, the vendor from whom IBM purchased the lock.

Attributes provided: Front Door with lock

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC02) - Rack Rear Door

This feature provides a rear door in flat black color for the
7953-94X rack. Either feature number EC02 or feature EC05 is required
on the 7953-94X. The door is the full width of the rack and the
hinges and lockplate can be moved from side to side allowing the door
to be opened on the left or on the right. IBM ships rack with the
handle on the right and hinges on the left viewed facing the rear of
the rack. The front doors, rear doors and side panels come with a
lock which is keyed the same as the front door or side panels.
Uniquely keyed locks can be obtained by the client directly from
Southco, the vendor from whom IBM purchased the lock.

Attributes provided: Rear Door with lock

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC03) - Rack Side Cover

This feature provides two side panels in black color for the
7953-94X rack. Each side panel can cover either the left or the right
side of the rack. These side covers are optional but recommended for
optimal airflow through a rack and for physical security. The front
door, rear doors and side panels come with a lock which is keyed the
same as the front door or side panels. Uniquely keyed locks can be
obtained by the client directly from Southco, the vendor from whom IBM
purchased the lock.

Attributes provided: Left and Right side panels for 7953 rack.

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC04) - Rack Suite Attachment Kit

This feature provides the attachment hardware to allow two IBM
PureFlex System 42U Racks with or without side panels to be bolted
together in a continuous suite. Order one EC04 feature for each 42U
rack attached to the starting rack. For example, in a three-rack
suite, order two EC04 features. When multiple racks are joined in
this way without internal side panels, cables can be more easily run
between racks without having to exit the continuous rack enclosure.
With the side panels installed optimum thermal efficiencies are
gained. If the optional rear door heat exchanger is chosen, side
panels must remain on racks in the suite. Side panels should be used
on the leftmost and rightmost racks of the suite.

Attributes provided: Hardware and trim to attach two racks

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC07) - Slim Rear Acoustic Door

Slim rear acoustic door for use on the 7014-T42 or server feature
#0553 2. 0M racks. Depth of this acoustic door is only about 113 mm
(4.45 inches).

Physically #EC07 and #EC08 are identical, but have two feature
codes to assist IBM sales configurator logic. #EC07 designates a rear
door and #EC08 designates a front door.

Attributes provided: Rear Acoustic door

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC08) - Slim Front Acoustic Door

Slim front acoustic door for use on the 7014-T42 or server feature
#0553 2. 0M racks. Depth of this acoustic door is only about 113 mm
(4.45 inches).

Physically #EC07 and #EC08 are identical, but have two feature
codes to assist IBM sales configurator logic. #EC07 designates a rear
door and #EC08 designates a front door.

Attributes provided: Acoustic front door

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC15) - Rear Door Heat Exchanger for 2.0 Meter Slim Rack

This feature indicates that the rear door heat exchanger (1164-95X)
is ordered for the 7953-94Y rack. Either feature EC02 or feature EC05
is required on the 7953-94Y.

Attributes provided: RDHX

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EC19) - CAPI Activation

This activation feature enables the use of the Coherent Accelerator
Processor Interface (CAPI) technology capabilities on POWER8
processor-based Power Systems using industry accelerator technologies.
This can increase performance through faster computational efficiency and
by requiring fewer resources to accomplish tasks through direct
integration of workload accelerator(s) in the system. When using this
optional feature (#EC19) one is required per server. The CAPI activation
feature is generic and can be used with separately provided CAPI
application or hardware from IBM or OEM.

Attributes provided: activation/enablement

Attributes required: None for enablement, but use of CAPI will
also require CAPI-aware hardware and applications

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: Max of one per socket

(#EC27) - PCIe2 LP 2-Port 10GbE RoCE SFP+ Adapter

(No Longer Available as of January 20, 2017)

This PCIe Gen2 adapter provides two 10 Gb copper SFP+ ports and
supports the IBTA RoCE standard. RoCE is Remote Direct Memory Access
(RDMA) over Converged Ethernet. The adapter can support significantly
greater bandwidth with low latency. It minimizes CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

Active Copper twinax cables up to 5 meter in length are supported
such as provided by feature #EN01, #EN02 or #EN03. Note that SFP+
twinax copper is NOT AS/400 5250 twinax or CX4 or 10 GBASE-T. Active
cables differ from passive cables.

Feature code #EC27 and #EC28 have identical electronics and
function and CCIN (EC27), but have different tail stocks. #EC27 is
low profile and #EC28 is full high. Compared to Feature #EC29/EC30,
the #EC27/EC28 have identical application capability and differ
electronically only in that #EC27/EC28 have a SFP+ copper interface
and a different CCIN where the #EC29/EC30 have an SFP+ SR optical
interface.

LIMITATIONS:

AIX NIM and Linux Network Install are not supported

This adapter supports RoCE and NIC functions but not concurrently
on same adapters.

Note: As of 3/12/2013 This feature is not available in the following EMEA
countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait, Lebanon, Libya,
Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi Arabia, Somalia,
Tunisia and United Arab Emirates which includes IBM Middle East -
Dubai and Abu Dhabi and Yemen.

Attributes provided: Dual-port 10GbE Adapter with RoCE capability

Attributes required: Available GEN2 PCIe Slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX supported

IBM i not supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

SUSE Linux Enterprise 12, or later

Ubuntu Server 16.04, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: 8 per CEC Max.

Note: VIOS supported NIC capability only.

(#EC28) - PCIe2 2-Port 10GbE RoCE SFP+ Adapter

(No Longer Available as of January 20, 2017)

This PCIe Gen2 adapter provides two 10 Gb copper SFP+ ports and
supports the IBTA RoCE standard. RoCE is Remote Direct Memory Access
(RDMA) over Converged Ethernet. The adapter can support significantly
greater bandwidth with low latency. It minimizes CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

Active Copper twinax cables up to 5 meter in length are supported
such as provided by feature #EN01, #EN02 or #EN03. Note that SFP+
twinax copper is NOT AS/400 5250 twinax or CX4 or 10 GBASE-T. Active
cables differ from passive cables.

Feature code #EC27 and #EC28 have identical electronics and
function and CCIN (EC27), but have different tail stocks. #EC27 is
low profile and #EC28 is full high. Compared to Feature #EC29/EC30,
the #EC27/EC28 have identical application capability and differ
electronically only in that #EC27/EC28 have a SFP+ copper interface
and a different CCIN where the #EC29/EC30 have an SFP+ SR optical
interface.

LIMITATIONS:

AIX NIM and Linux Network Install are not supported

This adapter supports RoCE and NIC functions but not concurrently
on same adapters.

Note: As of 3/12/2013 This feature is not available in the following EMEA
countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait, Lebanon, Libya,
Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi Arabia, Somalia,
Tunisia and United Arab Emirates which includes IBM Middle East -
Dubai and Abu Dhabi and Yemen.

Attributes provided: Dual-port 10GbE Adapter with RoCE capability

Attributes required: Available GEN2 PCIe Slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i requires VIOS. Only standard NIC is supported.

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS is supported for the ethernet capability.

(#EC29) - PCIe2 LP 2-Port 10GbE RoCE SR Adapter

(No Longer Available as of February 23, 2017)

This PCIe Gen2 adapter provides two 10 Gb SR optical ports and
supports the IBTA RoCE standard. RoCE is Remote Direct Memory Access
(RDMA) over Converged Ethernet. The adapter can support significantly
greater bandwidth with low latency. It minimizes CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

The adapter includes a pre-installed Optical Transceiver in the
adapter. LC connectors connect to standard 10Gb SR optical cabling
and provide up to 300m cable length.

Feature code #EC29 and #EC30 have identical electronics and
function and CCIN (EC29), but have different tail stocks. #EC29 is
low profile and #EC30 is full high, Compared to Feature #EC27/EC28,
the #EC29/EC30 have identical application capability and differ
electronically only in that #EC27/ EC28 have a SFP+ copper interface
and a different CCIN where the #EC29/EC30 have an SFP+ SR optical
interface.

LIMITATIONS:

AIX NIM and Linux Network Install are not supported

This adapter supports RoCE and NIC functions but not concurrently
on same adapters.

Note: As of 3/12/2013 This feature is not available in the following EMEA
countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait, Lebanon, Libya,
Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi Arabia, Somalia,
Tunisia and United Arab Emirates which includes IBM Middle East -
Dubai and Abu Dhabi and Yemen.

Attributes provided: Dual-port 10GbE Adapter with RoCE capability

Attributes required:

Available GEN2 PCIe Slot

Firmware Level 740 or greater

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i requires VIOS. Only standard NIC is supported.

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported.

(#EC2G) - PCIe2 LP 2-port 10GbE SFN6122F Adapter

(No Longer Available as of January 20, 2017)

This PCIe Gen2 2-port 10Gb Ethernet adapter is manufactured by
SolarFlare and supports their OpenOnload(R) application accelerator.
The adapter is supported by Linux and can be used by user-provided
applications written to leverage its interfaces.

This PCIe Gen2 adapter provides two 10 Gb copper SFP+ ports.
Active Copper twinax cables up to 5 meter in length are supported such
as provided by feature #EN01, #EN02 or #EN03. #EN01/EN02/EN03 cables
include transceivers for SFP+ ports. Note that SFP+ twinax copper is
NOT AS/400 5250 twinax or CX4 or 10 GBASE-T. Active cables differ
from passive cables

#EC2G and #EC2J are electronically identical adapters, but #EC2G
indicates a low profile tail stock and #EC2J indicates a full high
tail stock.

See also #EC2H and #EC2K which is also manufactured by SolarFlare,
but does not support the OpenOnload capability.

(#EC2J) - PCIe2 2-port 10GbE SFN6122F Adapter

(No Longer Available as of January 20, 2017)

This PCIe Gen2 2-port 10Gb Ethernet adapter is manufactured by
SolarFlare and supports their OpenOnload application accelerator.
The adapter is supported by Linux and can be used by user-provided
applications written to leverage its interfaces.

This PCIe Gen2 adapter provides two 10 Gb copper SFP+ ports.
Active Copper twinax cables up to 5 meter in length are supported such
as provided by feature #EN01, #EN02 or #EN03. #EN01/EN02/EN03 cables
include transceivers for SFP+ ports. Note that SFP+ twinax copper is
NOT AS/400 5250 twinax or CX4 or 10 GBASE-T. Active cables differ
from passive cables

#EC2G and #EC2J are electronically identical adapters, but #EC2G
indicates a low profile tail stock and #EC2J indicates a full high
tail stock.

See also #EC2H and #EC2K which is also manufactured by SolarFlare,
but does not support the OpenOnload capability.

(#EC2M) - PCIe3 LP 2-port 10GbE NIC&RoCE SR Adapter

This PCIe Gen3 adapter provides two 10 Gb SR optical fiber ports.
The adapter supports both NIC and IBTA RoCE standards. RoCE is Remote
Direct Memory Access (RDMA) over Converged Ethernet. Using RoCE, the
adapter can support significantly greater bandwidth with low latency
and minimize CPU overhead by more efficiently using memory access.
This offloads the CPU from I/O networking tasks, improving performance
and scalability.

The adapter includes two pre-installed Optical Transceivers in the
adapter. LC connectors connect to standard 10Gb SR optical cabling
and provide up to 300m cable length.

Feature code #EC2M and #EC2N have identical electronics and
function and CCIN (57BE), but have different tail stocks. #EC2M is
low profile and #EC2N is full high. Compared to EC37/EC38, the EC2M/
EC2N have identical application capability, but different cabling
(optical fiber vs copper twinax). The EC2M/EC2N is based on the
Mellanox ConnectX-3 Pro ASIC and is newer technology than previous
features EC29/EC30.

LIMITATION: This adapter supports RoCE and NIC functions but not
concurrently on same adapter.

(#EC2N) - PCIe3 2-port 10GbE NIC&RoCE SR Adapter

This PCIe Gen3 adapter provides two 10 Gb SR optical fiber ports.
The adapter supports both NIC and IBTA RoCE standards. RoCE is Remote
Direct Memory Access (RDMA) over Converged Ethernet. Using RoCE, the
adapter can support significantly greater bandwidth with low latency
and minimize CPU overhead by more efficiently using memory access.
This offloads the CPU from I/O networking tasks, improving performance
and scalability.

The adapter includes two pre-installed Optical Transceivers in the
adapter. LC connectors connect to standard 10Gb SR optical cabling
and provide up to 300m cable length.

Feature code #EC2M and #EC2N have identical electronics and
function and CCIN (57BE), but have different tail stocks. #EC2M is
low profile and #EC2N is full high. Compared to EC37/EC38, the EC2M/
EC2N have identical application capability, but different cabling
(optical fiber vs copper twinax). The EC2M/EC2N is based on the
Mellanox ConnectX-3 Pro ASIC and is newer technology than previous
features EC29/EC30.

LIMITATION: This adapter supports RoCE and NIC functions but not
concurrently on same adapter.

(#EC30) - PCIe2 2-Port 10GbE RoCE SR Adapter

This PCIe Gen2 adapter provides two 10 Gb SR optical ports and
supports the IBTA RoCE standard. RoCE is Remote Direct Memory Access
(RDMA) over Converged Ethernet. The adapter can support significantly
greater bandwidth with low latency. It minimizes CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

The adapter includes a pre-installed Optical Transceiver in the
adapter. LC connectors connect to standard 10Gb SR optical cabling
and provide up to 300m cable length.

Feature code #EC29 and #EC30 have identical electronics and
function and CCIN (EC29), but have different tail stocks. #EC29 is
low profile and #EC30 is full high, Compared to Feature #EC27/EC28,
the #EC29/EC30 have identical application capability and differ
electronically only in that #EC27/EC28 have a SFP+ copper interface
and a different CCIN where the #EC29/EC30 have an SFP+ SR optical
interface.

Attributes provided: Dual-port 10GbE Adapter with RoCE capability

Attributes required:

Available GEN2 PCIe Slot

Firmware Level 740 or greater

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i requires VIOS. Only standard NIC is supported.

Refer to Software requirements for specific O/S levels supported

Limitations:

AIX NIM and Linux Network Install are not supported

This adapter supports RoCE and NIC functions but not concurrently
on same adapters.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: As of 3/12/2013 This feature is not available in the following EMEA
countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait, Lebanon, Libya,
Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi Arabia, Somalia, Tunisia
and United Arab Emirates which includes IBM Middle East - Dubai and Abu
Dhabi and Yemen.

Note: VIOS is supported for the ethernet capability.

(#EC37) - PCIe3 LP 2-port 10GbE NIC&RoCE SFP+ Copper Adapter

This PCIe Gen3 adapter provides two 10 Gb SFP+ ports for copper
twinax cabling/transceivers. The adapter supports both NIC and IBTA
RoCE standards. RoCE is Remote Direct Memory Access (RDMA) over
Converged Ethernet. Using RoCE, the adapter can support significantly
greater bandwidth with low latency and minimize CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

Active Copper twinax cables up to 5 meter in length are supported
such as provided by feature #EN01, #EN02 or #EN03. These cables also
include copper transceivers. Active cables differ from passive
cables.

Feature code #EC37 and #EC38 have identical electronics and
function and CCIN (57BC), but have different tail stocks. #EC37 is
low profile and #EC38 is full high. Compared to EC2M/EC2N, the EC37/
EC38 have identical application capability, but different cabling
(optical fiber vs copper twinax). The EC37/EC38 is based on the
Mellanox ConnectX-3 Pro ASIC and is newer technology than previous
features EC27/EC28.

LIMITATION: This adapter supports RoCE and NIC functions but not
concurrently on same adapter.

(#EC38) - PCIe3 2-port 10GbE NIC&RoCE SFP+ Copper Adapter

This PCIe Gen3 adapter provides two 10 Gb SFP+ ports for copper
twinax cabling/transceivers. The adapter supports both NIC and IBTA
RoCE standards. RoCE is Remote Direct Memory Access (RDMA) over
Converged Ethernet. Using RoCE, the adapter can support significantly
greater bandwidth with low latency and minimize CPU overhead by more
efficiently using memory access. This offloads the CPU from I/O
networking tasks, improving performance and scalability.

Active Copper twinax cables up to 5 meter in length are supported
such as provided by feature #EN01, #EN02 or #EN03. These cables also
include copper transceivers. Active cables differ from passive
cables.

Feature code #EC37 and #EC38 have identical electronics and
function and CCIN (57BC), but have different tail stocks. #EC37 is
low profile and #EC38 is full high. Compared to EC2M/EC2N, the EC37/
EC38 have identical application capability, but different cabling
(optical fiber vs copper twinax). The EC37/EC38 is based on the
Mellanox ConnectX-3 Pro ASIC and is newer technology than previous
features EC27/EC28.

LIMITATION: This adapter supports RoCE and NIC functions but not
concurrently on same adapter.

(#EC3A) - PCIe3 LP 2-Port 40GbE NIC RoCE QSFP+ Adapter

RoCE is Remote Direct Memory Access (RDMA) over Converged Ethernet.
RoCE can support significantly greater bandwidth with low latency and
minimize CPU overhead by more efficiently using memory access. This
offloads the CPU from I/O networking tasks, improving performance and
scalability.

The #EC3A/EC3B adapter does not include transceivers. Shorter
distance copper cables include transceivers on the end of the QSFP+
cable (see #EB2B (1m), #EB2H (3m), #ECBN (5m)). For longer distance
use two optical SR transceiver (two #EB27). QSFP+ optical cables to
be used with the #EB27 transceiver are #EB2J (10m) or #EB2K (30m). Do
not mix copper and optical on the same adapter.

Feature code #EC3A and #EC3B have identical electronics and function
and the same CCIN (57BD), but they have different tail stocks. #EC3A
is low profile and #EC3B is full height.

AIX NIM and Linux Network Install are supported.

Limitation: As of April 2014, this feature is not available in the
following EMEA countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait,
Lebanon, Libya, Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi
Arabia, Somalia, Tunisia and United Arab Emirates which includes IBM
Middle East - Dubai and Abu Dhabi and Yemen.

(#EC3B) - PCIe3 2-Port 40GbE NIC RoCE QSFP+ Adapter

RoCE is Remote Direct Memory Access (RDMA) over Converged Ethernet.
RoCE can support significantly greater bandwidth with low latency and
minimize CPU overhead by more efficiently using memory access. This
offloads the CPU from I/O networking tasks, improving performance and
scalability.

The #EC3A/EC3B adapter does not include transceivers. Shorter
distance copper cables include transceivers on the end of the QSFP+
cable (see #EB2B (1m), #EB2H (3m), #ECBN (5m)). For longer distance
use two optical SR transceiver (two #EB27). QSFP+ optical cables to
be used with the #EB27 transceiver are #EB2J (10m) or #EB2K (30m). Do
not mix copper and optical on the same adapter.

Feature code #EC3A and #EC3B have identical electronics and function
and the same CCIN (57BD), but they have different tail stocks. #EC3A
is low profile and #EC3B is full height.

AIX NIM and Linux Network Install are supported.

Limitation: As of April 2014, this feature is not available in the
following EMEA countries, Bahrain, Comoros, Djibouti, Iraq, Kuwait,
Lebanon, Libya, Malaysia, Morocco, Oman, Pakistan, Qatar, Saudi
Arabia, Somalia, Tunisia and United Arab Emirates which includes IBM
Middle East - Dubai and Abu Dhabi and Yemen.

(#EC3E) - PCIe3 LP 2-port 100Gb EDR IB Adapter x16

The PCIe Gen3 x16 2-port InfiniBand EDR adapter provides high
speed connectivity with other servers or IB switches. Each port
maximum of 100Gb assumes no other system and/or switch bottlenecks
are present. A PCIe Gen3 x16 PCIe slot is required. This adapter is
sourced from Mellanox Corporation. based on ConnectX-4 technology.
The adapter supports the InfiniBand Trade Association (IBTA)
specification version 2.

The two 100Gb ports have QSFP+ connections which support EDR
cables, either EDR DAC or EDR optical. One adapter can support either
or both types of cable. The user can choose to cable up just one port
if they desire. Transceivers are included in the cables. IBM cable
features EB50-EB54 (copper shorter distance) and #EB5A-EB5H (optical
longer distance) are supported or their copper or optical Mellanox
equivalents are supported. Other cables are not supported.

#EC3E and #EC3F adapters are electronically and functionally
identical with the same CCIN of 2CEA. #EC3E has a low profile
tailstock bracket. #EC3F has a full high tailstock bracket. See also
#EC3T and #EC3U for a 1-port version of this adapter. Limitation:
Adapter does not fit in x8 PCIe slot. Adapter is not supported in a
PCIe Gen3 I/O drawer.

Attributes provided: EDR InfiniBand PCIe Adapter

Attributes required: available x16 PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 16 (Initial order maximum: 16)

OS level required:

Red Hat Enterprise Linux 7.2, or later, with Mellanox OFED 3.3,
or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later, with
Mellanox OFED 3.3, or later

SUSE Linux Enterprise Server 12, Service Pack 1, or later, with
Mellanox OFED 3.3, or later

Ubuntu Server 14.04.4, or later, with Mellanox OFED 3.3, or later

Ubuntu Server 16.04, or later, with Mellanox OFED 3.3, or later

AIX not supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: Limitation: Not supported by VIOS.

(#EC3L) - PCIe3 LP 2-port 100GbE (NIC& RoCE) QSFP28 Adapter x16

This PCIe Gen3 Ethernet x16 adapter provides two 100 Gb QSFP28
ports. The adapter supports both NIC and IBTA RoCE standards. RoCE is
Remote Direct Memory Access (RDMA) over Converged Ethernet. Using
RoCE, the adapter can support significantly greater bandwidth with
low latency and minimize CPU overhead by more efficiently using
memory access. This offloads the CPU from I/O networking tasks,
improving performance and scalability.

IBM offers either passive copper twinax cables up to 2 meter in
length or active optical cables up to 100 meters in length. See
features #EB5J - #EB5M for a 0.5M, 1.0M, 1.5M and 2.0M copper cable.
See features #EB5R - #EB5Y for a 3M, 5M, 10M, 15M, 20M, 30M, 50M or
100M active optical cable. Transceivers are included on each end of
these QSFP28 cables. Alternatively to the above supported cables, you
may chose to order an IBM qualified and supported QSFP28 optical
transceiver (feature #EB59) to put into the adapter and provide your
own 100GE optical cabling with your own QSP28 optical transceiver for
the other end.

Either one or both of the adapter's two QSP28 ports can be
populated. When two ports are filled, both can have copper cables,
both can have optical cables, or one can be copper and one can be
optical.

Feature code #EC3L and #EC3M have identical electronics and
function and CCIN (2CEC), but have different tail stock brackets.
#EC3L is low profile and #EC3M is full high. The adapter is based on
a Mellanox ConnectX-4 adapter which uses a ConnectX-4 EN Network
Controller. Attributes :

PCI Express 3.0 (up to 8GT/s) x16

PCIe Gen 3.0 compliant, 1.1 and 2.0 compatible

RDMA over Converged Ethernet (RoCE)

NIC and RoCE are concurrently supported

RoCE supported on Linux and AIX (7.2 and later)

NIC supported on all OSes

TCP/UDP/IP stateless offload

LSO, LRO, checksum offload

NIM boot support - Backward compatible with 40Gb Ethernet when
using compatible cables/transceivers.

Attributes provided: 2-port 100Gb Ethernet

Attributes required: x16 PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX Version 7.2 with the 7200-01 Technology Level or later

AIX Version 7.1 with the 7100-04 Technology Level and Service
Pack 3 and APAR IV88680 or later

AIX Version 6.1 with the 6100-09 Technology Level and Service
Pack 8 and APAR IV88679 or later

AIX Version 7.1 with the 7100-03 Technology Level and Service
Pack 8 or later (planned availability 1/27/2017)

AIX Version 7.2 with the 7200-00 Technology Level and Service
Pack 3 or later (planned availability 1/27/2017)

IBM i supported only with VIOS

Note: AIX 7.2 with the 7200-01 Technology level or later supports
both Ethernet and RoCE capability. AIX 6.1, 7.1 and VIOS support
Ethernet capability only.

(#EC3T) - PCIe3 LP 1-port 100Gb EDR IB Adapter x16

The PCIe Gen3 x16 1-port InfiniBand EDR adapter provides high
speed connectivity with other servers or IB switches. The port's
maximum of 100Gb assumes no other system and/or switch bottlenecks
are present. A PCIe Gen3 x16 PCIe slot is required. This adapter is
sourced from Mellanox Corporation. based on ConnectX-4 technology.
The adapter supports the InfiniBand Trade Association (IBTA)
specification version 2.

The 100Gb port has a QSFP+ connection which supports EDR cables,
either EDR DAC or EDR optical. Transceivers are included in the
cables. IBM cable features EB50-EB54 (copper shorter distance) and
#EB5A-EB5H (optical longer distance) are supported or their copper or
optical Mellanox equivalents are supported. Other cables are not
supported. #EC3T and #EC3U adapters are electronically and
functionally identical with the same CCIN of 2CEB. #EC3T has a low
profile tailstock bracket. Limitation: Adapter does not fit in x8
PCIe slot. Adapter is not supported in a PCIe Gen3 I/O drawer.

Attributes provided: EDR InfiniBand PCIe Adapter

Attributes required: available x16 PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 16 (Initial order maximum: 16)

OS level required:

Red Hat Enterprise Linux 7.2, or later, with Mellanox OFED 3.3,
or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later, with
Mellanox OFED 3.3, or later

SUSE Linux Enterprise Server 12, Service Pack 1, or later, with
Mellanox OFED 3.3, or later

(#EC41) - PCIe2 LP 3D Graphics Adapter x1

(No Longer Available as of December 29, 2017)

There are two identical 3D adapters EC41 with a low profile tail
stock and EC42 with full high tail stock.

When using graphic adapter for Partition Firmware Console to select
an install or boot device use #3632 display or rack mount 7316-TF4
display. Withdrawn display #3644 or withdrawn rack mount 7316-TF3
display may also be used.

Limit of one adapter per LPAR.

Limitation: Placement of this PCIe adapter is not supported in the
PCIe Gen3 I/O Drawer.

Attributes provided: Graphics adapter with DMS-59 Dual DVI cable

Attributes required: available PCIe2 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX not supported

IBM i not supported

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 12, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EC42) - PCIe2 3D Graphics Adapter x1

(No Longer Available as of November 5, 2015)

Note: Supported prior to 2015/05/11.

There are two identical 3D adapters EC41 with a low profile tail
stock and EC42 with full high tail stock.

When using graphic adapter for Partition Firmware Console to select
an install or boot device use #3632 display or rack mount 7316-TF4
display. Withdrawn display #3644 or withdrawn rack mount 7316-TF3
display may also be used.

Limit of one adapter per LPAR.

Limitation: Placement of this PCIe adapter is not supported in the
PCIe Gen3 I/O Drawer.

Attributes provided: Graphics adapter with DMS-59 Dual DVI cable

Attributes required: one available full height PCIe2 slot

Minimum required: 0

Maximum allowed: 0 (Initial order maximum: 0)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EC45) - PCIe2 LP 4-Port USB 3.0 Adapter

The PCIe Gen2 x8 short 4-port USB 3.0 adapter provides support for USB
devices. In applications that require the use of an USB extension
cable for keyboards, use one #4256 per port. The #EC45 and #EC46 USB
adapters are electroncially identical with the same 58F9 CCIN. They
differ physically in their tailstock. #EC45 is low profile and #EC46
is full high.

Attributes provided: Connectivity with USB 2.0 - 3.0 capable
devices

Attributes required: One low profile available PCIe slot.

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported.

(#EC46) - PCIe2 4-Port USB 3.0 Adapter

The PCIe Gen2 x8 short 4-port USB 3.0 adapter provides support for USB
devices. In applications that require the use of an USB extension
cable for keyboards, use one #4256 per port. The #EC45 and #EC46 USB
adapters are electronically identical with the same 58F9 CCIN. They
differ physically in their tailstock. #EC45 is low profile and #EC46
is full high.

Attributes provided: Connectivity with USB 2.0 - 3.0 capable
devices

Attributes required: One available full height PCIe slot.

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported.

(#EC54) - PCIe3 LP 1.6TB NVMe Flash Adapter

This feature ships 1.6 TB of low write latency, nonvolatile flash
memory on a PCIe Gen3 adapter. Adapter uses NVMe (Non-Volatile Memory
express) which is a high performance software interface to read/write
this flash memory. Adapter physically is half length x4 adapter which
can be used in either a x8 or x16 PCIe Gen3 slot in the system unit.
Compared to SAS/SATA SSD the NVMe adapter can provide significantly
more read or write IOPS and significantly larger throughput (GB/sec).
CCIN is 58CB. Adapter card is designed for read intensive workloads
with light write activity. Approximately 8,760 TB of data can be
written over the life of the adapter, but depending on the nature of
the workload may be somewhat larger. After the warranty period, if
the maximum write capability is achieved, the adapter's replacement
is not covered under IBM maintenance. For high value applications
where the content in the adapter must be protected, using additional
NVMe Flash adapter(s) with OS mirroring or software RAID is
recommended. #EC54 and #EC55 are identical cards except that the
tailstock bracket is different. #EC54 fits a low profile PCIe slot.
#EC55 fits a full high PCIe slot. See also #EC56/EC57 for a card with
more memory. Limitations: Not supported in PCIe Gen3 I/O drawer. Data
protection not implemented in the card and protection provided by OS
mirroring or software RAID.

Note: When using Linux OS SLES11 SP4 BE there is a maximum limitation of
12 per single LPAR.

(#EC56) - PCIe3 LP 3.2TB NVMe Flash Adapter

This feature ships 3.2 TB of write low latency, nonvolatile flash
memory on a PCIe Gen3 adapter. Adapter uses NVMe (Non-Volatile Memory
express) which is a high performance software interface to read/write
this flash memory. Adapter physically is half length x4 adapter which
can be used in either a x8 or x16 PCIe Gen3 slot in the system unit.
Compared to SAS/SATA SSD the NVMe adapter can provide significantly
more read or write IOPS and significantly larger throughput (GB/sec).
CCIN is 58CC. Adapter card is designed for read intensive workloads
with light write activity. Approximately 17,500 TB of data can be
written over the life of the adapter, but depending on the nature of
the workload may be somewhat larger. After the warranty period, if
the maximum write capability is achieved, the adapter's replacement
is not covered under IBM maintenance. For high value applications
where the content in the adapter must be protected, using additional
NVMe Flash adapter(s) with OS mirroring or software RAID is
recommended. #EC56 and #EC57 are identical cards except that the
tailstock bracket is different. #EC56 fits a low profile PCIe slot.
#EC57 fits a full high PCIe slot. See also #EC54/EC55 for a card with
less memory. Limitations: Not supported in PCIe Gen3 I/O drawer. Data
protection not implemented in the card and protection provided by OS
mirroring or software RAID.

Note: When using Linux OS SLES11 SP4 BE there is a maximum limitation of
12 per single LPAR.

(#ECBJ) - SAS X Cable 3m - HD Narrow 6Gb 2-Adapters to Enclosure

This 3 meter SAS cable connects two PCIe2 SAS adapters or two PCIe3
SAS adapters to a SAS I/O enclosure. This X cable has four
connectors, two Mini-SAS HD (High Density) Narrow connector and two
Mini SAS connectors. The Mini-SAS HD Narrow connectors attach to two
SAS adapters such as two #EJ0J or two #EJ0L or two #5913. Both Mini
SAS connectors attach to the same I/O drawer (enclosure) such as a
#5886 EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support
up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

This cable is almost identical to the #3454 3m SAS X cable, except
#ECBJ connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

(#ECBK) - SAS X Cable 6m - HD Narrow 6Gb 2-Adapters to Enclosure

This 3 meter SAS cable connects two PCIe2 SAS adapters or two PCIe3
SAS adapters to a SAS I/O enclosure. This X cable has four
connectors, two Mini-SAS HD (High Density) Narrow connector and two
Mini SAS connectors. The Mini-SAS HD Narrow connectors attach to two
SAS adapters such as two #EJ0J or two #EJ0L or two #5913. Both Mini
SAS connectors attach to the same I/O drawer (enclosure) such as a
#5886 EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support
up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

This cable is almost identical to the #3454 3m SAS X cable, except
#ECBJ connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

This 6 meter SAS cable
connects two SAS adapters to a SAS I/O enclosure. This X cable has
four connectors, two Mini-SAS HD (High Density) connector and two Mini
SAS connectors. The Mini-SAS HD connectors attaches to two SAS
adapters such as the #5913 1.8GB RAID SAS Adapter. Both Mini SAS
connectors attach to the same I/O drawer (enclosure) such as a #5886
EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support up to
6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapters can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Limitation: When this cable is ordered with a system in a rack
specifying IBM Plant integration, IBM Manufacturing will ship SAS
cables longer than 3 meters in a separate box and not attempt to
place the cable in the rack.

(#ECBL) - SAS X Cable 10m - HD Narrow 6Gb 2-Adapters to Enclosure

This 3 meter SAS cable connects two PCIe2 SAS adapters or two PCIe3
SAS adapters to a SAS I/O enclosure. This X cable has four
connectors, two Mini-SAS HD (High Density) Narrow connector and two
Mini SAS connectors. The Mini-SAS HD Narrow connectors attach to two
SAS adapters such as two #EJ0J or two #EJ0L or two #5913. Both Mini
SAS connectors attach to the same I/O drawer (enclosure) such as a
#5886 EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support
up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

This cable is almost identical to the #3454 3m SAS X cable, except
#ECBJ connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

Limitation: When this cable is ordered with a system in a rack
specifying IBM Plant integration, IBM Manufacturing will ship SAS
cables longer than 3 meters in a separate box and not attempt to
place the cable in the rack.

(#ECBM) - SAS X Cable 15m - HD Narrow 3Gb 2-Adapters to Enclosure

This 15 meter SAS cable connects two PCIe2 SAS adapters or two
PCIe3 SAS adapters to a SAS I/O enclosure. This X cable has four
connectors, two Mini-SAS HD (High Density) Narrow connector and two
Mini SAS connectors. The Mini-SAS HD Narrow connectors attach to two
SAS adapters such as two #EJ0J or two #EJ0L or two #5913. Both Mini
SAS connectors attach to the same I/O drawer (enclosure) such as a
#5886 EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support
up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

This cable is almost identical to the #3458 15m SAS X cable, except
#ECBM connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

Limitation: When this cable is ordered with a system in a rack
specifying IBM Plant integration, IBM Manufacturing will ship SAS
cables longer than 3 meters in a separate box and not attempt to
place the cable in the rack.

(#ECBN) - 5m (16.4-ft), IBM Passive QSFP+ to QSFP+ Cable (DAC)

QSFP+ cable is used for 40Gb-to-40Gb Ethernet connectivity.
clients can use this QSFP+ Direct Attach Cable for Ethernet
connectivity. This copper cable has QSFP+ transceivers already
attached to each end.

Attributes provided: 5m QSFP+ to QSFP+ Cable

Attributes required: QSFP/QSFP+ ports

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECBT) - SAS YO Cable 1.5m - HD Narrow 6Gb Adapter to Enclosure

This 1.5 meter SAS cable connects a PCIe2 SAS adapter or a PCIe3
SAS adapter to a SAS I/O enclosure. This YO cable has three
connectors, one Mini-SAS HD (High Density) Narrow connector and two
Mini SAS connectors. The Mini-SAS HD Narrow connector attaches to a
SAS adapter such as the #EJ0J or #EJ0L or #5913. Both Mini SAS
connectors attach to the same I/O drawer (enclosure) such as a #5886
EXP12S or #5887 EXP24S SAS I/O drawer. This cable can support up to
6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable.

This cable is almost identical to the #3450 1.5m SAS YO cable,
except the #ECBT connector for the SAS adapters is more narrow
allowing it to fit onto the newer PCIe3 SAS adapters.

(#ECBU) - SAS YO Cable 3m - HD Narrow 6Gb Adapter to Enclosure

This 3 meter SAS cable connects a PCIe2 SAS adapter or a PCIe3 SAS
adapter to a SAS I/O enclosure. This YO cable has three connectors,
one Mini-SAS HD (High Density) Narrow connector and two Mini SAS
connectors. The Mini-SAS HD Narrow connector attaches to a SAS
adapter such as the #EJ0J or #EJ0L or #5913. Both Mini SAS connectors
attach to the same I/O drawer (enclosure) such as a #5886 EXP12S or
#5887 EXP24S SAS I/O drawer. This cable can support up to 6Gb
throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable.

This cable is almost identical to the #3451 3m SAS YO cable, except
the #ECBU connector for the SAS adapters is more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

(#ECBV) - SAS YO Cable 6m - HD Narrow 6Gb Adapter to Enclosure

This 6 meter SAS cable connects a PCIe2 SAS adapter or a PCIe3 SAS
adapter to a SAS I/O enclosure. This YO cable has three connectors,
one Mini-SAS HD (High Density) Narrow connector and two Mini SAS
connectors. The Mini-SAS HD Narrow connector attaches to a SAS
adapter such as the #EJ0J or #EJ0L or #5913. Both Mini SAS connectors
attach to the same I/O drawer (enclosure) such as a #5886 EXP12S or
#5887 EXP24S SAS I/O drawer. This cable can support up to 6Gb
throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable.

This cable is almost identical to the #3452 6m SAS YO cable, except
the #ECBV connector for the SAS adapters is more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

Limitation: When this cable is ordered with a system in a rack
specifying IBM Plant integration, IBM Manufacturing will ship SAS
cables longer than 3 meters in a separate box and not attempt to
place the cable in the rack.

(#ECBW) - SAS YO Cable 10m - HD Narrow 6Gb Adapter to Enclosure

This 10 meter SAS cable connects a PCIe2 SAS adapter or a PCIe3 SAS
adapter to a SAS I/O enclosure. This YO cable has three connectors,
one Mini-SAS HD (High Density) Narrow connector and two Mini SAS
connectors. The Mini-SAS HD Narrow connector attaches to a SAS
adapter such as the #EJ0J or #EJ0L or #5913. Both Mini SAS connectors
attach to the same I/O drawer (enclosure) such as a #5886 EXP12S or
#5887 EXP24S SAS I/O drawer. This cable can support up to 6Gb
throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable.

This cable is almost identical to the #3453 10m SAS YO cable,
except the #ECBW connector for the SAS adapters is more narrow
allowing it to fit onto the newer PCIe3 SAS adapters.

Limitation: When this cable is ordered with a system in a rack
specifying IBM Plant integration, IBM Manufacturing will ship SAS
cables longer than 3 meters in a separate box and not attempt to
place the cable in the rack.

(#ECBX) - SAS YO Cable 15m - HD Narrow 3Gb Adapter to Enclosure

This 15 meter SAS cable connects a PCIe2 SAS adapter or a PCIe3 SAS
adapter to a SAS I/O enclosure. This YO cable has three connectors,
one Mini-SAS HD (High Density) Narrow connector and two Mini SAS
connectors. The Mini-SAS HD Narrow connector attaches to a SAS
adapter such as the #EJ0J or #EJ0L or #5913. Both Mini SAS connectors
attach to the same I/O drawer (enclosure) such as a #5887 EXP24S SAS
I/O drawer. This cable can support up to 3Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/O drawer. See SAS cabling documentation for the length of each
leg of the cable.

This cable is almost identical to the #3457 15m SAS YO cable,
except the #ECBX connector for the SAS adapters is more narrow
allowing it to fit onto the newer PCIe3 SAS adapters.

Limitation:

Does not support 6Gb throughput.

When this cable is ordered with a system in a rack specifying IBM
Plant integration, IBM Manufacturing will ship SAS cables longer
than 3 meters in a separate box and not attempt to place the cable
in the rack.

(#ECBY) - SAS AE1 Cable 4m - HD Narrow 6Gb Adapter to Enclosure

This 4 meter SAS cable connects a PCIe3 SAS adapter to a SAS tape
drive or DVD. The tape drive or DVD is probably in an I/O enclosure
such as a bridge box or 1U media enclosure or tape library. This AE
cable has two connectors, one Mini-SAS HD (High Density) Narrow
connector and one Mini-SAS connectors. The Mini-SAS HD Narrow
connector attaches to a SAS adapter such as the #EJ0X. The Mini-SAS
connector attaches to a SAS tape drive enclosure or DVD. This cable
can support up to 6Gb throughput.

Use #ECBY when ordering the cable as a feature code on a Power
System. Alternatively the same cable can be ordered using feature
code #5507 of the IBM tape enclosure or DVD.

Attributes required: available connectors on SAS controller such
as #EJ0X, #EJ10 or #EJ11 for use with an available SAS tape drive or
DVD.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECBZ) - SAS YE1 Cable 3m - HD Narrow 6Gb Adapter to Enclosure

This 3 meter SAS cable connects a PCIe3 SAS adapter to one or two
SAS tape drives. The tape drive(s) is probably in an I/O enclosure
such as a bridge box or 1U media enclosure or tape library. This YO
cable has three connectors, one Mini-SAS HD (High Density) Narrow
connector and two Mini-SAS connectors. The Mini-SAS HD Narrow
connector attaches to a SAS adapter such as the #EJ0X. Each Mini-SAS
connector attaches to a different SAS tape drive enclosure. This
cable can support up to 6Gb throughput.

Use #ECBZ when ordering the cable as a feature code on a Power
System. Alternatively the same cable can be ordered using feature
code #5509 of the IBM tape enclosure.

Attributes provided: connection between PCIe3 SAS adapter with
Mini-SAS HD Narrow connectors and one or two SAS tape drives with
Mini-SAS connectors.

Attributes required: available connectors on SAS controller such
as #EJ0X, #EJ10 or #EJ11 for use with an available SAS tape drive.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC0) - SAS AA Cable 0.6m - HD Narrow 6Gb Adapter to Adapter

This 0.6m SAS cable connects a pair of PCIe SAS adapters with
write cache to each other. The pair can be two PCIe3 SAS adapters or
can be two PCIe2 SAS adapters. This AA cable has two Mini-SAS HD
(High Density) Narrow connectors which connect the top connectors of
the two PCIe adapters providing a high performance path of all the
dual controller communication including mirroring the write cache and
status awareness of each card. This cable can support up to 6Gb
throughput.

Two AA cables are always required between a pair of PCIe3 SAS
adapters with write cache such as #EJ0L unless six (three pair) of
ports or unless eight (four pair) are used to attach I/O drawers. One
AA cable is always required between a pair of PCIe2 SAS adapters with
write cache (pair of #5913 or a pair of #ESA3) unless all six (three
pair) of ports are used to attach I/O drawers. If an AA cable fails
or is disconnected, then the information it was carrying is then sent
over the cables attached to the I/O drawers, sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

This cable is almost identical to the #5918 0.6m SAS AA cable,
except #ECC0 connectors for the SAS adapters are more narrow allowing
it to fit onto the newer PCIe3 SAS adapters.

Attributes provided: high speed connection between two PCIe2 or
two PCIe3 SAS adapters with write cache

Attributes required: available top connectors on pair of PCIe2 or
PCIe3 SAS adapters with write cache

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC2) - SAS AA Cable 1.5m - HD Narrow 6Gb Adapter to Adapter

This 1.5m SAS cable connects a pair of PCIe SAS adapters with
write cache to each other. The pair can be two PCIe3 SAS adapters or
can be two PCIe2 SAS adapters. This AA cable has two Mini-SAS HD
(High Density) Narrow connectors which connect the top connectors of
the two PCIe adapters providing a high performance path of all the
dual controller communication including mirroring the write cache and
status awareness of each card. This cable can support up to 6Gb
throughput.

Two AA cables are always required between a pair of PCIe3 SAS
adapters with write cache such as #EJ0L unless six (three pair) of
ports or unless eight (four pair) are used to attach I/O drawers. One
AA cable is always required between a pair of PCIe2 SAS adapters with
write cache (pair of #5913 or a pair of #ESA3) unless all six (three
pair) of ports are used to attach I/O drawers. If an AA cable fails
or is disconnected, then the information it was carrying is then sent
over the cables attached to the I/O drawers, sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

This cable is almost identical to the #5917 1.5m SAS AA cable,
except #ECC2 connectors for the SAS adapters are more narrow allowing
it to fit onto the newer PCIe3 SAS adapters.

Attributes provided: high speed connection between two PCIe2 or
two PCIe3 SAS adapters with write cache

Attributes required: available top connectors on pair of PCIe2 or
PCIe3 SAS adapters with write cache

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC3) - SAS AA Cable 3m - HD Narrow 6Gb Adapter to Adapter

This 3m SAS cable connects a pair of PCIe SAS adapters with write
cache to each other. The pair can be two PCIe3 SAS adapters or can be
two PCIe2 SAS adapters. This AA cable has two Mini-SAS HD (High
Density) Narrow connectors which connect the top connectors of the two
PCIe adapters providing a high performance path of all the dual
controller communication including mirroring the write cache and
status awareness of each card. This cable can support up to 6Gb
throughput.

Two AA cables are always required between a pair of PCIe3 SAS
adapters with write cache such as #EJ0L unless six (three pair) of
ports or unless eight (four pair) are used to attach I/O drawers. One
AA cable is always required between a pair of PCIe2 SAS adapters with
write cache (pair of #5913 or a pair of #ESA3) unless all six (three
pair) of ports are used to attach I/O drawers. If an AA cable fails
or is disconnected, then the information it was carrying is then sent
over the cables attached to the I/O drawers, sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

This cable is almost identical to the #5915 3m SAS AA cable, except
#ECC3 connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

Attributes provided: high speed connection between two PCIe2 or
two PCIe3 SAS adapters with write cache

Attributes required: available top connectors on pair of PCIe2 or
PCIe3 SAS adapters with write cache

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC4) - SAS AA Cable 6m - HD Narrow 6Gb Adapter to Adapter

This 6m SAS cable connects a pair of PCIe SAS adapters with write
cache to each other. The pair can be two PCIe3 SAS adapters or can be
two PCIe2 SAS adapters. This AA cable has two Mini-SAS HD (High
Density) Narrow connectors which connect the top connectors of the two
PCIe adapters providing a high performance path of all the dual
controller communication including mirroring the write cache and
status awareness of each card. This cable can support up to 6Gb
throughput.

Two AA cables are always required between a pair of PCIe3 SAS
adapters with write cache such as #EJ0L unless six (three pair) of
ports or unless eight (four pair) are used to attach I/O drawers. One
AA cable is always required between a pair of PCIe2 SAS adapters with
write cache (pair of #5913 or a pair of #ESA3) unless all six (three
pair) of ports are used to attach I/O drawers. If an AA cable fails
or is disconnected, then the information it was carrying is then sent
over the cables attached to the I/O drawers, sharing their bandwidth.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

This cable is almost identical to the #5916 6m SAS AA cable, except
#ECC4 connectors for the SAS adapters are more narrow allowing it to
fit onto the newer PCIe3 SAS adapters.

Attributes provided: high speed connection between two PCIe2 or
two PCIe3 SAS adapters with write cache

Attributes required: available top connectors on pair of PCIe2 or
PCIe3 SAS adapters with write cache

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC6) - 2M Optical Cable Pair for PCIe3 Expansion Drawer

The 2.0 meter active optical cable (AOC) pair connects a PCIe3
module in the PCIe Gen3 I/O Expansion Drawer to a PCIe3 Optical
Converter Adapter in the system unit. There are two identical cables
in the cable pair, each with two CXP connectors. One of the cables
attaches to the top CXP port of the PCIe3 module and to the top CXP
port of the PCIe3 Optical Converter Adapter. The other cable attaches
to the bottom CXP ports. See also other AOC cable length options such
as the feature #ECC8 (10 meter). The 2 meter length is suggested only
for intra-rack connections.

Attributes provided: Pair of 2 meter active optical cables

Attributes required: CXP ports on a PCIe3 Optical Cable Adapter
(#EJ07) and on a PCIe3 module such as a #EMXF in a PCIe Gen3 Expansion
Drawer (#EMX0)

Minimum required: 0

Maximum allowed: 8 (Initial order maximum: 8)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC8) - 10M Optical Cable Pair for PCIe3 Expansion Drawer

The 10.0 meter active optical cable (AOC) pair connects a PCIe3
module in the PCIe Gen3 I/O Expansion Drawer to a PCIe3 Optical
Converter Adapter in the system unit. There are two identical cables
in the cable pair, each with two CXP connectors. One of the cables
attaches to the top CXP port of the PCIe3 module and to the top CXP
port of the PCIe3 Optical Converter Adapter. The other cable attaches
to the bottom CXP ports. See also other AOC cable length options such
as the feature #ECC6 (2 meter). The 10 meter length is suggested for
cabling to a different rack.

Attributes provided: Pair of 10 meter active optical cables

Attributes required: CXP ports on a PCIe3 Optical Cable Adapter
(#EJ07) and on a PCIe3 module such as a #EMXF in a PCIe Gen3 Expansion
Drawer (#EMX0)

Minimum required: 0

Maximum allowed: 8 (Initial order maximum: 8)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECC9) - 20M Optical Cable Pair for PCIe3 Expansion Drawer

The 20 meter active optical cable (AOC) pair connects a PCIe3
module in the PCIe Gen3 I/O Expansion Drawer to a PCIe3 Optical
Converter Adapter in the system unit. There are two identical cables
in the cable pair, each with two CXP connectors. One of the cables
attaches to the top CXP port of the PCIe3 module and to the top CXP
port of the PCIe3 Optical Converter Adapter. The other cable attaches
to the bottom CXP ports.

See also other AOC cable length options such as the feature #ECC8
(10 meter). The 20 meter length is suggested only for extra-rack
connections.

Attributes provided: Pair of 20 meter active optical cables

Attributes required: CXP ports on a PCIe3 Optical Cable Adapter
and on a PCIe3 module in a PCIe Gen3 Expansion Drawer

Attributes provided: System Node to System Control Unit Cable Set
for Drawer 2

Attributes required: System Node Drawer 2 and System Control Unit

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECCC) - System Node to System Control Unit Cable Set for Drawer 3

Attributes provided: System Node to System Control Unit Cable Set
for Drawer 3

Attributes required: System Node Drawer 3 and System Control Unit

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECCD) - System Node to System Control Unit Cable Set for Drawer 4

Attributes provided: System Node to System Control Unit Cable Set
for Drawer 4

Attributes required: System Node Drawer 4 and System Control Unit

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECDJ) - 3.0M SAS X12 Cable (Two Adapter to Enclosure)

This 3 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four Mini-SAS HD (High Density)
connectors. Two of connectors attach to two SAS adapters such as in
the PCIe3 SAS Adapter. The other two connectors attach to one SAS I/
O Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability. Multiple cable
length feature codes are available. Choose the cable length that best
matches the distance between the adapter and the I/ O drawer. See SAS
cabling documentation for the length of each leg of the cable. The
SAS adapters can be in the same or in different PCIe I/O drawers. Or
one adapters can be in a supported Power System CEC and the other
adapter can be in a PCIe I/O drawer. Or both adapters can be in a
supported Power System CEC. Limitation: This cable can not be used
the EXP24S I/O drawer (#5887 or# EL1S) which uses Mini-SAS connectors
which are not HD.

This 4.5 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four Mini-SAS HD (High Density)
connectors. Two of connectors attach to two SAS adapters such as in
the PCIe3 SAS Adapter. The other two connectors attach to one SAS I/
O Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/ O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Limitation: This cable can not be used with the EXP24S I/O drawer
(#5887 or# EL1S) which uses Mini-SAS connectors which are not HD.

(#ECDL) - 10M SAS X12 Active Optical Cable (Two Adapter to Enclosure)

This 10 meter SAS cable connects two SAS adapters to a SAS I/O
enclosure. This X cable has four Mini-SAS HD (High Density)
connectors. Two of connectors attach to two SAS adapters such as in
the PCIe3 SAS Adapter. The other two connectors attach to one SAS I/
O Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/ O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Limitation: This cable can not be used with the EXP24S I/O drawer
(#5887 or# EL1S) which uses Mini-SAS connectors which are not HD.

(#ECDT) - 1.5M SAS YO12 Cable (Adapter to Enclosure)

This 1.5 meter SAS cable connects one SAS adapter to a SAS I/O
enclosure. This YO cable has three Mini-SAS HD (High Density)
connectors. One of connectors attach to a SAS adapter such as in the
PCIe3 SAS Adapter. The other two connectors attach to one SAS I/O
Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability. Multiple cable
length feature codes are available. Choose the cable length that best
matches the distance between the adapter and the I/ O drawer. See SAS
cabling documentation for the length of each leg of the cable. The
SAS adapters can be in the same or in different PCIe I/O drawers. Or
one adapters can be in a supported Power System CEC and the other
adapter can be in a PCIe I/O drawer. Or both adapters can be in a
supported Power System CEC. Limitation: This cable can not be used
the EXP24S I/O drawer (#5887 or# EL1S) which uses Mini-SAS connectors
which are not HD.

(#ECDU) - 3.0M SAS YO12 Cable (Adapter to Enclosure)

This 3 meter SAS cable connects one SAS adapter to a SAS I/O
enclosure. This YO cable has three Mini-SAS HD (High Density)
connectors. One of connectors attach to a SAS adapter such as in the
PCIe3 SAS Adapter. The other two connectors attach to one SAS I/O
Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability. Multiple cable
length feature codes are available. Choose the cable length that best
matches the distance between the adapter and the I/ O drawer. See SAS
cabling documentation for the length of each leg of the cable. The
SAS adapters can be in the same or in different PCIe I/O drawers. Or
one adapters can be in a supported Power System CEC and the other
adapter can be in a PCIe I/O drawer. Or both adapters can be in a
supported Power System CEC. Limitation: This cable can not be used
the EXP24S I/O drawer (#5887 or# EL1S) which uses Mini-SAS connectors
which are not HD.

(#ECDV) - 4.5M SAS YO12 Active Optical Cable (Adapter to Enclosure)

This 4.5 meter SAS cable connects one SAS adapter to a SAS I/O
enclosure. This YO cable has three Mini-SAS HD (High Density)
connectors. One of connectors attach to a SAS adapter such as in the
PCIe3 SAS Adapter. The other two connectors attach to one SAS I/O
Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/ O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Limitation: This cable can not be used with the EXP24S I/O drawer
(#5887 or# EL1S) which uses Mini-SAS connectors which are not HD.

(#ECDW) - 10M SAS YO12 Active Optical Cable (Adapter to Enclosure)

This 10 meter SAS cable connects one SAS adapter to a SAS I/O
enclosure. This YO cable has three Mini-SAS HD (High Density)
connectors. One of connectors attach to a SAS adapter such as in the
PCIe3 SAS Adapter. The other two connectors attach to one SAS I/O
Enclosure such as the EXP12SX (#ESLL/#ELLL ) or EXP24SX (#ESLS/
#ELLS). This cable is designed for high speed (DHS) to support up to
12Gb throughput if the adapter has that capability.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the I/ O drawer. See SAS cabling documentation for the length of each
leg of the cable. The SAS adapters can be in the same or in different
PCIe I/O drawers. Or one adapter can be in a supported Power System
CEC and the other adapter can be in a PCIe I/O drawer. Or both
adapters can be in a supported Power System CEC.

Limitation: This cable can not be used with the EXP24S I/O drawer
(#5887 or# EL1S) which uses Mini-SAS connectors which are not HD.

(#ECE0) - 0.6M SAS AA12 Cable (Adapter to Adapter)

This 0.6 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density) connectors
which connect the top connectors of two PCIe3 SAS adapters with write
cachce such as #EJ0L or #EJ14. The cable provides a high performance
path of all the dual controller communication including mirroring the
write cache and status awareness of each card. This cable is designed
for high speed to support up to 12Gb throughput. Note EJ0L/EJ14
support 6Gb. Two AA cable is always required between a pair of PCIe3
SAS adapters with write cache when just one or two I/O drawers are
attached. One AA cables is required if three I/O drawers are
attached. If four drawers are attached or if an AA cable fails or is
disconnected, then the information it was carrying is then sent over
the cables attached to the I/O drawers, sharing their bandwidth. AA
cables are not used with SAS adapters with no write cache. Multiple
cable length feature codes are available. Choose the cable length
that best matches the distance between the two adapters. The 6Gb
version of this cable is feature #5918. #5918 and #ECE0 can be mixed
on the same PCIe3 adapter pair.

(#ECE3) - 3.0M SAS AA12 Cable (Adapter to Adapter)

This 3.0 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density) connectors
which connect the top connectors of two PCIe3 SAS adapters with write
cachce such as #EJ0L or #EJ14. The cable provides a high performance
path of all the dual controller communication including mirroring the
write cache and status awareness of each card. This cable is designed
for high speed to support up to 12Gb throughput. Note EJ0L/EJ14
support 6Gb. Two AA cable is always required between a pair of PCIe3
SAS adapters with write cache when just one or two I/O drawers are
attached. One AA cables is required if three I/O drawers are
attached. If four drawers are attached or if an AA cable fails or is
disconnected, then the information it was carrying is then sent over
the cables attached to the I/O drawers, sharing their bandwidth. AA
cables are not used with SAS adapters with no write cache. Multiple
cable length feature codes are available. Choose the cable length
that best matches the distance between the two adapters. The 6Gb
version of this cable is feature #5915. #5915 and #ECE3 can be mixed
on the same PCIe3 adapter pair.

(#ECE4) - 4.5M SAS AA12 Active Optical Cable (Adapter to Adapter)

This 4.5 meter SAS cable connects a pair of SAS adapters to each
other. This AA cable has two Mini-SAS HD (High Density) connectors
which connect the top connectors of two PCIe3 SAS adapters with write
cachce such as #EJ0L or #EJ14. The cable provides a high performance
path of all the dual controller communication including mirroring the
write cache and status awareness of each card. This cable is designed
for high speed to support up to 12Gb throughput. Note EJ0L/EJ14
support 6Gb.

Two AA cables are always required between a pair of PCIe3 SAS
adapters with write cache when just one or two I/O drawers are
attached. One AA cable is required if three I/O drawers are attached.
If four drawers are attached or if an AA cable fails or is
disconnected, then the information it was carrying is then sent over
the cables attached to the I/O drawers, sharing their bandwidth. AA
cables are not used with SAS adapters with no write cache.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the two adapters.

Note: One of feature #ECRF, and one of feature #ECRG and two of
feature #ECRJ are required per one #ECR0 ordered.

(#ECRF) - Rack Front Door High-End appearance

This feature provides a front door in High-End appearance with an
IBM logo for the S42 rack. A front door such as #ECRF is recommended
on the S42 rack. A front door is required on ECR0 rack. IBM ships
rack with the handle on the right and hinges on the left viewed facing
the front of the rack.

The door comes with a lock which is keyed the same as the rear door
or side panels. Uniquely keyed locks can be obtained by the client
directly from Southco, the vendor from whom IBM purchased the lock.

Attributes provided: Front Door with lock

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ECRG) - Rack Rear Door Black

This feature provides a rear door in flat black color for the S42
rack. Feature number ECRF and ECR0 rack and ECRG are recommended on
the S42 rack. A front and rear door is required on the ECR0 rack.

The door is the full width of the rack and the hinges and lockplate
can be moved from side to side allowing the door to be opened on the
left or on the right. IBM ships rack with the handle on the right and
hinges on the left viewed facing the rear of the rack.

The front doors , rear doors and side panels come with a lock which
is keyed the same as the front door or side panels. Uniquely keyed
locks can be obtained by the client directly from Southco, the vendor
from whom IBM purchased the lock.

Attributes provided: Rear Door with lock

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ECRJ) - Rack Side Cover

This feature provides one side panels in black color for the S42 or
feature #ECR0 rack. Each side panel can cover either the left or the
right side of the rack. These side covers are optional but
recommended on S42 rack. 2x ECRJ are required on every ECR0 for
optimal airflow through a rack and for physical security.

The front door, rear doors and side panels come with a lock which
is keyed the same as the front door or side panels. Uniquely keyed
locks can be obtained by the client directly from Southco, the vendor
from whom IBM purchased the lock.

Attributes provided: One side panel for rack.

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ECRK) - Rack Rear Extension 5-In

Attributes provided: Rack Rear Extension

Attributes required: Maximum one per feature #ECR0 rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ECRM) - Rack Front Door for Rack (Black/Flat)

Attributes required: None

For 9119-MHE: (#ECRM)

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required:

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ECS0) - Indicator Assembled and Tested in China

This is indicator that the product was assembled and tested in
China.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#ECSF) - Custom Service Specify, Montpellier, France

Having #ECSF on the order, will cause the order to be routed to
France and the machine to be internally routed to the CSC build area.

Attributes provided: Customization

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ECSJ) - NeuCloud Indicator/Specify

(No Longer Available as of June 19, 2015)

This is a no-charge specify code for helping IBM track orders for
reporting purposes

Attributes provided: None

Attributes required: Used in mainland China

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: N/A

Return parts MES: No

(#ECSM) - Custom Service Specify, Mexico

Having #ECSM on the order, will cause the order to be routed to
Mexico and the machine to be internally routed to the CSC build area.

Attributes provided: Customization

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#ECSP) - Custom Service Specify, Poughkeepsie, NY, USA

Having #ECSP on the order, will cause the order to be routed to
Poughkeepsie, NY, USA and the machine to be internally routed to the CSC
build area.

Attributes provided: Customization

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#ECSS) - Integrated Solution Packing

This is a routing indicator for Solution packing.

Attributes provided: Routing instruction for manufacturing

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#ECW0) - Optical Wrap Plug

A wrap plug is a small connector designed to perform a diagnostic
test called a loopback test. This wrap plug is inserted into a SR
optical port on a PCIe Fibre Channel adapter or a SR or LR optical
port on a PCIe Ethernet adapter

This is a multi-mode LC fiber optic wrap plug with an inside/
outside optics diameter of 50/125. Its IBM part number as of early
2016 is 12R9314. An earlier equivalent function IBM part number which
is no longer shipped is 11P3847.

It is strongly recommended that Fibre Channel adapters (HBAs) fill
any empty adapter ports with a wrap plug. There is no technical issue
leaving a port empty. However, filling all ports with a cable to a
device/switch or with a wrap plug can speed the booting/IPLing of a
partition and can avoid error messages uselessly pointing to a planned
empty port.

There is no technical issue leaving an Ethernet port empty.
Whether an Ethernet port is empty or contains a wrap plug should not
impact boot/IPL time or impact empty-port messages.

Attributes provided: Wrap plug

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EH10) - 387GB SFF-2 SSD converted for AIX/Linux

This special feature code is used when "converting" a four-pack of
387GB SSD which are mounted on a SFF-1 carrier/trays to four features
all mounted on SFF-2 carrier/trays. SFF-2 carrier/trays can be placed
in an EXP24S I/O Drawer. A special feature code was created because
a) the originally announced 387GB SFF-2 feature code is either
withdrawn from marketing or b) the server on which it was initially
ordered is withdrawan fom marketing and the four-pack is an "iniital
only" orderable feature. IBM ordering/configuring systems don't
support converting to a withdrawn features or initial only features.
It was decided to use the same approach for all 4-pack features for
consistency even if all feature codes for your specific model are
still available. This is ordered using one "remove" and four "adds"
in the IBM ordering system and does not use the more common one-to-one
feature conversion. IBM ships one SFF-2 carrier/tray with each #EH10
feature ordered. SFF-2 #EH10 is equivalent to one SFF-2 #ES0C. Use
when converting from SFF-1 four pack #ESRA.

(#EH11) - 387GB SFF-2 SSD converted for AIX/Linux

This special feature code is used when "converting" a four-pack of
387GB SSD which are mounted on a SFF-1 carrier/trays to four features
all mounted on SFF-2 carrier/trays. SFF-2 carrier/trays can be placed
in an EXP24S I/O Drawer. A special feature code was created because
a) the originally announced 387GB SFF-2 feature code is either
withdrawn from marketing or b) the server on which it was initially
ordered is withdrawn from marketing and the four-pack is an "initial
only" orderable feature. IBM ordering/configuring systems don't
support converting to a withdrawn features or initial only features.
It was decided to use the same approach for all 4-pack features for
consistency even if all feature codes for your specific model are
still available. This is ordered using one "remove" and four "adds"
in the IBM ordering system and does not use the more common one-to-one
feature conversion. IBM ships one SFF-2 carrier/tray with each #EH11
feature ordered. SFF-2 #EH11 is equivalent to one SFF-2 #ES19. Use
when converting fom SFF-1 four pack #ESRE

(#EH12) - 387GB SFF-2 SSD converted for IBM i

This special feature code is used when "converting" a four-pack of
387GB SSD which are mounted on a SFF-1 carrier/trays to four features
all mounted on SFF-2 carrier/trays. SFF-2 carrier/trays can be placed
in an EXP24S I/O Drawer. A special feature code was created because
a) the originally announced 387GB SFF-2 feature code is either
withdrawn from marketing or b) the server on which it was initially
ordered is withdrawn from marketing and the four-pack is an "initial
only" orderable feature. IBM ordering/configuring systems don't
support converting to a withdrawn features or initial only features.
It was decided to use the same approach for all 4-pack features for
consistency even if all feature codes for your specific model are
still available. This is ordered using one "remove" and four "adds"
in the IBM ordering system and does not use the more common one-to-one
feature conversion. IBM ships one SFF-2 carrier/tray with each #EH12
feature ordered. SFF-2 #EH12 is equivalent to one SFF-2 #ES0D. Use
when converting fom SFF-1 four pack #ESRB

(#EH13) - 387GB SFF-2 SSD converted for IBM i

This special feature code is used when "converting" a four-pack of
387GB SSD which are mounted on a SFF-1 carrier/trays to four features
all mounted on SFF-2 carrier/trays. SFF-2 carrier/trays can be placed
in an EXP24S I/O Drawer. A special feature code was created because
a) the originally announced 387GB SFF-2 feature code is either
withdrawn from marketing or b) the server on which it was initially
ordered is withdrawn from marketing and the four-pack is an "initial
only" orderable feature. IBM ordering/configuring systems don't
support converting to a withdrawn features or initial only features.
It was decided to use the same approach for all 4-pack features for
consistency even if all feature codes for your specific model are
still available. This is ordered using one "remove" and four "adds"
in the IBM ordering system and does not use the more common one-to-one
feature conversion. IBM ships one SFF-2 carrier/tray with each #EH13
feature ordered. SFF-2 #EH13 is equivalent to one SFF-2 #ES1A. Use
when converting from SFF-1 four pack #ESRF.

(#EH14) - GEN2-S Conversion Carrier for Feature ES0C 387GB SSD

This feature delivers one GEN2-S carrier as a MES upgrade. The
carrier #EH14 accommodates one #ES0A 387GB SFF-1 SSD for AIX/Linux so
it can be installed as a conversion in a GEN2 bay such as #5887 EXP24S
SFF Gen2-bay Drawer. It is now functionally equivalent to #ES0C.

Attributes provided: One GEN2 Carrier

Attributes required: None

Minimum required: 0

Maximum allowed: 2304 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#EH15) Qty 150 Gen2-S Conversion Carriers for feature EQ0A 387GB SSD

Each instance of this feature #EH15 delivers 150 GEN2-S carriers as
an MES upgrade. These carriers convert one #EQ0A (Quantity 150 of
#ES0A, 387GB SFF-1 SSD for AIX/Linux). Once on the Carrier the drives
can be installed in a #5887 EXP24S SFF Gen2-bay Drawer. It is now
functionally equivalent #EQ0C quantity 150 of #ES0C.

Attributes provided: 150 GEN2 Carriers

Attributes required: None

Minimum required: 0

Maximum allowed: 9 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#EH16) GEN2-S Conversion Carrier for Feature ES0D 387GB SSD (IBM i)

This feature delivers a GEN2-S carrier as a MES upgrade. The
carrier #EH16 accommodates #ES0B 387GB SFF-2 SSD for IBM i with eMLC.
Once on the carrier it can be installed in a GEN2 bay such as #5887
EXP24S SFF Gen2-bay Drawer. It is now functionally equivalent to
#ES0D.

Each instance of this feature #EH17 delivers 150 GEN2-S carriers as
an MES upgrade. These carriers convert one #EQ0B (Quantity 150 of
#ES0B 387GB SFF-2 SSD for IBM i with eMLC). Once on the Carrier the
drives can be installed in a #5887 EXP24S SFF Gen2-bay drawer. It is
now functionally equivalent to #EQ0D quantity 150 of #ES0D.

Attributes provided: 150 GEN2 Carriers

Attributes required: None

Minimum required: 0

Maximum allowed: 9 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#EH35) - Mobile Enablement

Enterprise Pool Enablement

Attributes provided: Enterprise Pool Enablement Code

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 0)

OS level required:

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: Does not apply

(#EHCE) - IBM Cognos Business Intelligence

Routing indicator for IBM Cognos Business Intelligence on Power
Solution.

PCIe3 x16 slot in system unit plus a pair of active optical cables
(one optical pair feature such as #ECC6 or #ECC8

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJ0J) - PCIe3 RAID SAS Adapter Quad-port 6Gb x8

The PCIe3 RAID SAS Adapter is a high performance SSD/HDD controller
using PCIe Gen3 x8 technology. The adapter does not have write cache
and thus pairing with another PCIe3 RAID SAS Adapter (#EJ0J or #EJ0M)
is optional. Pairing can provide controller redundancy and enhance
performance. There are no batteries in the adapter to maintain.dapter
to maintain.

The adapter provides four Mini-SAS HD (high density) narrow
connectors for the attachment of SAS drives located in the EXP24S,
EXP12SX or EXP24SX storage enclosures or #5802/5803/EL36 12X PCIe I/O
drawers. X, YO or AT SAS cables with HD narrow connectors are used to
attach to these drawers. A max of 4 EXP24S/EXP12SX/EXP24SX can be
attached. A maximum of 48 SSD can be attached and a maximum of 96 HDD
can be attached per adapter or per adapter pair.

The adapter provides RAID 0, RAID 5, RAID 6 and RAID 10 for AIX and
Linux and VIOS. The adapter provides RAID 5 and RAID 6 for all levels
of IBM i and also provides RAID 10 for later levels of IBM i. IBM i
provides both OS mirroring and data spreading. AIX/Linux/VIOS provide
OS mirroring (LVM).

Features #EJ0J and #EJ0M are electronically identical with the same
CCIN of 57B4. #EJ0J has a full-high tailstock bracket and air baffle.
#EJ0M has a low profile tailstock bracket. #EJ10/#EJ11 are identical
with #EJ0J/#EJ0M, but have different feature codes to identify their
use as tape/DVD controllers to IBM configurator tools instead of
disk/SSD controllers.

Both 5xx and 4k byte sector HDD/SSD are supported for POWER8
servers. 5xx byte sector HDD/SSD are supported for earlier generation
servers. 5xx and 4k drives cannot be mixed in the same array..

Limitations:

HDD/SSD workloads which are performance sensitive to WRITES should
use the #EJ14 or #EJ0L controller which provides write cache.

HDD and SSD cannot be mixed on the same SAS port, but can be mixed
on the same adapter.

#5886 3.5-inch SAS Storage Drawer is not supported.

177 GB SSD are not supported.

Running SAS bays for both a #5887 EXP24S I/O drawer and a
12X-attached #5802/5803 I/O drawer on the same adapter or adapter pair
is not supported. Note mixing EXP24S or EXP12SX or EXP24SX is
supported.

If controlling drives in a #5802/5803/EL36 as a single controller,
the #EJ0J must be located in that #5802/5803/EL36. If controlling
drives in a #5802/5803/EL36 as a pair of controllers, at least one of
the SAS adapter pairs must be located in that #5802/5803/EL36.

Tape/DVD cannot be mixed with disk/SSD on the same adapter.

Attributes provided: full high PCIe3 four port x8 SAS RAID
adapter with no write cache and optional pairing.

Attributes required: One PCIe slot per adapter and Mini-SAS HD
narrow connector SAS cables such as #ECBJ-ECBL, #ECBT-ECBV, #ECCO-ECC4
or 12 Gb cables such as #ECDJ, ECDT or #ECDU.

Minimum required: 0

Maximum allowed: 128 (Initial order maximum: 128)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#EJ0L) - PCIe3 12GB Cache RAID SAS Adapter Quad-port 6Gb x8

The PCIe3 12 GB Cache RAID SAS Adapter provides high performance
HDD and/or SSD controller function using PCIe Gen3 technology. A pair
of adapters are required to provide mirrored write cache data and
adapter redundancy. Integrated flash memory provides protection of
the write cache without batteries in case of power failure.
Effectively up to 12 GB of write cache is provided using compression
of 4 GB of physical cache.

The adapter provides four Mini-SAS HD (high density) narrow
connectors for the attachment of SAS drives located in the SAS EXP24S
or EXP12SX or EXP24SX storage enclosures or #5802/5803/EL36 12X PCIe
I/O drawers. X, YO or AT SAS cables with HD narrow connectors are
used to attach to these drawers. A max of 4 EXP24S/EXP12SX/EXP24SX
can be attached with a maximum of 96 HDD or a maximum of 48 SSD. Two
AA SAS cable with HD narrow connectors are attached to the #EJ0L pair
to communicate status and cache content information and are required
unless three or four ports are being used to attach HDD/SSD.

The #EJ0L provides RAID 0, RAID 5, RAID 6 and RAID 10 for AIX and
Linux and VIOS. The adapter provides RAID 5 and RAID 6 for all levels
of IBM i and also provides RAID 10 for later levels of IBM i. IBM i
provides both OS mirroring and data spreading. AIX/Linux/VIOS provide
OS mirroring (LVM). This adapter can also support the Easy Tier
function (RAID 52T, 62T or 102T) for AIX and Linux. The adapter's
CCIN is 57CE.

Both 5xx and 4k byte sector HDD/SSD are supported for POWER8
servers. 5xx byte sector HDD/SSD are supported for earlier generation
servers. 5xx and 4k drives cannot be mixed in the same array.

Limitations:

HDD and SSD cannot be mixed on the same SAS port, but can be mixed
on the same adapter.

#5886 3.5-inch SAS I/O Drawer is not supported.

177 GB SSD are not supported.

Running SAS bays for both a #5887 EXP24S I/O drawer and a
12X-attached #5802/5803 I/O drawer on the same adapter pair is not
supported. Note mixing EXP24S or EXP12SX or EXP24SX is supported.

If controlling drives in a #5802/5803/EL36 at least one of the
#EJ0L pairs must be located in that #5802/5803/EL36.

Attributes provided: full high PCIe3 four port x8 adapter with up
to 12 GB write cache.

(#EJ0M) - PCIe3 LP RAID SAS ADAPTER Quad-Port 6Gb x8

The PCIe3 RAID SAS LP Adapter is a high performance SSD/HDD
controller using PCIe Gen3 x8 technology. The adapter does not have
write cache and thus pairing with another PCIe3 RAID SAS Adapter
(#EJ0M or #EJ0J) is optional. Pairing can provide controller
redundancy and enhance performance. There are no batteries in the
adapter to maintain.

The adapter provides four Mini-SAS HD (high density) narrow
connectors for the attachment of SAS drives located in the EXP24S,
EXP12SX, EXP24SX storage enclosure or #5802/EL36 12X PCIe I/O drawers.
X, YO or AT SAS cables with HD narrow connectors are used to attach to
these drawers. A max of 4 EXP24S/EXP12SX/EXP24SX can be attached. A
maximum of 48 SSD can be attached and a maximum of 96 HDD can be
attached per adapter or per adapter pair.

The adapter provides RAID 0, RAID 5, RAID 6 and RAID 10 for AIX and
Linux and VIOS. The adapter provides RAID 5 and RAID 6 for all levels
of IBM i and also provides RAID 10 for later levels of IBM i. IBM i
provides both OS mirroring and data spreading. AIX/Linux/VIOS provide
OS mirroring (LVM).

Features #EJ0J and #EJ0M are electronically identical with the same
CCIN of 57B4. #EJ0J has a full-high tailstock bracket and air baffle.
#EJ0M has a low profile tailstock bracket. #EJ10 / #EJ11 are
identical with #EJ0J / #EJ0M respectively, but have different feature
codes to identify their use to IBM configurator tools as tape/DVD
controllers instead of disk/SSD controllers.

Both 5xx and 4k byte sector HDD/SSD are supported for POWER8
servers. 5xx byte sector HDD/SSD are supported for earlier generation
servers. 5xx and 4k drives cannot be mixed in the same array.

Limitations:

HDD/SSD workloads which are performance sensitive to WRITES should
use the #EJ14 or #EJ0L controller which provide write cache.

HDD and SSD cannot be mixed on the same SAS port, but can be mixed
on the same adapter.

#5886 3.5-inch SAS Storage Drawer is not supported.

177 GB SSD are not supported.

Running SAS bays for both a EXP24S storage drawer and a
12X-attached #5802/5803 I/O drawer on the same adapter or adapter pair
is not supported. Note mixing EXP24S or EXP12SX or EXP24SX is
supported.

A single #EJ0M cannot control drives in a #5802/5803/EL36 because
the adapter must be located in that #5802/5803/ EL36 in a full high
slot. However, if controlling drives in a #5802/5803/EL36 as a pair
of controllers, one #EJ0J can be located in that #5802/5803/EL36 and
one #EJ0M can be in the system's low profile slot.

(#EJ10) - PCIe3 SAS Tape/DVD Adapter Quad-port 6Gb x8

The PCIe3 SAS Adapter is a high performance SAS tape controller
using PCIe Gen3 x8 technology. It supports LTO-5 or later tape drives
and DVD. The adapter provides four Mini-SAS HD (high density)
connectors to which AE1 SAS cables such as #ECBY and/or YE1 SAS Cables
such as #ECBZ with HD narrow connectors can be attached. A max of 4
tape drives per adapter can be attached using four AE1 cables. A max
of 8 tape drives can be attached using four YE1 cables.

#EJ10 (full high) and #EJ11 (low profile) are electronically the
same adapter with the same 57B4 CCIN, but differ in that their
tailstocks fit different size PCIe slots.

#EJ0J and #EJ10 are the same adapter with the same 57B4 CCIN, but
have different feature code numbers to indicate different usage to IBM
configurator tools. #EJ10 runs SAS LTO-5 or later drives and DVD.
Support of both tape/DVD and HDD/SSD on the same adapter is not
supported.

Note: The original #EJ0X adapter does not support DVD but also has
the same CCIN.

(#EJ11) - PCIe3 LP SAS Tape/DVD Adapter Quad-port 6Gb x8

The PCIe3 Low Profile SAS Adapter is a high performance SAS tape
controller using PCIe Gen3 x8 technology. It supports LTO-5 or later
tape drives and DVD. The adapter provides four Mini-SAS HD (high
density) connectors to which AE1 SAS cables such as #ECBY and/or YE1
SAS Cables such as #ECBZ with HD narrow connectors can be attached. A
max of 4 tape drives per adapter can be attached using four AE1
cables. A max of 8 tape drives can be attached using four YE1 cables.

#EJ10 (full high) and #EJ11 (low profile) are electronically the
same adapter with the same 57B4 CCIN, but differ in that their
tailstocks fit different size PCIe slots.

#EJ0M and #EJ11 are the same adapter with the same 57B4 CCIN, but
have different feature code numbers to indicate different usage to IBM
configurator tools. #EJ11 runs SAS LTO-5 or later tape drives and
DVD. Support of both tape/DVD and HDD/SSD on the same adapter is not
supported.

The #EJ14 provides four Mini-SAS HD (high density) narrow
connectors for the attachment of SAS drives located in the SAS EXP24S,
EXP12SX, or EXP24SX storage enclosures X, YO or AT SAS cables with HD
narrow connectors are used to attach to these drawers. A max of 4
EXP24S can be attached with a maximum of 96 HDD or a maximum of 72 SSD
per pair of #EJ14. If more than 48 SSD are attached, no HDD can be
attached. A mix of EXP24S, EXP12SX or EXP24SX is supported on the
same adapter pair.

Two AA SAS cable with HD narrow connectors are attached to the
#E14L pair to communicate status and cache content information and are
required unless three or four ports are being used to attach HDD/SSD.

Adapter must be placed in a x16 slot in the system unit which is
CAPI enabled. The server must have CAPI enablement feature.

Features #EJ17 and #EJ18 are electronically identical, but have
different tailstock brackets. #EJ17 is full high and #EJ18 is low
profile.

Limitation: Concurrent add, remove or replacement of this adapter
is not supported.

Attributes provided: CAPI enabled FPGA

Attributes required: One x16 CAPI enabled PCIe3 Slot

Minimum required: 0

Maximum allowed: 8 (Initial order maximum: 8)

Limitation: Concurrent add, remove or replacement of this adapter
is not supported.

OS level required:

AIX 7.2 or later supported

IBM i - not supported

Linux - not supported

Refer to Software Requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJ1P) - PCIe1 SAS Tape/DVD Dual-port 3Gb x8 Adapter

This is a PCIe Gen1 short x8 form factor adapter. It supports the
attachment of SAS tape and DVD using a pair of mini SAS 4x connectors.
The PCIe1 can be used for external tape drives which are not supported
on the newer and faster 4-port 6Gb PCIe3 adapter (see feature code
#EJ10/EJ11/EL60). The adapter supports external SAS tape drives such as
the DAT72, DAT160, LTO-4, LTO-5, LTO-6, and LTO-7 found in the IBM
multimedia drawers such as the 7226-1U3 or 7214-1U2 or tape units such
as the TS2240, TS2340, TS3100, TS3200, and TS3310. Other removable media
devices supported include IBM SAS/SATA DVD-ROM/RAM drives.

SAS adapter-to-enclosure (AE) 3Gb cables with mini-SAS connectors are
used to attach these drives. See feature codes #3684 (3 meter) and #3685
(6 meter). The same AE cables can often alternatively be ordered under
the tape enclosure or multimedia drawer.

Feature EJ1P and EJ1N are electrically and functionally identical with
the same CCIN of 57B3. EJ1P has a full-high tailstock bracket and EJ1N
has a low profile tailstock bracket. Feature EJ1P/EJ1N is the same
adapter as #5901/5278 but designates to IBM configurator tools that the
usage will be tape/DVD and will not be used for disk.

Attributes provided: Two mini SAS 4x connectors

Attributes required: One PCIe slot

Minimum required: 0

Maximum allowed: 8 (Initial order maximum: 8)

OS level required:

IBM i supported

AIX Version 6.1, or later

AIX Version 7.1, or later

AIX Version 7.2, or later

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise 11.3, or later

SUSE Linux Enterprise 12, or later

Ubuntu not supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EJ28) - PCIe Crypto Coprocessor Gen3 BSC 4765-001

(No Longer Available as of January 20, 2017)

Secure-key adapter provides both cryptographic coprocessor and
cryptographic accelerator functions in a single PCIe card. The
adapter is well suited to applications requiring high-speed,
security-sensitive, RSA acceleration, cryptographic operations for data
encryption and digital signing, secure management, and use of
cryptographic keys, or custom cryptographic applications. It provides
secure storage of cryptographic keys in a tamper-resistant hardware
security module designed to meet FIPS 140-2 level 4 security
requirements. The adapter is a PCIe 4x full height - short card.

#EJ27, #EJ28 and #EJ29 are all feature codes representing the same
physical card with the same CCIN of 4765, but different feature codes
are used to indicate if a blind swap cassette is used and its type.
#EJ27 indicates no blind swap cassette. #EJ28 indicates a Gen 3 blind
swap cassette. #EJ29 indicates a Gen 4 blind swap cassette. 2

#EJ27, EJ28 and EJ29 are identical to #4807, #4808 and #4809
adapters which were manufactured after 2012, but different from #4807,
#4808 and #4809 adapters manufactured prior to 2012.

Other IBM PCIe Cryptographic Coprocessor adapter highlights

Integrated Dual processors that operate in parallel for higher
reliability

Supports IBM Common Cryptographic Architecture or PKCS#11 standard

Ability to configure adapter as coprocessor or accelerator

Support for smart card applications using Europay, MasterCard and
Visa

Cryptographic key generation and random number generation

PIN processing - generation, verification, translation

Encrypt/Dectrypt using AES and DES keys

Refer to the following URL for the latest firmware and
software updates.

Attributes required: One full-high PCIe slot using a gen3 blind
swap cassette such as found in an EM0X Gen3 I/O drawer or
#5802/5803/5873/5877 12X-attached I/O drawer

Minimum required: 0

Maximum allowed: 10 (Initial order maximum: 10)

OS level required:

Linux not supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJ33) - PCIe3 Crypto Coprocessor BSC-Gen3 4767

Secure-key adapter provides both cryptographic coprocessor and
cryptographic accelerator functions in a single PCIe card. The
adapter is well suited to applications requiring high-speed,
security-sensitive, RSA acceleration, cryptographic operations for
data encryption and digital signing, secure management, and use of
cryptographic keys, or custom cryptographic applications. It provides
secure storage of cryptographic keys in a tamper-resistant hardware
security module designed to meet FIPS 140-2 level 4 security
requirements. The adapter is a PCIe Gen 1 x4 full height - short
card. The adapter runs in dedicated mode only (no PowerVM
virtualization).

#EJ32 and #EJ33 are both feature codes representing the same
physical card with the same CCIN of 4767. Different feature codes are
used to indicate if a blind swap cassette is used and its type. #EJ32
indicates no blind swap cassette. #EJ33 indicates a Gen 3 blind swap
cassette.

(#EJPJ) - Specify mode-2 (1)5901/5278 for EXP24 #5887 or #EL1S

Feature indicates that one EXP24 SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 2 (two groups of 12 drive bays) with only
one group of 12 drive bays enabled by connecting to one PCIe 3G SAS
controller (5901/5278) using one appropriate YO cable. IBM does not
provide changes to the mode setting after #5887 or #EL1S is shipped."

Attributes required: One YO cable (#3691/#EL1Z 1.5M, #3692/#EL20
3M, #3693/#EL1W 6M, #3694/#EL1U 15M) and one (5901/5278) dedicated to
the single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJPK) - Specify mode-2 (2)5901/5278 for EXP24 #5887 or #EL1S

Feature indicates that one EXP24 SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 2 (two groups of 12 drive bays) with only
one group of 12 drive bays enabled by connecting to two PCIe 3G SAS
controllers (5901/5278) using one appropriate X cable. IBM does not
provide changes to the mode setting after #5887 or #EL1S is shipped.

Attributes required: One dual X cable (#3661/#EL22 1 1.5M,
#3662/#EL23 3M, #3663/#EL21 15M) and two (5901/5278) dedicated to the
single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJPL) - Specify mode-4 (1)5901/5278 for EXP24 #5887 or #EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 4 (four groups of 6 drive bays) with only
one group of 6 drive bays enabled by connecting to one PCIe 3G SAS
controller (5901/5278) using one appropriate X cable connecting to
only one adapter port. IBM does not provide changes to the mode
setting after #5887 or #EL1S is shipped.

Attributes required: One dual X cable #3661/#EL22 1 1.5M,
#3662/#EL23 3M, #3663/#EL21 15M) and one (5901/5278) dedicated to the
single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJPM) - Specify mode-4 (2)5901/5278 for EXP24 #5887 or #EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 4 (four groups of 6 drive bays) with two
groups of 6 drive bays enabled by connecting to two PCIe 3G SAS
controllers (5901/5278) using one appropriate X cable connecting to
only one adapter port on each controller. IBM does not provide
changes to the mode setting after #5887 or #EL1S is shipped.

Attributes required: One dual X cable (#3661/#EL22 1 1.5M,
#3662/#EL23 3M, #3663/#EL21 15M) and two (5901/5278) dedicated to the
single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJPN) - Specify mode-4 (3)5901/5278 for EXP24 #5887 or #EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 4 (four groups of 6 drive bays) with three
groups of 6 drive bays enabled by connecting to three PCIe 3G SAS
controllers (5901/5278) using two appropriate X cables connecting to
only one adapter port on each controller. IBM does not provide
changes to the mode setting after #5887 or #EL1S is shipped.

Attributes required: Two dual X cables (#3661/#EL22 1 1.5M,
#3662/#EL23 3M, #3663/#EL21 15M) and three (5901/5278) dedicated to
the single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJPR) - Specify mode-2 (2)5903/5805 for EXP24 #5887 or #EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 2 (two groups of 12 drive bays) with only
one group of 12 drive bays enabled by connecting to one pair of PCIe
3G SAS RAID controllers (5903/5805) using one appropriate X cable.
IBM does not provide changes to the mode setting after #5887 or #EL1S
is shipped.

Attributes required: One dual X cable #3661/#EL22 1 1.5M,
#3662/#EL23 3M, #3663/#EL21 15M) and two (5903/5805) dedicated to the
single #5887 or #EL1S

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required:

IBM i not supported.

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EJPT) - Specify mode-2 (2)5913 for EXP24 #5887 or #EL1S

Feature indicates that EXP24S SFF Gen2-bay Drawer (#5887 or #EL1S)
will be configured to Mode 2 (two groups of 12 drive bays) with only
one group of 12 drive bays enabled by connecting to one pair of PCIe2
1.8GB Cache RAID SAS Adapter Tri-port 6Gb controllers (5913) using one
appropriate 6G X cable. IBM does not provide changes to the mode
setting after #5887 or #EL1S is shipped.

(#EJR1) - Specify Mode-1 1(1) EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 1 and be using one PCIe3 RAID SAS
adapter (#EJ0J/EJ0M/EL3B) and one 6G YO SAS Cable.

(#EJR2) - Specify Mode-1 1(2)EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 1 and be using two (one pair)
PCIe3 RAID SAS adapters (#EJ0J/EJ0M/EL3B) and two 6G YO SAS Cables.

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 2 and be using two (one pair) RAID
PCIe3 SAS adapters (#EJ0J/EJ0M/EL3B) and two 6G X SAS Cables.

(#EJR4) - Specify Mode-2 1(4) EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be
configured in Mode 2 and be using four (two pair) PCIe3 RAID SAS
adapters (#EJ0J/EJ0M/EL3B) and two 6G X SAS Cables.

(#EJR5) - Specify Mode-4 1(4) EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be
configured in Mode 4 and be using four (unpaired) PCIe3 RAID SAS
adapters (#EJ0J/EJ0M/EL3B) and two 6G X SAS Cables.

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be
configured in Mode 2 and be using one PCIe3 RAID SAS adapter
(#EJ0J/EJ0M/EL3B) and two 6G YO SAS Cables.

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be
configured in Mode 2 and be using two (nonpaired) PCIe3 RAID SAS
adapters (#EJ0J/EJ0M/EL3B) and two 6G YO SAS Cables.

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 2 and be using one PCIe3 RAID SAS
adapter (#EJ0J/EJ0M/EL3B) and one 6G YO SAS Cables.
This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24S.
If adapters/cables are MES added later to support the rest of
the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools
understand the expanded usage, probably specify feature #EJR7.

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 2 and be using two PCIe3 RAID SAS
adapters (#EJ0J/EJ0M/EL3B) and one 6G X SAS Cable.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24S.
If adapters/cables are MES added later to support the rest of
the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools
understand the expanded usage, probably specify feature #EJR4.

(#EJRC) - Specify Mode-4 1(1) EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 4 and be using one PCIe3 RAID SAS
adapter (#EJ0J/EJ0M/EL3B) and one 6G X SAS Cables. One leg of
the X cable is left unattached at the adapter end.

This indicates a "partial" configuration where there are
not enough adapters/cables to run all the SAS bays in the
EXP24S. If adapters/cables are MES added later to support the
rest of the SAS bays, then this specify code should be removed
and the appropriate specify feature added to help IBM config
tools understand the expanded usage, probably specify feature
#EJR5, #EJRD or #EJRE.

(#EJRD) - Specify Mode-4 1(2) EJ0J/EJ0M/EL3B for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 4 and be using two (nonpaired)
PCIe RAID SAS adapters (#EJ0J/EJ0M/EL3B) and one 6G X SAS
Cables.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24S.
If adapters/cables are MES added later to support the rest of
the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools
understand the expanded usage, probably specify feature #EJR5
or #EJRE.

(#EJRE) - Specify Mode-4 1(3)EJ0J/EJ0M/EL3B for EXP24S (#5888/EL1S)

This feature does not order or ship any hardware, but indicates
to IBM config tools and to IBM Manufacturing that in this
configuration one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S)
should be configured in Mode 4 and be using three (nonpaired)
PCIe3 RAID SAS adapter (#EJ0J/EJ0M/EL3B) and two 6G X SAS
Cables. One leg of one X cable is left unattached at the
adapter end.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24S.
If adapters/cables are MES added later to support the rest of
the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools
understand the expanded usage, probably specify feature #EJR5.

(#EJRF) - Specify Mode-1 1(2)EJ14 for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 1 and be using two PCIe3 12GB Cache RAID PLUS SAS Adapter
Quad-port 6Gb x8 (#EJ14) and two 6G YO SAS Cables.

(#EJRG) Specify Mode-2 1(2)EJ14 1(2) X for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using two PCIe3 12GB Cache RAID PLUS SAS Adapter
Quad-port 6Gb x8 (#EJ14) and two 6G X SAS Cables.

(#EJRH) Specify Mode-2 1(2)EJ14 1(1) X for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using two PCIe3 12GB Cache RAID PLUS SAS Adapter
Quad-port 6Gb x8 (#EJ14) and one 6G X SAS Cable.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24S. If
adapters/cables are MES added later to support the rest of the SAS
bays, then this specify code should be removed and the appropriate
specify feature added to help IBM config tools understand the
expanded usage.

(#EJRJ) - Specify Mode-2 1(4) EJ14 for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using four PCIe3 12GB Cache RAID PLUS SAS Adapter
Quad-port 6Gb x8 (#EJ14) and two 6G X SAS Cable.

(#EJRL) -Non-paired Indicator EJ14 PCIe SAS RAID+ Adapter

Feature EJRU must be added for every instance of a non-paired SAS
RAID adapter #E14L. It identifies a specific high availability
configuration supported by AIX or Linux which has one #E14L on one
system and the paired #EJ14 located on a second system. IBM i does
not support paired adapter on different servers.

SAS RAID adapter firmware disables write cache until a second SAS
RAID adapter is recognized as its pair. This specify indicates the
pairing will not be on just one server.

Attributes required: Only one #EJ14 on a server and its pair on a
different server.

Minimum required: 0

Maximum allowed: 128 (Initial order maximum: 128)

OS level required: Refer to #EJ14

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJRP) - Specify Mode-1 1(2) EJ0L for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 1 and be using two PCIe3 12GB Cache RAID SAS adapters (#EJ0L) and
two 6G YO SAS Cables.

Attributes required: Two dual X cables, four AA cables and one
port on each of four #EJ0Ls dedicated to the single #5887/EL1S.

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EJRS) Specify Mode-2 1(2)EJ0L 1(2) X for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using two PCIe3 12GB RAID SAS adapters (#EJ0L) and two
6G X SAS Cables.

(#EJRT) Specify Mode-2 1(2)EJ0L 1(1) X for EXP24S (#5887/EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using two 12GB Cache RAID SAS adapters (#EJ0L) and one
6G X SAS Cable.

This indicates a "partial" configuration where there are not enough
adapters/cables to run all the SAS bays in the EXP24S. If
adapters/cables are MES added later to support the rest of the SAS
bays, then this specify code should be removed and the appropriate
specify feature added to help IBM config tools understand the expanded
usage.

(#EJRU) - Non-paired Indicator EJ0L PCIe SAS RAID Adapter

Feature EJRU must be added for every instance of a non-paired SAS
RAID adapter #EJ0L. It identifies a specific high availability
configuration supported by AIX or Linux which has one #EJ0L on one
system and the paired #EJ0L located on a second system. IBM i does
not support paired adapter on different servers.

Attributes required: Every #EJ0L requires a 6Gb/s SAS RAID
adapter (#EJ0L) on both this server and on another server that will
pair up the SAS RAID adapter and enable the onboard caches to
function.

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EJS1) - Non-paired Indicator ESA3 PCIe SAS RAID Adapter

Feature EJS1 must be added for every instance of a non-paired SAS
RAID adapter #ESA3. It identifies a specific high availability
configuration supported by AIX or Linux which has one #ESA3 on one
system and the paired #ESA3 located on a second system. IBM i does
not support paired adapter on different servers.

Attributes required: Every #ESA3 requires a 6Gb/s SAS RAID
adapter CR (#ESA3) on both this server and on another server that will
pair up the SAS RAID adapter and enable the onboard caches to
function.

Minimum required: 0

Maximum allowed: 48 (Initial order maximum: 48)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EJS2) - Specify Mode-2 1(2) ESA3 for EXP24S (#5887/#EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using two (one pair) of PCIe2 1.8GB RAID SAS adapter
(#ESA3) and one X SAS Cable.

This indicates a "partial" configuration where there are not enough
adapters/cables to run all the SAS bays in the EXP24S. If additional
adapters/cables are MES added later to support the rest of the SAS
bays, then this specify code should be removed and the appropriate
specify feature added to help IBM config tools understand the expanded
usage, probably specify feature #EJS4.

(#EJS3) - Specify Mode-1 1(2)ESA3 for EXP24S (#5887/#EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 1 and be using two (one pair) of PCIe2 1.8GB RAID SAS adapter
(#ESA3) and two YO SAS Cables.

(#EJS4) - Specify Mode-2 1(4)ESA3 for EXP24S (#5887/#EL1S)

This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24S SFF Gen2-bay Drawer (#5887/#EL1S) should be configured in
Mode 2 and be using four (two pairs) of PCIe2 1.8GB RAID SAS adapters
(#ESA3) and two X SAS Cables.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 1 (one group
of 12 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

One YO12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 1 (one group
of 12 drive bays). It will be connected to two (one pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two (one pair) #EJ0J/#EL59/
#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to four (two pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 4 (four groups
of 3 drive bays). It will be connected to four (unpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two (nonpaired) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two YO12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBT, #ECBU, #ECBV or#ECBW.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

One YO12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBT, #ECBU, #ECBV or#ECBW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a "partial" configuration where
there are not enough adapters/cables to run all the SAS bays in the
SAS Storage Enclosure. If adapters/cables are MES added later to
support the rest of the SAS bays, then this specify code should be
removed and the appropriate specify feature added to help IBM config
tools understand the expanded usage, probably specify feature #EJV7.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two (one pair) #EJ0J/#EL59/
#EJ0M/#EL36 PCIe3 RAID SAS adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a "partial" configuration where
there are not enough adapters/cables to run all the SAS bays in the
SAS Storage Enclosure. If adapters/cables are MES added later to
support the rest of the SAS bays, then this specify code should be
removed and the appropriate specify feature added to help IBM config
tools understand the expanded usage, probably specify feature #EJV4.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 4 (four groups
of 3 drive bays). It will be connected to one #EJ0J/#EL59/
#EJ0M/#EL36 PCIe3 RAID SAS adapter.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL. One leg of the
X12 cable is left unattached at the adapter end.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJV5, #EJVD or #EJVE.
This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP12SX Drawer (#ESLL/ELLL) should be configured in Mode 4 and be
using one PCIe3 RAID SAS adapter (#EJ0J/EL59/EJ0M/EL3B) and one X12
SAS Cables. One leg of the X cable is left unattached at the adapter
end. This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP12SX. If
adapters/ cables are MES added later to support the rest of the SAS
bays, then this specify code should be removed and the appropriate
specify feature added to help IBM config tools understand the
expanded usage, probably specify feature #EJV5, #EJVD or #EJVE.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 4 (four groups
of 3 drive bays). It will be connected to Two (nonpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJV5 or #EJVE.

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 4 (four groups
of 3 drive bays). It will be connected to Three (nonpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL. One leg of one
of the two X12 cables is left unattached at the adapter end.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJV5.

(#EJVF) - Specify Mode-1 1(2)EJ14 1(2)YO12 for EXP12SX #ESLL/ELLL

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 1 (one group
of 12 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVG) - Specify Mode-2 1(2)EJ14 1(2)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVH) - Specify Mode-2 1(2)EJ14 1(1)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJVG.

(#EJVJ) - Specify Mode-2 1(4)EJ14 1(2)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to four #EJ14 PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVP) - Specify Mode-1 1(2)EJ0L 1(2)YO12 for EXP12SX #ESLL/ELLL

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 1 (one group
of 12 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVR) - Specify Mode-2 1(4)EJ0L 1(2)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to four #EJ0L PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVS) - Specify Mode-2 1(2)EJ0L 1(2)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJVT) - Specify Mode-2 1(2)EJ0L1(1)X12 for EXP12SX #ESLL/ELLL

(No Longer Available as of January 24, 2017)

Feature indicates that EXP12SX SAS Storage Enclosure (#ESLL or
#ELLL) will be configured by IBM Manufacturing in Mode 2 (two groups
of 6 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJVG.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 1 (one group
of 24 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

One YO12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 1 (one group
of 24 drive bays). It will be connected to two (one pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two (one pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to four (two pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 4 (four groups
of 6 drive bays). It will be connected to four (unpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two (nonpaired) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two YO12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBT, #ECBU, #ECBV or#ECBW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to one #EJ0J/#EL59/#EJ0M/
#EL36 PCIe3 RAID SAS adapter.

One YO12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBT, #ECBU, #ECBV or#ECBW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a "partial" configuration where
there are not enough adapters/cables to run all the SAS bays in the
SAS Storage Enclosure. If adapters/cables are MES added later to
support the rest of the SAS bays, then this specify code should be
removed and the appropriate specify feature added to help IBM config
tools understand the expanded usage, probably specify feature #EJW7.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two (one pair) #EJ0J/
#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a "partioal" configuration where
there are not enough adapters/cables to run all the SAS bays in the
SAS Storage Enclosure. If adapters/cables are MES added later to
support the rest of the SAS bays, then this specify code should be
removed and the appropriate specify feature added to help IBM config
tools understand the expanded usage, probably specify feature #EJW4.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 4 (four groups
of 6 drive bays). It will be connected to one #EJ0J/#EL59/
#EJ0M/#EL36 PCIe3 RAID SAS adapter.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL. One leg of the
X12 cable is left unattached at the adapter end.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJW5, #EJWD or #EJWE.
This feature does not order or ship any hardware, but indicates to
IBM config tools and to IBM Manufacturing that in this configuration
one EXP24SX Drawer (#ESLS/ELLS) should be configured in Mode 4 and be
using one PCIe3 RAID SAS adapter (#EJ0J/EL59/EJ0M/ EL3B) and one X12
SAS Cables. One leg of the X cable is left unattached at the adapter
end. This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the EXP24SX. If
adapters/cables are MES added later to support the rest of the SAS
bays, then this specify code should be removed and the appropriate
specify feature added to help IBM config tools understand the
expanded usage, probably specify feature #EJW5, #EJWD or #EJWE.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 4 (four groups
of 6 drive bays). It will be connected to Two (nonpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJW5 or #EJWE.

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 4 (four groups
of 6 drive bays). It will be connected to Three (nonpaired)
#EJ0J/#EL59/#EJ0M/#EL36 PCIe3 RAID SAS adapters.

Two X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL. One leg of one
of the two X12 cables is left unattached at the adapter end.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJW5.

(#EJWF) - Specify Mode-1 1(2)EJ14 1(2)YO12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 1 (one group
of 24 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWG) - Specify Mode-2 1(2)EJ14 1(2)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWH) - Specify Mode-2 1(2)EJ14 1(1)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two #EJ14 PCIe3 RAID SAS
adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJWG.

(#EJWJ) - Specify Mode-2 1(4)EJ14 1(2)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to four #EJ14 PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWP) - Specify Mode-1 1(2)EJ0L 1(2)YO12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 1 (one group
of 24 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

Two YO12 cables connect the SAS Storage Enclsoure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECDT, #ECDU, #ECDV or #ECDW.

One specify feature should ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWR) - Specify Mode-2 1(4)EJ0L 1(2)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to four #EJ0L PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWS) - Specify Mode-2 1(2)EJ0L 1(2)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

Two X12 cables connect the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This specify feature indicates a full set of adapters plus SAS
cables is used for the enclosure. A different specify feature code is
used to indicate a "partial" configuration with a subset of adapters
and cables is used.

(#EJWT) - Specify Mode-2 1(2)EJ0L1(1)X12 for EXP24SX #ESLS/ELLS

Feature indicates that EXP24SX SAS Storage Enclosure (#ESLS or
#ELLS) will be configured by IBM Manufacturing in Mode 2 (two groups
of 12 drive bays). It will be connected to two #EJ0L PCIe3 RAID SAS
adapters.

One X12 cable connects the SAS Storage Enclosure to the SAS
adapter ports. Use the cable length appropriate to the configuration
and select from features such as #ECBJ, #ECBK or#ECBL.

One specify feature should be ordered with each SAS Storage
Enclosure. This feature does not order or ship any hardware, but
indicates to IBM config tools and to IBM Manufacturing the
combination of enclosure mode, SAS adapter and SAS cable type which
will be used.

This indicates a "partial" configuration where there are not
enough adapters/cables to run all the SAS bays in the SAS Storage
Enclosure. If adapters/cables are MES added later to support the rest
of the SAS bays, then this specify code should be removed and the
appropriate specify feature added to help IBM config tools understand
the expanded usage, probably specify feature #EJWG.

(#ELC0) - PDU Access Cord 0.38m

This 0.38 meter (14 inch) cord is used with a vertically mounted
PDU (Power Distribution Unit) such as a #7188 or #7109 when the PDU is
located in a 7953-94Y or #ER05 Slim Rack. One end of this power cord
connects to the PDU. The other end of this cord connects to the power
cord running to the wall outlet or electrical power source.

One PDU Access Cord is required per vertically mounted PDU.
Without a PDU Access Cord, inserting and removing the wall outlet
power cord into the PDU can be very difficult in the narrow side
pockets of the Slim Rack. A PDU Access Cord is not required for PDUs
in wider racks such as the 7014-T42 or #0553.

Attributes provided: Power cord

Attributes required: None

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#ELJ6) - Power IFL Processor Activation

This no-charge feature delivers Qty = 4 processor activations in
support of a Power IFL chargeable package feature such as FC#ELJG

(#ELJG) - Power Integrated Facility for Linux Package

This is a chargeable feature code which will cause IBM
configuration tools to add the proper quantity of no-charge
features associated with processor activations, such as #ELJ5,
#ELJ6, #ELJ7 and memory activation #ELJH. When the number of
#ELJG features ordered are equal to defined system minimum order
requirement for static processor and memory activations, the
activation requirement is satisfied.

Attributes provided: Power IFL package of processor and memory
activation and PowerVM

This is a chargeable feature code that will cause IBM configuration
tools to add the proper quantity of no-charge features associated with
processor activations, such as #ELAA and memory activations #ELMA.
When the number of #ELPA features ordered are equal to defined system
minimum order requirements for static processor and memory
activations, the activation requirement is satisfied.

(#ELS0) - Boot From Existing Drive Specified

(No Longer Available as of July 8, 2016)

This feature is used when the client is providing their own EXP24S
SFF Gen2-bay Drawer (#5887/EL1S) and a disk which the new system will
boot from. Client will need to do a Record Purpose only (RPO) off
EXP24S SFF Gen2-bay Drawer from existing server and then RPO on to the
new server.

Attributes provided: None

Attributes required: None

Minimum required: 0

Maximum allowed: 1 (Initial order maximum: 1)

OS level required:

AIX - supported

IBM i - supported

Linux - supported

Initial Order/MES/ Both/Supported: Both

CSU: Yes

Return parts MES: No

(#ELS9) - #ES1A Load Source Specify (387GB SSD SFF-2)

This specify code indicates that a #ES1A Solid State Drive is being
used as the Load Source.

(#EMJ9) - $0 576 GB-Days of On/Off CoD Temporary Memory Resources

(No Longer Available as of August 10, 2016)

Provides a billing credit for 576 GB-Days of Elastic CoD memory
resources. #EMJ9 features must be ordered with a new server and the
quantity of #EMJ9 depends on the quantity of processor modules and the
number of processor cores per module. The credit is tracked by IBM
and is used to offset billing days ordered by the client for the
specific server for which it was ordered. This billing credit is not
transferable to a different server or different model. #EMJ9 is not
orderable by MES orders.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EM9T), usage reporting, billing
feature ordering (1GB memory-day or 100GB memory-days or 999GB memory
days), and billing processes are used on the server and are a pre-
requisite to using the credit balance.

For example, if ten #EMJ9 features had been ordered for a Power E880
or E880C #EPBD processor module feature, then credit for 5,760 GB memory
days would be known by IBM.
And if for example there were 576 GB of
memory temporarily activated using Elastic CoD, then that memory could
be used for 10 days and their billing features would be offset by the
#EMJ9 generated credit balance.

Initial Order minimum is equal to two identical features (4x16GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

(#EM8K) - 128GB (4X32GB) CDIMMs, 1600 MHz, 4GBIT DDR3 DRAM

Memory Feature is a total of 128GB of memory. 4 each 32GB CDIMMs,
1600 MHz, 4GBIT DDR3 DRAM, Short with Extender.

Initial Order minimum is equal to two identical features (4x32GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

(#EM8L) - 256GB (4X64GB) CDIMMs, 1600 MHz, 4GBIT DDR3 DRAM

Memory Feature is a total of 256GB of memory. 4 each 64GB CDIMMs,
1600 MHz, 4GBIT DDR3 DRAM, DDP Short with Extender.

Initial Order minimum is equal to two identical features (4x64GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

(#EM8M) - 512GB (4X128GB) CDIMMs, 1600 MHz, 4GBIT DDR3 DRAM

Memory Feature is a total of 512GB of memory. 4 each 128GB CDIMMs,
1600 MHz, 4GBIT DDR3 DRAM, DDP Tall.

Initial Order minimum is equal to two identical features (4x128GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

(#EM9T) - 90 Days Elastic CoD Memory Enablement

The 90 Days Elastic CoD Memory Enablement (#EM9T) feature can be
ordered to enable your server for Elastic Capacity on Demand. You must
sign an
Elastic Capacity on Demand contract and the Sales Channel has to register
the server before an MES order can be fulfilled for #EM9T. Feature code
#EM9T provides access to Elastic CoD memory resources for 90 days. Access
to these resources is measured in GB-days. For example, if there are 100
GBs of inactive, CoD memory installed, this feature will enable usage of
9000 GB-days (9 TB-days) before disabling Elastic CoD access to the
inactive memory.

After usage of this feature has been exhausted, additional temporary
usage of Elastic CoD memory resources may be activated by ordering
another Elastic CoD enablement code, #EM9T. Before ordering a fourth
enablement code, after the system maximum of 3 has been reached, the
current code must be RPO deleted from the configuration file with a MES
delete order.

Usage of Elastic temporary CoD memory is billed by ordering #EMJ4 -
8 GB-Day Memory Activation, #EMJ5 - Quantity of 800 #EMJ4 Memory
Activations, and #EMJ6 - Quantity of 999 #EMJ4 Memory Activations. The
9119 - MHE maximum Elastic GB-days billable on any one order using #EMJ4,
#EMJ5 and #EMJ6 is 6,144,000 GBs. The 9119 - MME maximum Elastic GB-days
billable on any one order using #EMJ4, #EMJ5 and #EMJ6 is 3,072,000 GBs.

Initial Order minimum is equal to two identical features (4x256GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

Initial Order minimum is equal to two identical features (4x256GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

Initial Order minimum is equal to two identical features (4x256GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

Initial Order minimum is equal to two identical features (4x256GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

Initial Order minimum is equal to two identical features (4x256GB
CDIMM) per processor drawer.

Minimum of 50% of Initial memory orders must be activated via
memory activation feature #EMA5 and #EMA6 at the time of purchase.

MES order quantity can equal one memory feature (4 CDIMMs) per
processor drawer.

Memory sizes can differ across CPU modules, but the eight CDIMM
slots connected to the same processor module must be filled with
identical CDIMMs.

CDIMMs must be evenly distributed across processor modules within
a drawer, always filling up four or eight CDIMM slots per module, and
balanced across drawers; placement order must take into account the
number of drawers.

If memory is either removed and/or moved to another system, none
of that memory carries with it any activations. All CoD activations
remain with the system unless authorization to remove the activations
is obtained from the CoD project office and installed.

(#EMB6) - Bundle of eight #EM8M, 512GB 1600 MHz Memory

When ordering #EMB6, a quantity of 8 of #EMB7 must be ordered for
each #EMB6 ordered.

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

(#EMB7) - Memory Activations for #EMB6 or #EMBA

It must be ordered in quantities of 8 (providing a total of 4 TB of
memory activation).

This feature must be ordered with #EMB6. For each #EMB6 ordered a
quantity of 8 #EMB7 features must be ordered.

This feature must be ordered with #EMBA. For each #EMBA ordered a
quantity of 16 #EMB7 features must be ordered.

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

Attributes provided: Activates memory hardware only in the system
serial number for which they are purchased.

Attributes required: #EMB6 or #EMBA bundle memory feature.

Minimum required: 0

Maximum allowed: 64 (Initial order maximum: 64)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: #EMB7 is not supported in China and Korea.

(#EMB8) - Five Hundred and Twelve Memory Activations for IFL

This CoD memory feature makes available the equivalent of five
hundred and twelve FC #EMA5, Quantity of 1 GB of DDR3/DDR4, POWER8
memory activation.

#EMB8 is only available when one of the following Power IFL
(Integrated Facility for Linux) FC #ELAA, #ELAC, #EPBF, #ELJ5, #ELJ6,
or #ELJ7 is ordered or already on the system. The quantity of #EMB8
must be equal to or less than the quantity of IFL FCs #ELAA, #ELAC,
#EPBF, #ELJ5, #ELJ6, or #ELJ7 on the system (so minimum of 512GB
inactive memory must be available on the system).

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

Attributes provided: Activates memory hardware only in the system
serial number for which they are purchased.

Attributes required: Inactive Memory available for temporary use
installed

(#EMBA) - Bundle of eight #EM8Y, 1024GB 1600 MHz DDR4 Memory

When ordering #EMBA, static activation EMA5 or EMA6 or
EMB7 must be ordered for each #EMBA ordered.

Feature #EMA5 will permanently activate 1 GB of memory.

Feature #EMA6 will permanently activate 100 GB of memory.

Feature #EMB7 makes available 512GB of memory activations (It must
be ordered in quantities of 8 (providing a total of 4 TB of memory
activation).

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

(#EMBB) - Memory Activations for #EMBA

It must be ordered in quantities of 8 (providing a total of 8TB of
memory activation). This feature must be ordered with #EMBA. For
each #EMBA ordered 8 #EMBB features must be ordered.

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

Attributes provided: Activates memory hardware only in the system
serial number for which they are purchased.

Attributes required: #EMBA memory feature

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: #EMBB is not supported in China and Korea.

(#EMBC) - 1,024 Memory Activations for IFL

#EMBC is only available when one of the following Power IFL
(Integrated Facility for Linux) FC #ELJ5, #ELJ6, or #ELJ7 is ordered
or already on the system. The quantity of #EMB8 must be equal to or
less than the quantity of IFL FC's #ELJ5, #ELJ6, or #ELJ7 on the
system (so minimum of 1024GB inactive memory must be available on the
system).

Memory CoD activations will activate memory hardware only in the
system serial number for which they are purchased. If you move memory
hardware to another system, the memory may not be functional in that
system until arrangements are made to move the memory activations or
purchase additional memory activations.

Attributes provided: Activates memory hardware only in the system
serial number for which they are purchased.

Attributes required: Inactive Memory available for temporary use
installed

Minimum required: 0

Maximum allowed: 48 (Initial order maximum: 48)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: #EMBC is not supported in China and Korea.

(#EME0) - Static to Mobile Memory Auto Conversion

Feature is used to order automatic conversion of a static action to
a mobile activation without IBM intervention. The term "static
activations" also includes Mobile Enabled activations which are
technically static. Firmware 840 or later is a pre-requisite. Use of
this feature on servers with firmware 830 or earlier will require IBM
intervention through the pcod@us.ibm.com

Attributes provided: Automated conversion

Attributes required: Firmware 840

Minimum required: 0

Maximum allowed: 246 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EMJ4) - 8 GB-Day billing for Elastic CoD memory

Provides 8 GB-Day of billing based on the usage of #EM9T - 90 Days
Elastic CoD Memory Enablement.

Attributes provided: N/A

Attributes required: Elastic Capacity on Demand contract

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EMJ5) - 800 GB-Day billing for Elastic CoD memory

Provides 800 GB-Days of billing based on the usage of #EM9T - 90 Days
Elastic CoD Memory Enablement.

Attributes provided: N/A

Attributes required: Elastic Capacity on Demand contract

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EMJ6) - 999 GB-Day billing for Elastic CoD memory

(No Longer Available as of November 8, 2016)

Provides 999 GB-Days of billing based on the usage of #EM9T - 90
Days Elastic CoD Memory Enablement.

Attributes provided: N/A

Attributes required: Elastic Capacity on Demand contract

Minimum required: 0

Maximum allowed: 100 (Initial order maximum: 100)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EMJ8) - 384 GB-Days of Elastic CoD Memory Resources

(No Longer Available as of August 10, 2016)

Provides access a billing credit for 384 GB-Days of Elastic CoD
memory resources. #EMJ8 features must be ordered with a new server or
with an upgrade to a new server and the quantity of #EMJ8 depends on the
quantity of system nodes and the number of processor cores per node. The
credit is tracked by IBM and is used to offset billing days ordered by
the client for the specific server for which it was ordered. This
billing credit is not transferable to a different server or different
model. #EMJ8 is not orderable by MES orders except with a model change
upgrade.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EM9T), usage reporting, billing feature
ordering (1GB memory-day or 100GB memory-days or 999GB memory days), and
billing processes are used on the server and are a pre-requisite to using
the credit balance.

For example, if twenty #EMJ8 features had been ordered for a Power
E880 or E880C #EPBB processor feature, then credit for 7,680 GB memory
days would be known by IBM.
And if for example there were 768 GB of memory
temporarily activated using Elastic CoD, then that memory could be used
for 10 days and their billing features would be offset by the #EMJ8
generated credit balance.

Attributes required: New Power E880 or E880C server or model upgrade
to E880 or E880C.

Minimum required: 0

Maximum allowed: 80 (Initial order maximum: 80)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EMJD) - 480 GB-Days of Elastic CoD Memory Resources

(No Longer Available as of August 10, 2016)

Provides access a billing credit for 480 GB-Days of Elastic CoD
memory resources. #EMJD features must be ordered with a new server or
with an upgrade to a new server and the quantity of #EMJD depends on
the quantity of system nodes and the number of processor cores per
node. The credit is tracked by IBM and is used to offset billing days
ordered by the client for the specific server for which it was
ordered. This billing credit is not transferable to a different
server or different model. #EMJD is not orderable by MES orders
except with a model change upgrade.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EM9T), usage reporting, billing
feature ordering (1GB memory-day or 100GB memory-days or 999GB memory
days), and billing processes are used on the server and are a
pre-requisite to using the credit balance.

For example, if twenty #EMJD features had been ordered for a Power
E880 or E880C #EPBS processor feature, then credit for 7,680 GB memory
days would be known by IBM.
And if for example there were 768 GB of memory
temporarily activated using Elastic CoD, then that memory could be
used for 10 days and their billing features would be offset by the
#EMJD generated credit balance.

Attributes required: New Power E880 or E880C server or model upgrade
to E880 or E880C.

Minimum required: 0

Maximum allowed: 80 (Initial order maximum: 80)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EMX0) - PCIe Gen3 I/O Expansion Drawer

This 19-inch, 4U (4 EIA) enclosure provides PCIe Gen3 slots outside of
the system unit. It has two module bays. One 6-Slot Fanout Module
(#EMXF or EMXG) placed in each module bay. Two 6-slot modules provide a
total of 12 PCIe Gen3 slots. Each fanout module is connected to a PCIe3
Optical Cable Adapter located in the system unit over an active optical
cable (AOC) pair.

The PCIe Gen3 I/O Expansion Drawer has two redundant, hotplug power
supplies. Each power supply has its own separately ordered power
cord. The two power cords plug into a Power supply conduit which
connects to the power supply. The single-phase AC power supply is
rated at 1030 Watt and can use 100-120V or 200-240V. If using
100-120V, then the maximum is 950 Watt. It's recommended the power
supply connect to a PDU in the rack. Power Systems PDUs are designed
for 200-240V electrical source.

The drawer has fixed rails which can accommodate racks with depths
from 27.5 inches to 30.5 inches.

Limitations:

#EMX0 has a cable management bracket located at the rear of the
drawer which swings up to provide service access to the PCIe adapters.
2U (2 EIA) of space is required to swing up the bracket. Thus the
drawer can not be placed in the very top 2U of a rack.

There is a power cord access consideration with vertically mounted
PDUs on the right hand side of the rack when viewed from the rear of
the rack. The #EMX0 cable management bracket makes accessing some of
the PDU outlets located at the same rack height as the #EMX0 drawer
more challenging. Using a horizontally mounted PDU or locating the
PDU or #EMX0 at a different vertical location is recommended.

(#EMXA) - AC Power Supply Conduit for PCIe3 Expansion Drawer

Provides two 320-C14 inlet electrical connections for two
separately ordered AC power cords with C13 connector plugs. Conduit
provides electrical power connection between two power supplies
located in the front of a PCIe Gen3 I/O Expansion Drawer (#EMX0) and
two power cords which connect on the rear of the PCIe Gen3 I/O
Expansion Drawer.

(#EMXB) - DC Power Supply Conduit for PCIe3 Expansion Drawer

Provides two DC inlet electrical connections for two separately
ordered DC power cords. Conduit provides electrical power connection
between two power supplies located in the front of a PCIe Gen3 I/O
Expansion Drawer (#EMX0) and two power cords which connect on the rear
of the PCIe Gen3 I/O Expansion Drawer.

The module has two CXP ports which are connected two CXP ports on a
PCIe Optical Cable Adapter #EJ05 or #EJ07 or #EJ08 depending on the
server selected. A pair of active optical CXP cables (AOC) or a pair
of CXP copper cables are used for this connection. The top CXP port
of the fanout module is cabled to the top CXP port of the PCIe3
Optical Cable Adapter. The bottom CXP port of the fanout module is
cabled to the bottom CXP port of the same PCIe3 Optical Cable Adapter.

(#EMXG) - PCIe3 6-Slot Fanout Module for PCIe3 Expansion Drawer

PCIe3 fanout module for PCIe Gen3 I/O Expansion Drawer. Provides
six PCIe Gen3 full high, full length slots (two x16 and four x8). The
PCIe slots are hot plug. With firmware 8.4 or later, the fanout
module supports concurrent maintenance, though obviously while
off-line all its PCIe slots are also off-line. Blind swap cassettes
(BSC) are used for the PCIe slots. The BSC are interchangeable with
the #5802/5877/5803/5873 12X PCIe I/O Drawer BSC.

The module has two CXP ports which are connected two CXP ports on
a PCIe Optical Cable Adapter #EJ05 or #EJ07 or #EJ08 depending on the
server selected. A pair of active optical CXP cables (AOC) or a pair
of CXP copper cables are used for this connection. The top CXP port
of the fanout module is cabled to the top CXP port of the PCIe3
Optical Cable Adapter. The bottom CXP port of the fanout module is
cabled to the bottom CXP port of the same PCIe3 Optical Cable
Adapter.

EMXG is a follow-on to the original EMXF fanout module. EMXG and
EMXF are functionally identical but EMXG implements a small physical
change to enable a larger set of potential PCIe adapters to be
housed. EMXG and EMXF can be intermixed in the same drawer. There is
no difference in firmware or software prerequisites for the EMXF and
EMXG. EMXG and for EMXF use the same BSC.

(#EN01) - 1m (3.3-ft), 10GbE'Net Cable SFP+ Act Twinax Copper

1m (3.3-ft) copper active twinax Ethernet cable which supports
Ethernet data transfer rates up to 10 Gb/s. The cable has a copper
twinax transceiver on each end which is placed in an SFP+ port of an
adapter and/or a switch. This cabling option can be a cost effective
alternative to optical cable for short reach link high-speed
connection.

Attributes provided: 10Gb/s copper active twinax Ethernet cable

Attributes required: One available SFP+ 10Gb/s Ethernet Port

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN02) - 3m (9.8-ft), 10Gb E'Net Cable SFP+ Act Twinax Copper

3m (9.8-ft) copper active twinax Ethernet cable which supports
Ethernet data transfer rates up to 10 Gb/s. The cable has a copper
twinax transceiver on each end which is placed in an SFP+ port of an
adapter and/or a switch. This cabling option can be a cost effective
alternative to optical cable for short reach link high-speed
connection.

Attributes provided: 10Gb/s copper active twinax Ethernet cable

Attributes required: One available SFP+ 10Gb/s Ethernet Port

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN03) - 5m (16.4-ft), 10Gb E'Net Cable SFP+ Act Twinax Copper

5m (16.4-ft) copper active twinax Ethernet cable which supports
Ethernet data transfer rates up to 10 Gb/s. The cable has a copper
twinax transceiver on each end which is placed in an SFP+ port of an
adapter and/or a switch. This cabling option can be a cost effective
alternative to optical cable for short reach link high-speed
connection.

Attributes provided: 10Gb/s copper active twinax Ethernet cable

Attributes required: One available SFP+ 10Gb/s Ethernet Port

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN0A) - PCIe3 16Gb 2-port Fibre Channel Adapter

PCIe Gen3 16 Gigabit dual-port Fibre Channel (FC) Adapter is a
high-performance 8x short form adapter based on the Emulex LPe16002B PCIe
Host Bus Adapter (HBA). The adapter provides two ports of 16 Gb Fibre
Channel capability using SR optics. Each port can provide up to 16 Gb
Fibre Channel functions simultaneously. The adapter was initially
announced as PCIe Gen2, but upgrades to the adapter firmware and
introduction of POWER8 servers have enabled PCIe Gen3.

Each port provides single initiator capability over a fibre link or
with NPIV, multiple initiator capability is provided. The ports are
SFP+ and include an optical SR transceiver. The ports have LC type
connectors and utilize shortwave laser optics. The adapter operates
at link speeds of 4, 8 and 16 Gbps and will automatically negotiate to
the highest speed possible. LEDs on each port provide information on
the status and link speed of the port.

The adapter connects to a Fibre Channel switch at 4Gb, 8Gb or 16Gb.
It can directly attach to a device without a switch at 16Gb.
Attachment without a switch is not supported at 4Gb or 8Gb.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Feature #EN0A and #EN0B are electronically identical. They differ
physically only that EN0A has a tail stock for full high PCIe slots
and #EN0B has a tail stock for low profile PCIe slots. CCIN is 577F
for both features.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: Two 16Gb FC ports (with LC connectors)

Attributes required: Available PCIe Gen2 slot in supported server

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

IBM i 7.1 with IBM 7.1.0 machine code RS710-S, or later

IBM i 7.2 TR4, or later

IBM i 7.3, or later

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 11, Service Pack 3, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#EN0B) - PCIe3 LP 16Gb 2-port Fibre Channel Adapter

PCIe Gen3 16 Gigabit dual-port Fibre Channel (FC) Adapter is a
high-performance 8x short form adapter based on the Emulex LPe16002B PCIe
Host Bus Adapter (HBA). The adapter provides two ports of 16 Gb Fibre
Channel capability using SR optics. Each port can provide up to 16 Gb
Fibre Channel functions simultaneously. The adapter was initially
announced as PCIe Gen2, but upgrades to the adapter firmware and
introduction of POWER8 servers have enabled PCIe Gen3.

Each port provides single initiator capability over a fibre link or
with NPIV, multiple initiator capability is provided. The ports are
SFP+ and include an optical SR transceiver. The ports have LC type
connectors and utilize shortwave laser optics. The adapter operates
at link speeds of 4, 8 and 16 Gbps and will automatically negotiate to
the highest speed possible. LEDs on each port provide information on
the status and link speed of the port.

The adapter connects to a Fibre Channel switch at 4Gb, 8Gb or 16Gb.
It can directly attach to a device without a switch at 16Gb under AIX
or VIOS or Linux. Attachment without a switch is not supported at 4Gb
or 8Gb.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Feature #EN0A and #EN0B are electronically identical. They differ
physically only that EN0A has a tail stock for full high PCIe slots
and #EN0B has a tail stock for low profile PCIe slots. CCIN is 577F
for both features.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: Two 16Gb FC ports (with LC connectors)

Attributes required: Available PCIe Gen2 slot in supported server

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

IBM i 7.1 with IBM 7.1.0 machine code RS710-S, or later

IBM i 7.2 TR4, or later

IBM i 7.3, or later

AIX supported

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 11, Service Pack 3, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

VIOS supported

(#EN0F) - PCIe2 LP 8Gb 2-Port Fibre Channel Adapter

This feature ships a two-port, 8 Gb PCIe Gen2 Fibre Channel
Adapter based on the QLogic QLE2562 Host Bus Adapter (HBA). Each port
provides single initiator capability over a fibre link. The ports
have LC type connectors and utilize shortwave laser optics. The
adapter operates at link speeds of 2, 4, and 8 Gbps and will
automatically negotiates to the highest speed possible. LEDs on each
port provide information on the status and link speed of the port.
The adapter connects to a Fibre Channel switch or can directly attach
to a Fibre Channel port on a supported storage unit. N_Port ID
Virtualization (NPIV) capability is supported through VIOS. Cables
are the responsibility of the customer. Use multimode fibre optic
cables with short-wave lasers that adhere to the following
specifications:

OM3 - multimode 50/125 micron fibre, 2000 MHz*km bandwidth

OM2 - multimode 50/125 micron fibre, 500 MHz*km bandwidth

OM1 - multimode 62.5/125 micron fibre, 200 MHz*km bandwidth

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables. The following table shows the supported
distances for the three different cable types at the three different
link speeds.

#EN0F and #EN0F are
electronically identical with the same CCIN of 578D. #EN0F has a low
profile tailstock bracket. #EN0G has a full high tailstock bracket.

See also feature #5273 or #5735 for a 2-port 8Gb Fibre Channel
adapter based on an Emulex adapter.

See also optional wrap plug feature #ECW0 which is a) required to
run some diagnostic procedures and b) in some cases may speed system
boot when placed in empty ports.

Attributes provided: Dual port Fibre Channel adapter

Attributes required: Low Profile PCIe slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Red Hat Enterprise Linux 7.2, little endian, or later

Red Hat Enterprise Linux 7.2, big endian, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

SUSE Linux Enterprise Server 12, Service Pack 1, or later

Ubuntu 16.04, or later

AIX Version 7.2 with the 7200-00 Technology Level and Service
Pack 2

AIX Version 7.1 with the 7100-04 Technology Level and Service
Pack 2

AIX version 7.1 with the 7100-03 Technology Level and Service
Pack 7 (planned availability September 8, 2016)

AIX Version 6.1 with the 6100-09 Technology Level and Service
Pack 7

IBM i supported only with VIOS

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: Assignment to the VIOS requires VIOS 2.2.4.2 or later

(#EN0G) - PCIe2 8Gb 2-Port Fibre Channel Adapter

This feature ships a two-port, 8 Gb PCIe Gen2 Fibre Channel
Adapter based on the QLogic QLE2562 Host Bus Adapter (HBA). Each port
provides single initiator capability over a fibre link. The ports
have LC type connectors and utilize shortwave laser optics. The
adapter operates at link speeds of 2, 4, and 8 Gbps and will
automatically negotiates to the highest speed possible. LEDs on each
port provide information on the status and link speed of the port.
The adapter connects to a Fibre Channel switch or can directly attach
to a Fibre Channel port on a supported storage unit. N_Port ID
Virtualization (NPIV) capability is supported through VIOS. Cables
are the responsibility of the customer. Use multimode fibre optic
cables with short-wave lasers that adhere to the following
specifications:

OM3 - multimode 50/125 micron fibre, 2000 MHz*km bandwidth

OM2 - multimode 50/125 micron fibre, 500 MHz*km bandwidth

OM1 - multimode 62.5/125 micron fibre, 200 MHz*km bandwidth

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables. The following table shows the supported
distances for the three different cable types at the three different
link speeds.

#EN0F and #EN0F are
electronically identical with the same CCIN of 578D. #EN0F has a low
profile tailstock bracket. #EN0G has a full high tailstock bracket.

See also feature #5273 or #5735 for a 2-port 8Gb Fibre Channel
adapter based on an Emulex adapter.

See also optional wrap plug feature #ECW0 which is a) required to
run some diagnostic procedures and b) in some cases may speed system
boot when placed in empty ports.

Attributes provided: Dual port Fibre Channel adapter

Attributes required: Full High PCIe slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Red Hat Enterprise Linux 7.2, little endian, or later

Red Hat Enterprise Linux 7.2, big endian, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

SUSE Linux Enterprise Server 12, Service Pack 1, or later

Ubuntu 16.04, or later

AIX Version 7.2 with the 7200-00 Technology Level and Service
Pack 2

AIX Version 7.1 with the 7100-04 Technology Level and Service
Pack 2

AIX version 7.1 with the 7100-03 Technology Level and Service
Pack 7 (planned availability September 8, 2016)

AIX Version 6.1 with the 6100-09 Technology Level and Service
Pack 7

IBM i supported only with VIOS

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: Assignment to the VIOS requires VIOS 2.2.4.2 or later

(#EN0H) - PCIe3 4-port (10Gb FCoE & 1GbE) SR&RJ45

This PCIe Gen3 Converged Network Adapter (CNA) supports both
Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb SR
optical ports and two 1 Gb RJ45 ports in a PCIe 8x short form adapter.
The 10Gb ports have CNA functionality. The 1Gb ports have Ethernet
capability. The adapter was initially announced as PCIe Gen2, but
upgrades to the adapter firmware and introduction of POWER8 servers
have enabled PCIe Gen3.

NPIV capability is provided through VIOS

SR-IOV capability for the NIC function is supported on specific
servers with the appropriate firmware and OS level for any of the four
ports.

The 10Gb ports are SFP+ and include an optical SR transceiver. The
ports have LC Duplex type connectors and utilize shortwave laser
optics and MMF-850nm fiber cabling. With 62.5 micron OM1, up to 33
meter length fiber cables are supported. With 50 micron OM2, up to 82
meter fiber cable lengths are supported. With 50 micron OM3 or OM4,
up to 300 meter fiber cable lengths are supported. Note that an FCoE
switch is required for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0H and #EN0J adapters are electronically identical. They are
physically identical except #EN0H has a tail stock for full high PCIe
slots and #EN0J has a tail stock for low profile slots. The CCIN is
2B93 for both features.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four ports - two 10Gb CNA and two 1Gb E

Attributes required: PCIe Gen2 slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i - 7.1 TR9 and IBM i - 7.2 TR1 supported only with VIOS

IBM i - 7.1 TR10, or later supported with only with VIOS or SRIOV

Note: An HMC is required to setup the adapter for SR-IOV. Only
Ethernet NIC is supported with SRIOV

IBM i - 7.2 TR2 or later supported only with VIOS or SRIOV. An HMC
is required to setup the adapter for SR-IOV. Only Ethernet NIC is
supported with SRIOV

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EN0J) - PCIe3 LP 4-port (10Gb FCoE & 1GbE) SR&RJ45

This PCIe Gen3 Converged Network Adapter (CNA) supports both
Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb SR
optical ports and two 1 Gb RJ45 ports in a PCIe 8x short form adapter.
The 10Gb ports have CNA functionality. The 1Gb ports have Ethernet
capability. The adapter was initially announced as PCIe Gen2, but
upgrades to the adapter firmware and introduction of POWER8 servers
have enabled PCIe Gen3.

NPIV capability is provided through VIOS

The 10Gb ports are SFP+ and include an optical SR transceiver. The
ports have LC Duplex type connectors and utilize shortwave laser
optics and MMF-850nm fiber cabling. With 62.5 micron OM1, up to 33
meter length fiber cables are supported. With 50 micron OM2, up to 82
meter fiber cable lengths are supported. With 50 micron OM3 or OM4,
up to 300 meter fiber cable lengths are supported. Note that an FCoE
switch is required for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0H and #EN0J adapters are electronically identical. They are
physically identical except #EN0H has a tail stock for full high PCIe
slots and #EN0J has a tail stock for low profile slots. The CCIN is
2B93 for both features.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four ports - two 10Gb CNA and two 1Gb E

Attributes required: PCIe Gen2 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 or later and IBM i 7.2 TR1 or later supported only with
VIOS or SRIOV, Note - An HMC is required to setup the adapter for
SR-IOV. Only Ethernet NIC is supported with SRIOV

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

SR-IOV requires:

Requires new firmware level 820.10 SP

HMC

AIX V7.1 with the 7100-03 Technology Level and Service Pack 4 or
later

AIX V6.1 with the 6100-09 Technology Level and Service Pack 4 or
later

IBM i 7.1 Technology Release 9 or later

IBM i 7.2 Technology Release 1 or later

SUSE Linux Enterprise Server 11 Service Pack 3 and later Service
Packs

Red Hat Enterprise Linux 6.5 or later

Red Hat Enterprise Linux 7 or later

SR-IOV Assigned to the VIOS requires VIOS 2.2.3.4 or later

(#EN0K) - PCIe3 4-port (10Gb FCoE & 1GbE) SFP+Copper&RJ45

This PCIe Gen3 Converged Network Adapter (CNA) supports both
Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb ports
for active copper twinax cables and two 1 Gb RJ45 ports in a PCIe 8x
short form adapter. The 10Gb ports have CNA (both NIC and FCoE)
functionality. The 1Gb ports have Ethernet capability. The adapter
was initially announced as PCIe Gen2, but upgrades to the adapter
firmware and introduction of POWER8 servers have enabled PCIe Gen3.

NPIV capability is provided through VIOS

SR-IOV capability for the NIC function is supported on specific
servers with the appropriate firmware and OS level for any of the four
ports.

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables. Note that an FCoE switch is required for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0K and #EN0L adapters are electronically identical. They are
physically identical except #EN0K has a tail stock for full high PCIe
slots and #EN0L has a tail stock for low profile slots. The CCIN is
2CC1 for both features.

#EN0K/EN0L are very similar to the #EN0H/#EN0J adapters except the
#EN0H/EN0K use SR optical cabling and has a different CCIN.

AIX and Linux provide software iSCSI support through the adapter.
Linux can also leverage adapter hardware support including initiator
and header & data digest (CRC) generation and checking

Attributes provided: Four ports - two 10Gb CNA and two 1Gb E

Attributes required: PCIe Gen2 or Gen3 slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i - 7.1 TR9 and IBM i - 7.2 TR1 supported only with VIOS

IBM i - 7.1 TR10, or later supported with only with VIOS or SRIOV

Note: An HMC is required to setup the adapter for SR-IOV. Only
Ethernet NIC is supported with SRIOV

IBM i - 7.2 TR2 or later supported only with VIOS or SRIOV. An HMC
is required to setup the adapter for SR-IOV. Only Ethernet NIC is
supported with SRIOV

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EN0L) - PCIe3 LP 4-port(10Gb FCoE & 1GbE) SFP+Copper&RJ45

This low profile PCIe Gen3 Converged Network Adapter (CNA) supports
both Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb ports
for active copper twinax cables and two 1 Gb RJ45 ports in a PCIe 8x
short form adapter. The 10Gb ports have CNA (both NIC and FCoE)
functionality. The 1Gb ports have Ethernet capability. The adapter
was initially announced as PCIe Gen2, but upgrades to the adapter
firmware and introduction of POWER8 servers have enabled PCIe Gen3.

NPIV capability is provided through VIOS.

SR-IOV capability for the NIC function is supported on specific
servers with the appropriate firmware and OS level for any of the four
ports.

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables. Note that an FCoE switch is required for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0K and #EN0L adapters are electronically identical. They are
physically identical except #EN0K has a tail stock for full high PCIe
slots and #EN0L has a tail stock for low profile slots. The CCIN is
2CC1 for both features.

#EN0K/EN0L are very similar to the #EN0H/#EN0J adapters except the
#EN0H/EN0K use SR optical cabling and has a different CCIN.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four ports - two 10Gb CNA and two 1Gb E

Attributes required: PCIe Gen2 or Gen3 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 or later and IBM i 7.2 TR1 or later supported only with
VIOS or SRIOV, Note - An HMC is required to setup the adapter for
SR-IOV. Only Ethernet NIC is supported with SRIOV

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note:

SR-IOV requires:

Requires new firmware level 820.10 SP

HMC

AIX V7.1 with the 7100-03 Technology Level and Service Pack 4 or
later

AIX V6.1 with the 6100-09 Technology Level and Service Pack 4 or
later

IBM i 7.1 Technology Release 9 or later

IBM i 7.2 Technology Release 1 or later

SUSE Linux Enterprise Server 11 Service Pack 3 and later Service
Packs

Red Hat Enterprise Linux 6.5 or later

Red Hat Enterprise Linux 7 or later

SR-IOV assigned to the VIOS requires VIOS 2.2.3.4 or later

(#EN0M) - PCIe3 4-port(10Gb FCoE & 1GbE) LR&RJ45 Adapter

(No Longer Available as of April 23, 2017)

This PCIe Gen3 Converged Network Adapter (CNA) supports both
Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb LR
optical ports and two 1 Gb RJ45 ports in a PCIe 8x short form adapter.
The 10Gb ports have CNA functionality. The 1Gb ports have Ethernet
capability. NPIV capability is provided through VIOS. The adapter
was initially announced as PCIe Gen2, but upgrades to the adapter
firmware and introduction of POWER8 servers have enabled PCIe Gen3.

SR-IOV capacity for the NIC function is supported on specific
servers with the appropriate firmware and OS level for any of the four
ports.

The 10Gb ports are SFP+ and include an optical LR transceiver. The
ports have LC Duplex type connectors and utilize longtwave laser
optics and 1310nm fiber cabling. With 9 micron OS1, up to 10
kilometer length fiber cables are supported. Priority Flow Control
(PFC) and Fibre Channel over Ethernet (FCoE) are only supported for
distances of 300 meters or less. Note that an FCoE switch is required
for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0M and #EN0N adapters are electronically identical. They are
physically identical except #EN0M has a tail stock for full high PCIe
slots and #EN0N has a tail stock for low profile slots. The CCIN is
2CC0 for both features.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four ports - two 10Gb CNA and two 1Gb E

Attributes required: PCIe Gen2 or Gen3 slot - full high

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

AIX supported

IBM i - 7.1 TR9 and IBM i 7.2 TR1 supported only with VIOS

IBM i - 7.1 TR10, or later suppored only with VIOS or SRIOV. Note
- An HMC is required to setup the adapter for SR-IOV. Only Ethernet
NIC is supported with SRIOV

IBM i - 7.2 TR 2 or later supported only with VIOS or SRIOV Note -
IBM i supported only with VIOS or SR-IOV. An HMC is required to setup
the adapter for SR-IOV. Only Ethernet NIC is supported with SRIOV

Red Hat Enterprise Linux 6.5, or later

Red Hat Enterprise Linux 7, or later

SUSE Linux Enterprise 11 Service Pack 3, or later

SUSE Linux Enterprise 12, or later

Ubuntu 15.04, or later

Red Hat Enterprise Linux 7.1, little endian, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EN0N) - PCIe3 LP 4-port (10Gb FCoE & 1GbE) LR&RJ45 Adapter

(No Longer Available as of April 23, 2017)

This PCIe Gen3 Converged Network Adapter (CNA) supports both
Ethernet NIC (Network Interface Card) traffic and Fibre Channel
traffic over Ethernet (FCoE). The adapter provides two 10 Gb LR
optical ports and two 1 Gb RJ45 ports in a PCIe 8x short form adapter.
The 10Gb ports have CNA functionality. The 1Gb ports have Ethernet
capability. NPIV capability is provided through VIOS. The adapter
was initially announced as PCIe Gen2, but upgrades to the adapter
firmware and introduction of POWER8 servers have enabled PCIe Gen3.

The 10Gb ports are SFP+ and include an optical LR transceiver. The
ports have LC Duplex type connectors and utilize longtwave laser
optics and 1310nm fiber cabling. With 9 micron OS1, up to 10
kilometer length fiber cables are supported. Priority Flow Control
(PFC) and Fibre Channel over Ethernet (FCoE) are only supported for
distances of 300 meters or less. Note that an FCoE switch is required
for any FCoE traffic.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0M and #EN0N adapters are electronically identical. They are
physically identical except #EN0M has a tail stock for full high PCIe
slots and #EN0N has a tail stock for low profile slots. The CCIN is
2CC0 for both features.

(#EN0S) - PCIe2 4-Port (10Gb+1GbE) SR+RJ45 Adapter

The 10Gb ports are SFP+ and include optical SR transceivers. The
ports have LC Duplex type connectors and utilize shortwave laser
optics and MMF-850nm fiber cabling. With 62.5 micron OM1, up to 33
meter length fiber cables are supported. With 50 micron OM2, up to 82
meter fiber cable lengths are supported. With 50 micron OM3 or OM4,
up to 300 meter fiber cable lengths are supported.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0S and #EN0T adapters are electronically identical. They are
physically identical except #EN0S has a tail stock for full high PCIe
slots and #EN0T has a tail stock for low profile slots. The CCIN is
2CC3 for both features.

(#EN0T) - PCIe2 LP 4-Port (10Gb+1GbE) SR+RJ45 Adapter

The 10Gb ports are SFP+ and include optical SR transceivers. The
ports have LC Duplex type connectors and utilize shortwave laser
optics and MMF-850nm fiber cabling. With 62.5 micron OM1, up to 33
meter length fiber cables are supported. With 50 micron OM2, up to 82
meter fiber cable lengths are supported. With 50 micron OM3 or OM4,
up to 300 meter fiber cable lengths are supported.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0S and #EN0T adapters are electronically identical. They are
physically identical except #EN0S has a tail stock for full high PCIe
slots and #EN0T has a tail stock for low profile slots. The CCIN is
2CC3 for both features.

(#EN0U) - PCIe2 4-port (10Gb+1GbE) Copper SFP+RJ45 Adapter

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0U and #EN0V adapters are electronically identical. They are
physically identical except #EN0U has a tail stock for full high PCIe
slots and #EN0V has a tail stock for low profile slots. The CCIN is
2CC3 for both features.

(#EN0V) - PCIe2 LP 4-port (10Gb+1GbE) Copper SFP+RJ45 Adapter

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables.

For the 1Gb RJ45 ports, 4-pair CAT-5 Unshielded Twisted Pair (UTP)
cable or higher is supported for distances of up to 100 meters. In
addition to 1Gb (1000 Mb) networks, 100 Mb networks are also
supported, but 10Mb networks are not supported.

#EN0U and #EN0V adapters are electronically identical. They are
physically identical except #EN0U has a tail stock for full high PCIe
slots and #EN0V has a tail stock for low profile slots. The CCIN is
2CC3 for both features.

AIX, IBM i, and Linux provide software iSCSI support through the
the adapter.

Attributes provided: Two 10G-BaseT ports

Attributes required: PCIe Gen2 or Gen3 slot - low profile

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 or later and IBM i 7.2 TR1 or later supported only
with VIOS

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EN0Y) - PCIe2 LP 8Gb 4-port Fibre Channel Adapter

PCIe Gen2 8 Gigabit quad port Fibre Channel Adapter is a low
profile, high-performance 8x short form Host Bus Adapter (HBA). Each
port provides single initiator capability over a fiber link or with
NPIV, multiple initiator capability is provided. The ports have LC
type connectors and utilize shortwave laser optics. The adapter
operates at link speeds of 2, 4, and 8 Gbps and will automatically
negotiate to the highest speed possible. LEDs on each port provide
information on the status and link speed of the port. CCIN is EN0Y.
This adapter is based on the QLogic QLE2564 PCIe Host Bus Adapter (HBA).

The adapter connects to a Fibre Channel switch. Direct device
attachment has not been tested and is not supported.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications:

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables.

The following table shows the supported distances for the three
different cable types at the three different link speeds.

A Gen2 or later PCIe slot is required to provide the bandwidth for
all four ports to operate at full speed. Use in Gen1 slots has not
been tested and is not supported.

The EN12 and EN0Y adapters are electronically and functionally
identical. Feature EN12 indicates a full high tailstock bracket.
Feature EN0Y indicates a low profile tailstock bracket.

Consult with your IBM representative or Business Partner for
additional information relative to any third party attachment.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: 4-port 8Gb Fibre Channel Adapter

Attributes required: 1 Empty PCIe Gen2 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR9 or later and IBM i 7.2 TR1 or later supported only
with VIOS

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS supported

(#EN12) - PCIe2 8Gb 4-port Fibre Channel Adapter

PCIe Gen2 8 Gigabit quad port Fibre Channel Adapter is a high-
performance 8x short form Host Bus Adapter (HBA). Each port provides
single initiator capability over a fiber link or with NPIV, multiple
initiator capability is provided. The ports have LC type connectors
and use shortwave laser optics. The adapter operates at link speeds
of 2, 4, and 8 Gbps and will automatically negotiate to the highest
speed possible. LEDs on each port provide information on the status
and link speed of the port.
This adapter is based on the QLogic QLE2564 PCIe Host Bus Adapter (HBA).

The adapter connects to a Fibre Channel switch. Direct device
attachment has not been tested and is not supported.

N_Port ID Virtualization (NPIV) capability is supported through
VIOS.

Cables are the responsibility of the customer. Use multimode fibre
optic cables with short-wave lasers that adhere to the following
specifications: OM3 - multimode 50/125 micron fibre, 2000 MHz*km
bandwidth

Because core sizes are different, OM1 cables can only be connected
to other OM1 cables. For best results, OM2 cables should not be
connected to OM3 cables. However, if an OM2 cable is connected to an
OM3 cable, the characteristics of the OM2 cable apply to the entire
length of the cables.

The following table shows the supported distances for the three
different cable types at the three different link speeds. Cable

The EN12 and EN0Y adapters are electronically and functionally
identical with the same CCIN of EN0Y. # EN12 indicates a full high
tailstock bracket. #EN0Y indicates a low profile tailstock bracket.

Consult with your IBM representative or Business Partner for
additional information relative to any third party attachment.

See also optional wrap plug feature #ECW0 which is:

a) Required to run some diagnostic procedures and

b) in some cases may speed system boot when placed in empty ports
as well as avoid useless messages pointing to a planned empty port.

Attributes provided: 4-port 8Gb Fibre Channel Adapter

Attributes required: Available PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

Red Hat Enterprise Linux Server 6.7, or later

Red Hat Enterprise Linux Server 7.1, big endian, for Power, or
later

Red Hat Enterprise Linux Server 7.1, little endian, for Power or
later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu Server 16.04, or later

Ubuntu Server 14.04.3, or later

AIX 6.1 and 7.1 or later supported

Refer to Software Requirements for specific O/S levels supported

IBM i supported only with VIOS

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN13) - PCIe 1-port Bisync Adapter

(No Longer Available as of May 13, 2016)

The #EN13 is a full high PCIe adapter which provides 1 RVX port for
Bisync capability. This RVX port is labeled "Port 1". Its CCIN =
576C Bisync functionality identical to #2893 PCIe adapter is provided.

Select one of the following cables to attach to port 1 (RVX port):

#0348 - V.24/EIA232 20-Ft PCI Cable

#0353 - V.35 20-Ft PCI Cable

#0359 - X.21 20-Ft PCI Cable

Limitations

The #2893 does not support the remote ring indicate function.

There is a second port on the card with an integrated modem, but
Bisync is not supported on this port (labeled "Port 0").

Attributes required: PCIe full high slot, external modem, cable
for the modem.

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN14) - PCIe 1-port Bisync Adapter CIM

(No Longer Available as of May 13, 2016)

The #EN14 is a full high PCIe adapter which provides 1 RVX port for
Bisync capability. This RVX port is labeled "Port 1". Its CCIN =
576C Bisync functionality identical to #2894 PCIe adapter is provided.

Attributes required: PCIe full high slot, external modem, cable
for the modem.

Minimum required: 0

Maximum allowed: 14 (Initial order maximum: 14)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EN15) - PCIe3 4-port 10GbE SR Adapter

This PCIe Gen3 supports Ethernet NIC (Network Interface Card)
traffic and also supports SR-IOV capability. The adapter provides
four 10 Gb SR optical ports in a PCIe 8x short form adapter. SR-IOV
capability for the NIC function is supported with the appropriate
firmware and OS level for any of the four ports. Enabling SR-IOV
function requires an HMC.

The four 10Gb ports are SFP+ and include four optical SR
transceivers. The ports have LC Duplex type connectors and utilize
shortwave laser optics and MMF-850nm fiber cabling. With 62.5 micron
OM1, up to 33 meter length fiber cables are supported. With 50 micron
OM2, up to 82 meter fiber cable lengths are supported. With 50 micron
OM3 or OM4, up to 300 meter fiber cable lengths are supported.

#EN15 and #EN16 adapters are electronically identical.
They are physically identical except #EN15 has a tail stock for full high
PCIe slots and #EN16 has a tail stock allowing it to fit in a Power
E870/E880 or Power E870C/E880C system node PCIe slot.
The CCIN is 2CE3 for both features.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four 10GbE ports

Attributes required: full high PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

AIX Version 6.1 with the 6100-09 Technology Level and service pack 5
and APAR IV68443 or later

AIX Version 7.1 with the 7100-03 Technology Level and Service Pack 5
and APAR IV68444 or later

AIX Version 7.1 with the 7100-02 Technology Level and Service Pack 7
or later

AIX Version 6.1 with the 6100-08 Technology Level and service Pack 7
or later

IBM i 7.1 TR10 or later

IBM i 7.2 TR2 or later

Red Hat Enterprise Linux 6.6, or later

Red Hat Enterprise Linux 7.1, big endian, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 12, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

Ubuntu Server 16.04, or later

Ubuntu Server 14.04.03, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS 2.2.3.3 or later

(#EN16) - PCIe3 LPX 4-port 10GbE SR Adapter

This PCIe Gen3 supports Ethernet NIC (Network Interface Card)
traffic and also supports SR-IOV capability. The adapter provides
four 10 Gb SR optical ports in a PCIe 8x short form adapter. SR-IOV
capability for the NIC function is supported with the appropriate
firmware and OS level for any of the four ports. Enabling SR-IOV
function requires an HMC.

The four 10Gb ports are SFP+ and include four optical SR
transceivers. The ports have LC Duplex type connectors and utilize
shortwave laser optics and MMF-850nm fiber cabling. With 62.5 micron
OM1, up to 33 meter length fiber cables are supported. With 50 micron
OM2, up to 82 meter fiber cable lengths are supported. With 50 micron
OM3 or OM4, up to 300 meter fiber cable lengths are supported.

#EN15 and #EN16 adapters are electronically identical.
They are physically identical except #EN15 has a tail stock for full high
PCIe slots and #EN16 has a tail stock allowing it to fit in a Power
E870/E880 or Power E870C/E880C system node PCIe slot.
The CCIN is 2CE3 for both features.

Limitation: The adapter is too tall to fit in 2U Power S812L,
S822L, S822 system unit low profile PCIe slots.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four 10GbE ports

Attributes required: system node PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX Version 6.1 with the 6100-09 Technology Level and service pack 5
and APAR IV68443 or later

AIX Version 7.1 with the 7100-03 Technology Level and Service Pack 5
and APAR IV68444 or later

AIX Version 7.1 with the 7100-02 Technology Level and Service Pack 7
or later

AIX Version 6.1 with the 6100-08 Technology Level and service Pack 7
or later

IBM i 7.1 TR10 or later

IBM i 7.2 TR2 or later

Red Hat Enterprise Linux 6.6, or later

Red Hat Enterprise Linux 7.1, big endian, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 12, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

Ubuntu Server 16.04, or later

Ubuntu Server 14.04.03, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS 2.2.3.3 or later

(#EN17) - PCIe3 4-port 10GbE SFP+ Copper Adapter

This PCIe Gen3 supports Ethernet NIC (Network Interface Card)
traffic and also supports SR-IOV capability. The adapter provides
four 10GbE SFP+ ports into which copper twinax transceivers will be
placed. It is a PCIe 8x short form adapter. SR-IOV capability for
the NIC function is supported with the appropriate firmware and OS
level for any of the four ports. Enabling SR-IOV function requires an
HMC.

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables.

#EN17 and #EN18 adapters are electronically identical.
They are physically identical except #EN17 has a tail stock for full high
PCIe slots and #EN18 has a tail stock allowing it to fit in a Power
E870/E880 or Power E870C/E880C system node PCIe slot.
The CCIN is 2CE4 for both features.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four 10GbE ports

Attributes required: full high PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 192 (Initial order maximum: 192)

OS level required:

AIX Version 6.1 with the 6100-09 Technology Level and service pack
5 and APAR IV68443 or later

AIX Version 7.1 with the 7100-03 Technology Level and Service Pack
5 and APAR IV68444 or later

AIX Version 7.1 with the 7100-02 Technology Level and Service Pack
7 or later (planned availability 9/30/2015)

AIX Version 6.1 with the 6100-08 Technology Level and service Pack
7 or later (planned availability 9/30/2015)

IBM i 7.1 TR10 or later

IBM i 7.2 TR2 or later

Red Hat Enterprise Linux 6.6, or later

Red Hat Enterprise Linux 7.1, big endian, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS 2.2.3.3 or later

(#EN18) - PCIe3 LPX 4-port 10GbE SFP+ Copper Adapter

This PCIe Gen3 supports Ethernet NIC (Network Interface Card)
traffic and also supports SR-IOV capability. The adapter provides
four 10GbE SFP+ ports into which copper twinax transceivers will be
placed. It is a PCIe 8x short form adapter. SR-IOV capability for
the NIC function is supported with the appropriate firmware and OS
level for any of the four ports. Enabling SR-IOV function requires an
HMC.

The 10Gb ports are SFP+ and do not include a transceiver. Active
Copper twinax cables up to 5 meter in length are supported such as
provided by feature #EN01, #EN02 or #EN03. A transceiver is included
with these cables. Note that SFP+ twinax copper is NOT AS/400 5250
twinax or CX4 or 10 GBase-T. Active cables differ from passive
cables.

#EN17 and #EN18 adapters are electronically identical.
They are physically identical except #EN17 has a tail stock for full high
PCIe slots and #EN18 has a tail stock allowing it to fit in a Power
E870/E880 or Power E870C/E880C system node PCIe slot. The CCIN is 2CE4
for both features.
The CCIN is 2CE4 for both features.

Limitation: The adapter is too tall to fit in 2U Power S812L,
S822L, S822 system unit low profile PCIe slots.

AIX, IBM i and Linux provide software iSCSI support through the
adapter. Linux can also leverage adapter hardware support including
initiator and header & data digest (CRC) generation and checking

Attributes provided: Four 10GbE ports

Attributes required: system node PCIe Gen3 slot

Minimum required: 0

Maximum allowed: 32 (Initial order maximum: 32)

OS level required:

AIX Version 6.1 with the 6100-09 Technology Level and service pack
5 and APAR IV68443 or later

AIX Version 7.1 with the 7100-03 Technology Level and Service Pack
5 and APAR IV68444 or later

AIX Version 7.1 with the 7100-02 Technology Level and Service Pack
7 or later (planned availability 9/30/2015)

AIX Version 6.1 with the 6100-08 Technology Level and service Pack
7 or later (planned availability 9/30/2015)

IBM i 7.1 TR10 or later

IBM i 7.2 TR2 or later

Red Hat Enterprise Linux 6.6, or later

Red Hat Enterprise Linux 7.1, big endian, or later

Red Hat Enterprise Linux 7.1, little endian, or later

SUSE Linux Enterprise Server 12, or later

Ubuntu Server 14.04.3, or later

Ubuntu Server 16.04, or later

SUSE Linux Enterprise Server 11, Service Pack 4, or later

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

Note: VIOS 2.2.3.51 or later. VIOS supports NIC capability only.

(#EN27) - 2 Port Async EIA-232 PCIe Adapter

(No Longer Available as of March 11, 2016)

Adapter provides connection for 2 asynchronous EIA-232 devices.
Ports are programmable to support EIA-232 protocols, at a line speed
of 128K bps. Two RJ45 connections are located on the rear of the
adapter. To attach to devices using a 9-pin (DB9) connection, use an
RJ45-to-DB9 converter. For convenience, one converter is included
with this feature. One converter for each connector needing a DB9
connector is needed.

The converter cable shipped with the feature is a 10-pin RJ-45
approximately 30 cm or 1 foot in length. 10-pin is often used for DCE
devices like modems. DTE devices may be able to use fewer pin
connections. Multiple RJ-45 pin out options are used and sold across
the industry. The #3930 feature provides an 8-pin RJ45-to-DB9
converter cable. If an addition 10-pin RJ45-to-DB9 connector like the
cable provided with the adapter is desired, it is commonly available
from multiple sources and identified as a serial cable with a Digi
part number 76000239. It is also possible to look up the cable's pin
out specifications and build your own cables.

Note #EN27 and # EN28 are physically and electrically identical
adapters, except for the type of PCIe slot used (full-high or low
profile).

(#EN28) - PCIe LP 2-Port Async EIA-232 Adapter

(No Longer Available as of February 9, 2016)

Adapter provides connection for 2 asynchronous EIA-232 devices.
Ports are programmable to support EIA-232 protocols, at a line speed
of 128K bps. Two RJ45 connections are located on the rear of the
adapter. To attach to devices using a 9-pin (DB9) connection, use an
RJ45-to-DB9 converter. For convenience, one converter is included
with this feature. One converter for each connector needing a DB9
connector is needed.

The converter cable shipped with the feature is a 10-pin RJ-45
approximately 30 cm or 1 foot in length. 10-pin is often used for DCE
devices like modems. DTE devices may be able to use fewer pin
connections. Multiple RJ-45 pin out options are used and sold across
the industry. The #3930 feature provides an 8-pin RJ45-to-DB9
converter cable. If an addition 10-pin RJ45-to-DB9 connector like the
cable provided with the adapter is desired, it is commonly available
from multiple sources and identified as a serial cable with a Digi
part number 76000239. It is also possible to look up the cable's pin
out specifications and build your own cables.

Note #EN27 and #EN28 are physically and electrically identical
adapters, except for the type of PICe Slot used (full-high or low
profile).

(#EN29) - 2 Port Async EIA-232 PCIe Adapter

Adapter provides connection for 2 asynchronous EIA-232 devices.
Ports are programmable to support EIA-232 protocols, at a line speed
of 128K bps. Two RJ45 connections are located on the rear of the
adapter. To attach to devices using a 9-pin (DB9) connection, use an
RJ45-to-DB9 converter. For convenience, one converter is included
with this feature. One converter for each connector needing a DB9
connector is needed.

The converter cable shipped with the feature is a 10-pin RJ-45
approximately 30 cm or 1 foot in length. 10-pin is often used for DCE
devices like modems. DTE devices may be able to use fewer pin
connections. Multiple RJ-45 pin out options are used and sold across
the industry. The #3930 feature provides an 8-pin RJ45-to-DB9
converter cable. If an addition 10-pin RJ45-to-DB9 connector like the
cable provided with the adapter is desired, it is commonly available
from multiple sources and identified as a serial cable with a Digi
part number 76000239. It is also possible to look up the cable's pin
out specifications and build your own cables.

This adapter is functionally and electronically equivalent to
#EN27 and is functionally equivalent to the earlier #5289 Async
adapters.

(#EP9T) - 90 Days Elastic CoD Processor Core Enablement

90 Days Elastic CoD Processor Core Enablement (#EP9T) feature can
be ordered to enable your server for Capacity on Demand. You must
sign an Elastic Capacity on Demand contract and the Sales Channel has
to register the server before an MES order can be fulfilled for #EP9T.
Feature code #EP9T provides access to Elastic CoD processor resources
for 90 days. Access to these resources is measured in processor-days.
For example, if there are 32-cores of inactive, CoD processors
installed, this feature will enable using 2880 processor-days before
disabling CoD access to the inactive CoD processors.

After usage of this feature has been exhausted, additional
temporary usage of Elastic CoD processor resources may be activated by
ordering another Elastic CoD enablement code #EP9T. Before ordering a
fourth enablement code, after the system maximum of 4 has been
reached, the current code must be RPO deleted from the configuration
file with a MES RPO delete order.

Usage of Elastic CoD processors is billed by ordering either
Elastic Proc. CoD Billing for 1 Proc-day. There is a different
Elastic Billing feature code available for each processor feature
installed in a 8412, 9117, 9119, 9179, 9080, and 8408 MT. The maximum
Elastic proc-days billable on any one order is 4

Note: Customer is limited to select a quantity between 0 and 4 for this
feature on a single MES order.

(#EPAA) - HVDC PDU - 90A 6xOutlet

Feature ships a high voltage dc (HVDC) Power Distribution Unit
(PDU) with six customer-usable Rong Feng RF-203P outlets. Each outlet
is rated at up to 10 amp and has its own circuit breaker. The PDU is
most typically used for 380 volt configurations, but is rated for
240-380 volt. The PDU is rated for 90 amps. The PDU comes with a
permanently attached 4.3 m (14 ft) line cord which is un-terminated
(no plug) which must be directly wired into a HVDC power source.

The PDU can be either vertically mounted in rack side pockets or
horizontally mounted. Using the feature #EBA5 mounting specify
communicates to IBM manufacturing it should be horizontally mounted.
If horizontally mounted, the PDU uses 1U rack space and makes
accessing power cords much easier.

Use HVDC power cords ordered with the server or I/O drawer to plug
into the PDU RF-203P outlets. See HVDC power cord features #EPAD or
#EPAC.

If ordering this PDU with a 7014-T00 or 7014-T42 rack which is
being shipped from IBM at the same time, see PDU feature #EPAF (HVDC
PDU - 90A 6xOutlet Alternate Base).

Limitation:

Not supported in North America

IBM Manufacturing does not factory integrate AC and DC PDUs in the
same rack

Customers mixing AC and DC in the same rack must carefully follow
power wiring guidelines

Attributes provided: HVDC PDU

Attributes required: HVDC power supplies and power cords on the
server or I/O drawer

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 0)

OS level required: None

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EPAC) - Auto Selected HVDC Power Cord

This feature ships a high voltage dc (HVDC) power cord which is
1.0M, 1.5M or 2.5M in length. The specific length will be selected by
IBM Manufacturing when the system is being manufactured. Cord is
connected to an HVDC PDU and to a HVDC power supply in a server or I/
O drawer. Cord has two Rong Feng RF-203P plugs, one at each end.
Cord is 3 connector 1.3 mm2 gauge or 16 AWG and is rated for 10 amp
and 192-400 volt dc Order one feature #EPAC for each HVDC power supply
in the server or I/O drawer.

See feature #EPAD to specify a 2.5 meter (8 foot) HVDC cable and
not have IBM Manufacturing select a length. There are no feature
codes to specify a 1.0M or 1.5M cord.

Attributes provided: HVDC power cord

Attributes required: HVDC power supply

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: No

(#EPAD) - 2.5 Meter HVDC Power Cord

2.5 meter (8 foot) high voltage dc (HVDC) power cord. Cord is
connected to an HVDC PDU and to a HVDC power supply in a server or I/
O drawer. Cord has two Rong Feng RF-203P plugs, one at each end.
Cord is 3 conductor,1.3 mm2 gauge or 16 AWG and is rated for 10 amp
and 192-400 volt dc For other length HVDC power cords, see feature
#EPAC.

Attributes provided: HVDC power cord

Attributes required: HVDC power supply

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EPBB) - 4.35 GHz, 32-core POWER8 processor

Thirty two core 4.356 GHz POWER8 processor planar containing four
eight-core processor SCMs. Each processor core has 512 KB of L2 cache
and 8 MB of L3 cache. There are 8 CDIMM slots per processor and a
total of 32 CDIMM slots per system node.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

(#EPBD) - 4.02 GHz 48-core POWER8 processor

Forty eight core 4.024 GHz POWER8 processor planar containing
four twelve-core processor SCMs. Each processor core has 512 KB of L2
cache and 8 MB of L3 cache. There are 8 CDIMM slots per processor and
a total of 32 CDIMM slots per system node.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

This feature provides thirty two core 4.356 GHz POWER8 processor
planar containing four eight-core processor SCMs. Each processor core
has 512 KB of L2 cache and 8 MB of L3 cache. There are 8 CDIMM slots
per processor and a total of 32 CDIMM slots per system node for a
Capacity Backup (CBU) system in deployment for the high availability
or disaster recovery operation.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

Attributes provided: Four 8-core 4.356 GHz SCMs providing 32-core
system node for a CBU Power Enterprise System

This feature provides fourthy eight core 4.024 GHz POWER8
processor planar containing four twelve-core processor SCMs. Each
processor core has 512 KB of L2 cache and 8 MB of L3 cache. There are
8 CDIMM slots per processor and a total of 32 CDIMM slots per system
node for a Capacity Backup (CBU) system in deployment for the high
availability or disaster recovery operation.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

Attributes provided: Four 12-core 4.024 GHz SCMs providing
48-core system node for a CBU Power Enterprise System

Attributes required: EB3J

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required:

Linux supported AIX 6.1 and 7.1 or later supported

IBM i 7.1 TR10 and 7.2 TR2 or later supported

Refer to Software Requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#EPBK) - 1 core Processor Activation for #EPBB or #EPBG

Each occurrence of this feature will permanently activate one
processor core on Processor Card #EPBB or #EPBG.

Attributes provided: One processor core activation for #EPBB or
#EPBG

Attributes required: #EPBB or #EPBG with inactive processor cores

Minimum required: 0

Maximum allowed: 128 (Initial order maximum: 128)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EPBM) - 1 core Processor Activation for #EPBD or #EPBH

Each occurrence of this feature will permanently activate one
processor core on Processor Card #EPBD or #EPBH.

Static processor core activation which can be converted to a mobile
activation at lower charge with a future MES conversion order.

Use this feature with an initial server order when mobile
activations are not available. Once a server is established as part
of a Power Enterprise Pool, additional mobile activations can be
ordered directly without using this "two-step" order.

Static processor core activation which can be converted to a mobile
activation at lower charge with a future MES conversion order.

Use this feature with an initial server order when mobile
activations are not available. Once a server is established as part
of a Power Enterprise Pool, additional mobile activations can be
ordered directly without using this "two-step" order.

Attributes provided: Processor core activation

Attributes required: #EPBD or #EPBH processor core

Minimum required: 0

Maximum allowed: 184 (Initial order maximum: 184)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EPBS) - 4.19GHz, 40-core POWER8 processor

Forty core 4.19 GHz POWER8 processor planar containing four
ten-core processor SCMs. Each processor core has 512 KB of L2 cache and
8 MB of L3 cache. There are 8 CDIMM slots per processor and a total of
32 CDIMM slots per system node.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

This feature provides forty core 4.19 GHz POWER8 processor planar
containing four ten-core processor SCMs. Each processor core has 512
KB of L2 cache and 8 MB of L3 cache. There are 8 CDIMM slots per
processor and a total of 32 CDIMM slots per system node for a
Capacity Backup (CBU) system in deployment for the high availability
or disaster recovery operation.

Each processor feature will deliver a set of four identical
processors that populate four sockets in a given system node. All
processors in the system node must be identical. Activations are
applied per core with a minimum of eight activation.

Attributes provided: Four 10-core 4.19 GHz SCMs providing 40-core
system node for a CBU Power Enterprise System

Attributes required: EB3J

Minimum required: 0

Maximum allowed: 4 (Initial order maximum: 4)

OS level required:

Red Hat Enterprise Linux 6.5 or later

Red Hat Enterprise Linux 7 or later

Red Hat Enterprise Linux 7.1 (le) or later

SUSE Linux Enterprise Server 11 Service Pack 3 or later

SUSE Linux Enterprise Server 12 or later

Ubuntu Server 14.04.3 or later

AIX 6.1, 7.1, and 7.2 or later

IBM i 7.1 TR11 and 7.2 TR3 or later

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

(#EPBU) - 1 core Processor Activation for #EPBS or #EPBT

Each occurrence of this feature will permanently activate one
processor core on Processor Card #EPBS or #EPBT.

Static processor core activation which can be converted to a mobile
activation at lower charge with a future MES conversion order.

Use this feature with an initial server order when mobile
activations are not available. Once a server is established as part
of a Power Enterprise Pool, additional mobile activations can be
ordered directly without using this "two-step" order.

Attributes provided: Processor core activation

Attributes required: #EPBS or #EPBT processor core

Minimum required: 0

Maximum allowed: 152 (Initial order maximum: 152)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

(#EPE0) - Static to Mobile Processor Auto Conversion

Feature is used to order automatic conversion of a static action to
a mobile activation without IBM intervention. The term "static
activations" also includes Mobile Enabled activations which are
technically static. Firmware 840 or later is a pre-requisite. Use of
this feature on servers with firmware 830 or earlier will require IBM
intervention through the pcod@us.ibm.com

Attributes provided: Automated conversion

Attributes required: Firmware 840

Minimum required: 0

Maximum allowed: 184 (Initial order maximum: 0)

OS level required: N/A

Initial Order/MES/Both/Supported: MES

CSU: Yes

Return parts MES: No

(#EPJ3) - 48 Proc-Days of Elastic CoD Processor Resources

(No Longer Available as of August 10, 2016)

Feature code #EPJ3 provides a billing credit for 48 Proc-Days of
Elastic CoD processor resources. #EPJ3 features must be ordered with a
new server or with an upgrade to a new server and the quantity of #EJP3
depends on the quantity of system nodes and the number of processor cores
per node. The credit is tracked by IBM and is used to offset billing
days ordered by the client for the specific server for which it was
ordered. This billing credit is not transferable to a different server
or different model. #EPJ3 is not orderable by MES orders except with a
model change upgrade.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EP9T), usage reporting, billing feature
ordering (1 proc-day or 100 proc-days), and billing processes are used
on the server and are a pre-requisite to using the credit balance

For example, if ten #EPJ3 features had been ordered for a Power E880
or E880C #EPBB processor feature, then credit for 480 days would be known
by IBM.
And if for example there were 24 inactive, CoD processor cores
temporarily activated using Elastic CoD, then those cores could be used
for 20 days and their billing features would be offset by the #EPJ3
generated credit balance.

Attributes required: New Power E880 or E880C server or model upgrade
to E880 or E880C.

Minimum required: 0

Maximum allowed: 40 (Initial order maximum: 40)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: N/A

Return parts MES: Does not apply

Note: Ordered as quantity of 10

(#EPJ5) - 72 Proc-Days of Elastic CoD Temporary Processor Resources

(No Longer Available as of August 10, 2016)

Feature code #EPJ5 provides a billing credit for 72 Proc-Days of
Elastic
CoD processor resources. #EPJ5 features must be ordered with a new server
or with an upgrade to a new server and the quantity of #EJP5 depends on
the quantity of system nodes and the number of processor cores per node.
The credit is tracked by IBM and is used to offset billing days ordered
by the client for the specific server for which it was ordered. This
billing credit is not transferable to a different server or different
model. #EPJ5 is not orderable by MES orders except with a model change
upgrade.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EP9T), usage reporting, billing feature
ordering (1 proc-day or 100 proc-days), and billing processes are used on
the server and are a pre-requisite to using the credit balance.

For example, if ten #EPJ5 features had been ordered for a Power E880
or E880C processor feature EPBD, then credit for 720 days would be known
by IBM. And if for example there were 36 inactive, CoD processor cores
temporarily activated using Elastic CoD, then those cores could be used
for 20 days and their billing features would be offset by the #EPJ5
generated credit balance.

Attributes required: New Power E880 or E880C server or model upgrade
to E880 or E880C.

Minimum required: 0

Maximum allowed: 40 (Initial order maximum: 40)

OS level required: N/A

Initial Order/MES/Both/Supported: Initial

CSU: N/A

Return parts MES: Does not apply

Note: Ordered as quantity of 10

(#EPJC) - 1 Proc-Day Elastic CoD Billing for #EPBB, AIX/Linux

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPJC
should be ordered for each billable processor core day of feature
#EPBB used by a supported AIX or Linux operating system.

Attributes provided: One processor core day billing for feature
#EPBB (AIX or Linux)

Attributes required: At least one processor core on #EPBB that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPJD) - 1 Proc-Day Elastic CoD Billing for #EPBB, IBM i

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPJD
should be ordered for each billable processor core day of feature
#EPBB used by a supported IBM i operating system.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPJE provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBB for AIX/Linux.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPJF provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBB for IBM i.

Attributes required: At least one processor core on #EPBB that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 99 (Initial order maximum: 99)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPJG) Proc CoD Utility Billing, 100 Proc-mins. for #EPBB, AIX/Linux

Provides payment for temporary use of processor feature #EPBB with
supported AIX or Linux operating systems. Each occurrence of this
feature will pay for 100 minutes of usage. The purchase of this
feature occurs after the customer has 100 minutes of use on processor
cores in the shared processor pool that are not permanently active.

Attributes provided: Payment for Utility CoD use of processor
#EPBB (AIX or Linux)

Attributes required: At least one processor core on #EPBB that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPJH) - Proc CoD Utility Billing, 100 Proc-mins. for #EPBB, IBM i

Provides payment for temporary use of processor feature #EPBB with
supported IBM i operating systems. Each occurrence of this feature
will pay for 100 minutes of usage. The purchase of this feature
occurs after the customer has 100 minutes of use on processor cores in
the shared processor pool that are not permanently active.

Attributes required: At least one processor core on #EPBB that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPJQ) - 1 Proc-Day Elastic CoD Billing for #EPBD, AIX/Linux

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPJQ
should be ordered for each billable processor core day of feature
#EPBD used by a supported AIX or Linux operating system.

See also billing feature code #EPJS for quantity 100 processor days

Attributes provided: One processor core day billing for feature
#EPBD (AIX or Linux)

Attributes required: At least one processor core on #EPBD that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPJR) - 1 Proc-Day Elastic CoD Billing for #EPBD, IBM i

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPJR
should be ordered for each billable processor core day of feature
#EPBD used by a supported IBM i operating system.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPJS provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBD for AIX/ Linux.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPJT provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBD for IBM i.

Attributes required: At least one processor core on #EPBD that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 99 (Initial order maximum: 99)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPJU) Proc CoD Utility Billing, 100 Proc-mins. for #EPBD, AIX/Linux

Provides billing feature which is ordered to pay for temporary use
of processor feature #EPBD with supported AIX or Linux operating
systems. Each occurrence of this feature will pay for 100 minutes of
usage. The purchase of this feature occurs after the customer has 100
minutes of use on processor cores in the shared processor pool that
are not permanently active.

Attributes provided: Payment for Utility CoD use of processor
#EPBD (AIX or Linux)

Attributes required: At least one processor core on #EPBD that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPJV) Proc CoD Utility Billing, 100 Proc-mins. for #EPBD, IBM i

Provides payment for temporary use of processor feature #EPBD with
supported IBM i operating systems. Each occurrence of this feature
will pay for 100 minutes of usage. The purchase of this feature
occurs after the customer has 100 minutes of use on processor cores in
the shared processor pool that are not permanently active.

Attributes required: At least one processor core on #EPBD that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPKK) - 48 Proc-Days of Elastic CoD Processor Resources

(No Longer Available as of August 10, 2016)

Feature code #EPKK provides a billing credit for 48 Proc-Days of
Elastic CoD processor resources. #EPKK features must be ordered with
a new server or with an upgrade to a new server and the quantity of
#EPKK depends on the quantity of system nodes and the number of
processor cores per node. The credit is tracked by IBM and is used to
offset billing days ordered by the client for the specific server for
which it was ordered. This billing credit is not transferable to a
different server or different model. #EPKK is not orderable by MES
orders except with a model change upgrade.

Normal Elastic Capacity on Demand contracts, sales channel
registration, enablement codes (#EP9T), usage reporting, billing
feature ordering (1 proc-day or 100 proc-days), and billing processes
are used on the server and are a pre-requisite to using the credit
balance.

For example, if ten #EPKK features had been ordered for a Power E880
or E880C #EPBS processor feature, then credit for 480 days would be known
by IBM.
And if for example there were 24 inactive, CoD processor
cores temporarily activated using Elastic CoD, then those cores could
be used for 20 days and their billing features would be offset by the
#EPKK generated credit balance.

Attributes required: New Power E880 server or model upgrade to E880
or E880C.

Minimum required: 0

Maximum allowed: 40 (Initial order maximum: 40)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Ordered as quantity of 10

(#EPKL) - 1 Proc-Day Elastic CoD Billing for #EPBS, AIX/Linux

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPKL
should be ordered for each billable processor core day of feature
#EPBB used by a supported AIX or Linux operating system.

Attributes provided: One processor core day billing for feature
#EPBS (AIX or Linux)

Attributes required: At least one processor core on #EPBS that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPKM) - 1 Proc-Day Elastic CoD Billing for #EPBS, IBM i

Once an Elastic Processor Enablement feature is ordered and the
associated enablement code is entered into the system, you must report
your Elastic CoD usage to IBM at least monthly. This information,
used to compute your billing data, is then provided to your sales
channel. The sales channel will place an order for a quantity of
Elastic Processor Core Day Billing features and bill you. One #EPJD
should be ordered for each billable processor core day of feature
#EPBB used by a supported IBM i operating system.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPKN provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBS for AIX/ Linux.

After the Elastic Processor function is enabled in a system, you
must report your Elastic CoD usage to IBM at least monthly. This
information, used to compute your billing data, is provided to your
sales channel. The sales channel will place an order on your behalf
for the quantity of this feature that matches your reported use. One
#EPKP provides 100 days of Elastic CoD processor billing for POWER8
CoD Processor Book #EPBS for IBM i.

Attributes required: At least one processor core on #EPBS that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 99 (Initial order maximum: 99)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPKQ) Proc CoD Utility Billing, 100 Proc-mins. for #EPBS, AIX/Linux

Provides payment for temporary use of processor feature #EPBS with
supported AIX or Linux operating systems. Each occurrence of this
feature will pay for 100 minutes of usage. The purchase of this
feature occurs after the customer has 100 minutes of use on processor
cores in the shared processor pool that are not permanently active.

Attributes provided: Payment for Utility CoD use of processor
#EPBS (AIX or Linux)

Attributes required: At least one processor core on #EPBS that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports AIX and Linux

(#EPKR) Proc CoD Utility Billing, 100 Proc-mins. for #EPBS, IBM i

Provides payment for temporary use of processor feature #EPBS with
supported IBM i operating systems. Each occurrence of this feature
will pay for 100 minutes of usage. The purchase of this feature
occurs after the customer has 100 minutes of use on processor cores in
the shared processor pool that are not permanently active.

Attributes required: At least one processor core on #EPBS that is
not permanently active & Elastic CoD Processor Enablement.

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: N/A

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: Does not apply

Note: Supports IBM i

(#EPTH) - Horizontal PDU Mounting Hardware

This feature ships the hardware required to properly horizontally
mount one #EPTG/EPTJ, #EPTK/EPTL, #EPTM/EPTN or #EPTP/EPTQ PDU in a
1U 19-inch rack. A 1U blank panel for the front of the rack for
air-flow control is included.

Without this hardware, the PDU can be mounted vertically in the
rack's side pockets, but can only be poorly mounted horizontally. The
front end of the PDU will be firmly attached to the rear of the rack.
But the front of the PDU will be unsupported toward the middle of the
rack. Without this hardware, the unsupported end of the PDU will rest
on the hardware mounted immediately below it. If that underlying
hardware is removed from the rack there is no support for the PDU.

Important Note: This feature code is typically used for an MES
order and not for an original order of a new rack with #EPTn PDUs. As
part of factory integration, IBM Manufacturing automatically adds
this hardware without a feature code and at no additional charge when
its #EPTn PDU placement logic calls for horizontally mounted PDUs.
Use this feature code when (1) converting an existing vertically
mounted #EPTn PDU to horizontal mounting or (2) separately ordering a
#EPTn PDU for horizontal field installation.

(#EPTJ) - High Function 9xC19 PDU

Switched, Monitoring

This is an intelligent, switched 200-240 volt AC Power
Distribution Unit (PDU) with nine C19 receptacles on the front of the
PDU. The PDU is mounted on the rear of the rack making the nine C19
receptacles easily accessible. Each receptacle has a 20 amp circuit
breaker. Depending on country wiring standards the PDU is
single-phase or three-phase wye. See three-phase #EPTK/EPTL for
countries which do not use wye wiring.

The PDU can be mounted vertically in rack side pockets or it can
be mounted horizontally. If mounted horizontally, it uses 1 EIA (1U)
of rack space. See feature #EPTH for horizontal mounting hardware.

Device power cords with a C20 plug connect to C19 PDU receptacles
and are ordered separately. One country-specific wall line cord is
also ordered separately and attaches to a UTG524-7 connector on the
front of the PDU. Supported line cords include features #6489, #6491,
#6492, #6653, #6654, #6655, #6656, #6657, #6658, and #6667.

Two RJ45 ports on the front of the PDU enable the client to
monitor each receptacle's electrical power usage and to remotely
switch any receptacle on or off. The PDU is shipped with a generic
PDU password and IBM strongly urges clients to change it upon
installation.

There are also three C13 receptacles on the rear of the PDU
positioned toward the middle of the rack. These are generally not
easily accessed and therefore IBM does not generally recommend
their use.

#EPTG and #EPTJ are identical PDUs. Up to one lower price #EPTG
can be ordered with a new 7014-T42/T00 rack in place of a no-charge
#9188 PDU.

For comparison, this is most similar to the earlier generation
#7189 PDU

Attributes provided: Nine C19 PDU - switched, power monitoring

Attributes required: PDU wall line cord & space in 19-inch rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EPTL) - High Function 9xC19 PDU 3-Phase

Switched, Monitoring

This is an intelligent, switched 208 volt 3-phase AC Power
Distribution Unit (PDU) with nine C19 receptacles on the front of the
PDU. The PDU is mounted on the rear of the rack making the nine C19
receptacles easily accessible. Each receptacle has a 20 amp circuit
breaker.

The PDU can be mounted vertically in rack side pockets or it can
be mounted horizontally. If mounted horizontally, it uses 1 EIA (1U)
of rack space. See feature #EPTH for horizontal mounting hardware.

Device power cords with a C20 plug connect to C19 PDU receptacles
and are ordered separately. One wall line cord is provided with the
PDU (no separate feature code) and has a IEC60309 60A plug (3P+G).
The PDU supports up to 48 amps.

Two RJ45 ports on the front of the PDU enable the client to monitor
each receptacle's electrical power usage and to remotely switch
any receptacle on or off. The PDU is shipped with a generic PDU
password and IBM strongly urges clients to change it upon
installation.

There are also three C13 receptacles on the rear of the PDU
positioned toward the middle of the rack. These are generally not
easily accessed and therefore IBM does not generally recommend
their use.

#EPTK and #EPTL are identical PDUs. Up to one lower price #EPTK
can be ordered with a new 7014-T42/T00 rack in place of a no-charge
#9188 PDU.

For comparison, this is most similar to the earlier generation
#7196 PDU

(#EPTN) - High Function 12xC13 PDU

Switched, Monitoring

This is an intelligent, switched 200-240 volt AC Power
Distribution Unit (PDU) with twelve C13 receptacles on the front of
the PDU. The PDU is mounted on the rear of the rack making the twelve
C13 receptacles easily accessible. Each receptacle has a 20 amp
circuit breaker. Depending on country wiring standards the PDU is
single-phase or three-phase wye. See three-phase #EPTK/EPTL for
countries which do not use wye wiring.

The PDU can be mounted vertically in rack side pockets or it can
be mounted horizontally. If mounted horizontally, it uses 1 EIA (1U)
of rack space. See feature #EPTH for horizontal mounting hardware.

Device power cords with a C14 plug connect to C13 PDU receptacles
and are ordered separately. One country-specific wall line cord is
also ordered separately and attaches to a UTG524-7 connector on the
front of the PDU. Supported line cords include features #6489, #6491,
#6492, #6653, #6654, #6655, #6656, #6657, #6658, and #6667.

Two RJ45 ports on the front of the PDU enable the client to monitor
each receptacle's electrical power usage and to remotely switch any
receptacle on or off. The PDU is shipped with a generic PDU
password and IBM strongly urges clients to change it upon installation.

#EPTM and #EPTN are identical PDUs. Up to one lower price #EPTM
can be ordered with a new 7014-T42/T00 rack in place of a no-charge
#9188 PDU.

For comparison, this is most similar to the earlier generation
#7109 PDU

Attributes provided: Twelve C13 PDU - switched, power monitoring

Attributes required: PDU wall line cord & space in 19-inch rack

Minimum required: 0

Maximum allowed: 9999 (Initial order maximum: 250)

OS level required: None

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EPTQ) - High Function 12xC13 PDU 3-Phase

Switched, Monitoring

This is an intelligent, switched 208 volt 3-phase AC Power
Distribution Unit (PDU) with twelve C13 receptacles on the front of
the PDU. The PDU is mounted on the rear of the rack making the twelve
C13 receptacles easily accessible. Each receptacle has a 20 amp
circuit breaker.

The PDU can be mounted vertically in rack side pockets or it can
be mounted horizontally. If mounted horizontally, it uses 1 EIA (1U)
of rack space. See feature #EPTH for horizontal mounting hardware.

Device power cords with a C14 plug connect to C13 PDU receptacles
and are ordered separately. One wall line cord is provided with the
PDU (no separate feature code) and has a IEC60309 60A plug (3P+G).
The PDU supports up to 48 amps.

Two RJ45 ports on the front of the PDU enable the client to monitor
each receptacle's electrical power usage and to remotely switch any
receptacle on or off. The PDU is shipped with a generic PDU
password and IBM strongly urges clients to change it upon installation.

#EPTP and #EPTQ are identical PDUs. Up to one lower price #EPTP
can be ordered with a new 7014-T42/T00 rack in place of a no-charge
#9188 PDU.

For comparison, this is most similar to the earlier generation
#7196 PDU, but offers C13 receptacles

This feature provides 150 cables of feature 3452 which is a 6 meter
SAS cable that connects a SAS adapter to a SAS I/O enclosure. This YO
cable has three connectors, one Mini-SAS HD (High Density)connector
and two Mini SAS connectors. The Mini-SAS HD connector attaches to a
SAS adapter such as the #5913 1.8GB RAID SAS Adapter. Both Mini SAS
connectors attach to the same I/O drawer (enclosure) such as a
#5887/EL1S EXP24S SAS I/O drawer. This cable can support up to 6Gb
throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

This feature provides 150 10 meter SAS cables that connects a SAS
adapter to a SAS I/O enclosure. This YO cable has three connectors,
one Mini-SAS HD (High Density)connector and two Mini SAS connectors.
The Mini-SAS HD connector attaches to a SAS adapter such as the #5913
1.8GB RAID SAS Adapter. Both Mini SAS connectors attach to the same
I/O drawer (enclosure) such as a #5887/EL1S EXP24S SAS I/O drawer.
This cable can support up to 6Gb throughput.

Multiple cable length feature codes are available. Choose the
cable length that best matches the distance between the adapter and
the SAS I/O drawer.

(#EQ0C) - Quantity of 150 #ES0C

This feature ships a quantity of 150 #ES0C disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES0C

Attributes required: See Feature ES0C

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#EQ0D) - Quantity of 150 #ES0D

This feature ships a quantity of 150 #ES0D disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES0D

Attributes required: See feature ES0D

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Supported

CSU: Yes

Return parts MES: No

(#EQ0G) - Quantity 150 of #ES0G (775GB SSD SFF-2)

(No Longer Available as of August 31, 2017)

This feature ships a quantity of 150 #ES0G disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES0G

Attributes required: See Feature ES0G

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ0H) - Quantity 150 of #ES0H (775GB SSD SFF-2)

(No Longer Available as of August 31, 2017)

This feature ships a quantity of 150 #ES0H disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES0H

Attributes required: See feature ES0H

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported only with VIOS

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ0Q) - Quantity 150 of #ES0Q 387GB SFF-2 4k SSD (AIX/Linux)

(No Longer Available as of August 31, 2017)

This feature ships a quantity 150 of #ES0Q SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single disk unit feature. This feature
remains on the inventory records

Attributes provided: Quantity of 150 #ES0Q

Attributes required: See #ES0Q

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ0R) - Quantity 150 of #ES0R 387GB SFF-2 4k SSD (IBM i)

(No Longer Available as of August 31, 2017)

This feature ships a quantity 150 of #ES0R SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single disk unit feature. This feature
remains on the inventory records.

Attributes provided: Quantity of 150 #ES0R

Attributes required: See #ES0R

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ0S) - Quantity 150 of #ES0S 775GB SFF-2 4k SSD (AIX/Linux)

(No Longer Available as of August 31, 2017)

This feature ships a quantity 150 of #ES0S SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single disk unit feature. This feature
remains on the inventory records.

Attributes provided: Quantity of 150 #ES0S

Attributes required: See #ES0S

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ0T) - Quantity 150 of #ES0T 775GB SFF-2 4k SSD (IBM i)

(No Longer Available as of August 31, 2017)

This feature ships a quantity 150 of #ES0T SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single disk unit feature. This feature
remains on the inventory records.

Attributes provided: Quantity of 150 #ES0T

Attributes required: See #ES0T

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ19) - Quantity 150 of #ES19 (387GB SSD SFF-2)

(No Longer Available as of August 31, 2017)

This feature ships a quantity of 150 #ES19 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES19

Attributes required: See Feature ES19

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ1A) - Quantity 150 of #ES1A (387GB SSD SFF-2)

(No Longer Available as of August 31, 2017)

This feature ships a quantity of 150 #ES1A disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records

Attributes provided: Quantity of 150 #ES1A

Attributes required: See feature ES1A

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ38) - Quantity 150 of #1738 (856GB SFF-2 disk)

This feature ships a quantity of 150 #1738 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1738

Attributes required: See feature #1738

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux not supported

AIX not supported

IBM i 7.1 TR9 and 7.2 TR1 or later supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ52) - Quantity 150 of #1752 (900GB SFF-2 disk)

(No Longer Available as of January 20, 2017)

This feature ships a quantity of 150 #1752 disk units. The
configurator may either generate this feature or allow users to select
this feature as they would any other single disk unit feature. This
feature remains on the inventory records.

Attributes provided: See feature #1752

Attributes required: See feature #1752

Minimum required: 0

Maximum allowed: 26 (Initial order maximum: 26)

OS level required:

Linux supported

AIX 6.1 and 7.1 or later supported

IBM i not supported

Refer to Software requirements for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ62) - Quantity 150 of #ES62 3.86-4.0 TB 7200 rpm 4k LFF-1 Disk

This feature ships a quantity of 150 #ES62 drives. The
configurator may either generate this feature or allow users to
select this feature as they would any other single drive feature.
This feature remains on the inventory records.

(#EQ64) - Quantity 150 of #ES64 7.72-8.0 TB 7200 rpm 4k LFF-1 Disk

This feature ships a quantity of 150 #ES64 drives. The
configurator may either generate this feature or allow users to
select this feature as they would any other single drive feature.
This feature remains on the inventory records.

(#EQ78) - Quantity 150 of #ES78 387GB SFF-2 SSD 5xx

This feature ships a quantity of 150 #ES78 solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES78

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES78

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ79) - Quantity 150 of #ES79 387GB SFF-2 SSD 5xx

This feature ships a quantity of 150 #ES79 solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES79

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES79

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ7E) - Quantity 150 of #ES7E 775GB SFF-2 SSD 5xx

This feature ships a quantity of 150 #ES7E solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES7E

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES7E

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ7F) - Quantity 150 of #ES7F 775GB SFF-2 SSD 5xx

This feature ships a quantity of 150 #ES7F solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES7F

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES7F

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ80) - Quantity 150 of #ES80 1.9TB SFF-2 SSD 4k

(No longer available as of January 23, 2018.)

This feature ships a quantity of 150 #ES80 SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single SSD feature. This feature remains on
the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feature #ES80

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

See #ES80 for specific O/S levels supported

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ81) - Quantity 150 of ES81 1.9TB SFF-2 SSD 4k

(No longer available as of January 23, 2018.)

This feature ships a quantity of 150 #ES81 SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single SSD feature. This feature remains on
the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feature #ES81

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required:

Refer to #ES81

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ85) - Quantity 150 of #ES85 387GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES85 solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES85

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES85

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ86) - Quantity 150 of #ES86 387GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES86 solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES86

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES86

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8C) - Quantity 150 of #ES8C 775GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8C solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES8C

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES8C

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8D) - Quantity 150 of #ES8D 775GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8D solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES8D

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES8D

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8F) - Quantity 150 of #ES8F 1.55TB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8F solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES8F

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES8F

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8G) - Quantity 150 of #ES8G 1.55TB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8G solid state drives
(SSDs). The configurator may either generate this feature or allow
users to select this feature as they would any other single SSD
feature. This feature remains on the inventory records.

Attributes provided: Ships 150 SSDs

Attributes required: See feat #ES8G

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: Refer to feature #ES8G

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8Y) - Quantity 150 of #ES8Y 931GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8Y SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single SSD feature. This feature remains on
the inventory records.

Attributes provided: see feature #ES8Y

Attributes required: 150 SFF-2 SAS bays in EXP24S or EXP24SX
drawers

Minimum required: 0

Maximum allowed: 13 (Initial order maximum: 13)

OS level required: see feature ES8Y

Initial Order/MES/Both/Supported: Both

CSU: Yes

Return parts MES: No

(#EQ8Z) - Quantity 150 of ES8Z 931GB SFF-2 SSD 4k

This feature ships a quantity of 150 #ES8Z SSD. The configurator
may either generate this feature or allow users to select this feature
as they would any other single SSD feature. This feature remains on
the inventory records.