In-Memory Computing With SAP HANA On IBM EX5 and X6 Systems
In-Memory Computing With SAP HANA On IBM EX5 and X6 Systems
In-Memory Computing With SAP HANA On IBM EX5 and X6 Systems
In-memory Computing
with SAP HANA on IBM
eX5 and X6 Systems
IBM System x solution for
SAP HANA
SAP HANA overview and
use cases
Operational aspects for
SAP HANA appliances
Martin Bachmaier
Ilya Krutov
ibm.com/redbooks
SG24-8086-02
Note: Before using this information and the product it supports, read the information in
Notices on page vii.
Contents
Notices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Trademarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix
Authors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . x
Now you can become a published author, too! . . . . . . . . . . . . . . . . . . . . . . . . . xi
Comments welcome. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi
Stay connected to IBM Redbooks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xii
Chapter 1. Basic concepts of in-memory computing . . . . . . . . . . . . . . . . . 1
1.1 Keeping data in-memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.1 Using main memory as the data store . . . . . . . . . . . . . . . . . . . . . . . . 2
1.1.2 Data persistence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Minimizing data movement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.1 Compression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.2 Columnar storage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.3 Pushing application logic to the database . . . . . . . . . . . . . . . . . . . . . 13
1.3 Dividing and conquering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3.1 Parallelization on multi-core systems . . . . . . . . . . . . . . . . . . . . . . . . 14
1.3.2 Data partitioning and scale-out . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
Chapter 2. SAP HANA overview. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
2.1 SAP HANA overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.1 SAP HANA architecture . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
2.1.2 SAP HANA appliance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.2 SAP HANA delivery model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.2.1 SAP HANA as an appliance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.2.2 SAP HANA tailored data center integration . . . . . . . . . . . . . . . . . . . 22
2.3 Sizing SAP HANA . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
2.3.1 Memory per core ratio for SAP HANA appliances . . . . . . . . . . . . . . 23
2.3.2 Sizing approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
2.4 SAP HANA software licensing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
Chapter 3. Software components and data replication methods . . . . . . . 29
3.1 SAP HANA software components . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
3.1.1 SAP HANA database . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
3.1.2 SAP HANA client. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.1.3 SAP HANA studio . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.1.4 SAP HANA studio repository. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
iii
iv
Contents
7.4.1
7.4.2
7.4.3
7.4.4
7.4.5
7.4.6
7.4.7
vi
Notices
This information was developed for products and services offered in the U.S.A.
IBM may not offer the products, services, or features discussed in this document in other countries. Consult your
local IBM representative for information on the products and services currently available in your area. Any
reference to an IBM product, program, or service is not intended to state or imply that only that IBM product,
program, or service may be used. Any functionally equivalent product, program, or service that does not infringe
any IBM intellectual property right may be used instead. However, it is the user's responsibility to evaluate and
verify the operation of any non-IBM product, program, or service.
IBM may have patents or pending patent applications covering subject matter described in this document. The
furnishing of this document does not grant you any license to these patents. You can send license inquiries, in
writing, to:
IBM Director of Licensing, IBM Corporation, North Castle Drive, Armonk, NY 10504-1785 U.S.A.
The following paragraph does not apply to the United Kingdom or any other country where such
provisions are inconsistent with local law: INTERNATIONAL BUSINESS MACHINES CORPORATION
PROVIDES THIS PUBLICATION "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR
IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some states do not allow disclaimer of
express or implied warranties in certain transactions, therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made to the
information herein; these changes will be incorporated in new editions of the publication. IBM may make
improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time
without notice.
Any references in this information to non-IBM websites are provided for convenience only and do not in any
manner serve as an endorsement of those websites. The materials at those websites are not part of the materials
for this IBM product and use of those websites is at your own risk.
IBM may use or distribute any of the information you supply in any way it believes appropriate without incurring any
obligation to you.
Any performance data contained herein was determined in a controlled environment. Therefore, the results
obtained in other operating environments may vary significantly. Some measurements may have been made on
development-level systems and there is no guarantee that these measurements will be the same on generally
available systems. Furthermore, some measurements may have been estimated through extrapolation. Actual
results may vary. Users of this document should verify the applicable data for their specific environment.
Information concerning non-IBM products was obtained from the suppliers of those products, their published
announcements or other publicly available sources. IBM has not tested those products and cannot confirm the
accuracy of performance, compatibility or any other claims related to non-IBM products. Questions on the
capabilities of non-IBM products should be addressed to the suppliers of those products.
This information contains examples of data and reports used in daily business operations. To illustrate them as
completely as possible, the examples include the names of individuals, companies, brands, and products. All of
these names are fictitious and any similarity to the names and addresses used by an actual business enterprise is
entirely coincidental.
COPYRIGHT LICENSE:
This information contains sample application programs in source language, which illustrate programming
techniques on various operating platforms. You may copy, modify, and distribute these sample programs in any
form without payment to IBM, for the purposes of developing, using, marketing or distributing application programs
conforming to the application programming interface for the operating platform for which the sample programs are
written. These examples have not been thoroughly tested under all conditions. IBM, therefore, cannot guarantee or
imply reliability, serviceability, or function of these programs. You may copy, modify, and distribute these sample
programs in any form without payment to IBM for the purposes of developing, using, marketing, or distributing
application programs conforming to IBM's application programming interfaces.
vii
Trademarks
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business
Machines Corporation in the United States, other countries, or both. These and other IBM trademarked
terms are marked on their first occurrence in this information with the appropriate symbol ( or ),
indicating US registered or common law trademarks owned by IBM at the time this information was
published. Such trademarks may also be registered or common law trademarks in other countries. A current
list of IBM trademarks is available on the Web at http://www.ibm.com/legal/copytrade.shtml
The following terms are trademarks of the International Business Machines Corporation in the United States,
other countries, or both:
AIX
BladeCenter
DB2
Global Business Services
Global Technology Services
GPFS
IBM SmartCloud
IBM
Passport Advantage
POWER
PureFlex
RackSwitch
Redbooks
Redbooks (logo)
System x
System z
Tivoli
X-Architecture
z/OS
viii
Preface
This third edition of this IBM Redbooks publication describes in-memory
computing appliances from IBM and SAP that are based on IBM eX5 and X6
flagship systems and SAP HANA. It covers the basic principles of in-memory
computing, describes the IBM eX5 and X6 hardware offerings, and explains the
corresponding SAP HANA IT landscapes using these offerings.
This book also describes the architecture and components of the IBM System x
solution for SAP HANA, with IBM General Parallel File System (GPFS) as a
cornerstone. The SAP HANA operational disciplines are explained in detail:
Scalability options, high availability and disaster recovery, backup and restore,
and virtualization possibilities for SAP HANA appliances.
The following topics are covered:
Basic principles of in-memory computing
SAP HANA overview
Software components and replication methods
SAP HANA use cases and integration scenarios
The System x solution for SAP HANA
SAP IT landscapes using the System x solution for SAP HANA
Options for business continuity (high availability, disaster recovery, and
backup and restore)
SAP HANA operations
This book is intended for SAP administrators and technical solution architects. It
is also for IBM Business Partners and IBM employees who want to know more
about the SAP HANA offering and other available IBM solutions for SAP clients.
ix
Authors
This book was produced by a team of specialists from around the world working
at the IBM International Technical Support Organization (ITSO), Raleigh Center.
Martin Bachmaier is an IT Versatilist in the IBM
hardware development lab at Boeblingen, Germany.
He is part of the team developing the System x solution
for SAP HANA. Martin has a deep background in
designing, implementing, and managing scale-out data
centers, HPC clusters, and cloud environments, and
has worked with GPFS for eight years. He gives
university lectures, and likes to push IT limits. Martin is
an IBM Certified Systems Expert. He holds the CCNA,
CCNA Security, and VMware Certified Professional
credentials and has authored over a dozen books and
papers.
Ilya Krutov is a Project Leader at the ITSO Center in
Raleigh and has been with IBM since 1998. Before
joining the ITSO, Ilya served in IBM as a Run Rate
Team Leader, Portfolio Manager, Brand Manager,
Technical Sales Specialist, and Certified Instructor. Ilya
has expertise in IBM System x, BladeCenter and
PureFlex System products, server operating
systems, and networking solutions. He has authored
over 170 books, papers, and product guides. He has a
Bachelors degree in Computer Engineering from the
Moscow Engineering and Physics Institute.
Thanks to the authors of the previous edition of this book:
Gereon Vey
Tomas Krojzl
Special thanks to Irene Hopf, a Senior Architect in the IBM SAP International
Competence Center (ISICC) in Walldorf, Germany, who made a significant
contribution to the development of this book by extensively consulting and
guiding the team on SAP HANA topics
Thanks to the following people for their contributions to this project:
Tamikia Barrow, Cheryl Gera, Chris Rayns, Wade Wallace, David Watts,
Debbie Willmschen
International Technical Support Organization, Raleigh Center
Comments welcome
Your comments are important to us!
We want our books to be as helpful as possible. Send us your comments about
this book or other IBM Redbooks publications in one of the following ways:
Use the online Contact us review Redbooks form found at:
ibm.com/redbooks
Send your comments in an email to:
[email protected]
Mail your comments to:
IBM Corporation, International Technical Support Organization
Dept. HYTD Mail Station P099
2455 South Road
Poughkeepsie, NY 12601-5400
Preface
xi
xii
Chapter 1.
Basic concepts of
in-memory computing
In-memory computing is a technology that allows the processing of massive
quantities of data in main memory to provide immediate results from analysis and
transaction. The data that is processed is ideally real-time data (that is, data that
is available for processing or analysis immediately after it is created).
To achieve the preferred performance, in-memory computing follows these basic
concepts:
Keep data in main memory to speed up data access.
Minimize data movement by using the columnar storage concept,
compression, and performing calculations at the database level.
Divide and conquer. Use the multi-core architecture of modern processors
and multi-processor servers, or even scale out into a distributed landscape, to
grow beyond what can be supplied by a single server.
This chapter describes those basic concepts with the help of a few examples. It
does not describe the full set of technologies that are employed with in-memory
databases, such as SAP HANA, but it does provide an overview of how
in-memory computing is different from traditional concepts.
1,000,000
100,000
150x
10,000
1,000
100
2,000x
10
1
17x
0,1
12x
0,01
0,001
CPU register
CPU Cache
Volatile
RAM
SSD/Flash
Hard disk
Non-volatile
Figure 1-1 Data access times of various storage types relative to RAM (logarithmic scale)
The main memory is the fastest storage type that can hold a significant amount
of data. Although CPU registers and CPU cache are faster to access, their usage
is limited to the actual processing of data. Data in main memory can be accessed
more than a hundred thousand times faster than data on a spinning hard disk
drive (HDD), and even flash technology storage is about a thousand times slower
than main memory. Main memory is connected directly to the processors through
a high-speed bus, and hard disks are connected through a chain of buses (QPI,
PCIe, and SAN) and controllers (I/O hub, RAID controller or SAN adapter, and
storage controller).
Compared with keeping data on disk, keeping the data in main memory can
improve database performance just by the advantage in access time.
Time
Data savepoint
to persistent
storage
Log written
to persistent storage
(committed transactions)
Power failure
After a power failure, the database can be restarted much like a disk-based
database. The database pages are restored from the savepoints and then the
database logs are applied (rolled forward) to restore the changes that were not
captured in the savepoints. This action ensures that the database can be
restored in memory to the same state as before the power failure.
1.2.1 Compression
Even though todays memory capacities allow keeping enormous amounts of
data in-memory, compressing the data in-memory is still preferable. The goal is
to compress data in a way that does not use up the performance that is gained
while still minimizing data movement from RAM to the processor.
By working with dictionaries to represent text as integer numbers, the database
can compress data significantly and thus reduce data movement while not
imposing additional CPU load for decompression; in fact, it can add to the
performance, as shown in Figure 1-5 on page 10.
Row
ID
Date/
Time
Material
Customer
Name
Quantity
Customers
Material
Chevrier
MP3 Player
Di Dio
Radio
Dubois
Refrigerator
Miller
Stove
Newman
Laptop
14:05
Radio
Dubois
14:11
Laptop
Di Dio
14:32
Stove
Miller
14:38
MP3 Player
Newman
Row
ID
14:48
Radio
Dubois
14:55
Refrigerator
Miller
15:01
Stove
Chevrier
Date/
Time
Material
Customer
Name
Quantity
845
851
872
878
888
895
901
On the left side of Figure 1-3, the original table is shown, and it contains text
attributes (that is, material and customer name) in their original representation.
The text attribute values are stored in a dictionary (upper right), and an integer
value is assigned to each distinct attribute value. In the table, the text is replaced
by the corresponding integer value, as defined in the dictionary. The date and
time attribute also are converted to an integer representation. Using dictionaries
for text attributes reduces the size of the table because each distinct attribute
value must be stored only once in the dictionary; therefore, each additional
occurrence in the table must be referred to by the corresponding integer value.
The compression factor that is achieved by this method highly depends on data
being compressed. Attributes with few distinct values compress well, but
attributes with many distinct values do not benefit as much.
Although there are other, more effective compression methods that can be
employed with in-memory computing, for these methods to be useful, they must
have the correct balance between compression effectiveness, which gives you
more data in your memory, or less data movement (that is, higher performance),
resources that are needed for decompression, and data accessibility (that is, how
much unrelated data must be uncompressed to get to the data that you need).
Dictionary compression combines good compression effectiveness with low
decompression resources and high data access flexibility.
Row-based
Row
ID
Date/
Time
Column-based
Material
Customer
Name
Quantity
Row
ID
Date/
Time
Material
Customer
Name
Quantity
845
845
851
851
872
872
878
878
888
888
895
895
901
901
Row-based store
1
845
851
851
872
878
872
878
Column-based store
1
845
Both storage models have benefits and drawbacks, which are listed in Table 1-1.
Table 1-1 Benefits and drawbacks of row-based and column-based storage
Row-based storage
Benefits
Column-based storage
Drawbacks
The drawbacks of column-based storage are not as grave as they seem. In most
cases, not all attributes (that is, column values) of a row are needed for
processing, especially in analytic queries. Also, inserts or updates to the data are
less frequent in an analytical environment.1 SAP HANA implements both a
row-based storage and a column-based storage; however, its performance
originates in the usage of column-based storage in memory. The following
sections describe how column-based storage is beneficial to query performance
and how SAP HANA handles the drawbacks of column-based storage.
An exception is bulk loads (for example, when replicating data in the in-memory database, which can be
handled differently).
Get all records with Customer Name Miller and Material Refrigerator
Dictionary lookup of the strings
Strings are only compared once!
Customers
Material
Chevrier
MP3 Player
Di Dio
Radio
Dubois
Refrigerator
Miller
Stove
Newman
Laptop
Customer
Material
0 0 1 0 0 1 0
0 0 0 0 0 1 0
Combine
bit-wise AND
0 0 0 0 0 1 0
Resultset
The resulting records can be assembled from the column stores fast, because positions are known
(here: 6th position in every column)
The query asks to get all records with Miller as the customer name and
Refrigerator as the material.
First, the strings in the query condition are looked up in the dictionary. Miller is
represented by the number 4 in the customer name column. Refrigerator is
represented by the number 3 in the material column. This lookup must be done
only once. Subsequent comparisons with the values in the table are based on
integer comparisons, which are less resource-intensive than string comparisons.
10
In a second step, the columns are read that are part of the query condition (that
is, the Customer and Material columns). The other columns of the table are not
needed for the selection process. The columns are then scanned for values
matching the query condition. That is, in the Customer column, all occurrences of
4 are marked as selected, and in the Material column, all occurrences of 3 are
marked.
These selection marks can be represented as bitmaps, which are data structures
that allows efficient Boolean operations on them, which are used to combine the
bitmaps of the individual columns to a bitmap representing the selection or
records matching the entire query condition. In this example, the record number
6 is the only matching record. Depending on the columns that are selected for
the result, the additional columns must be read to compile the entire record to
return it. But because the position within the column is known (record number 6),
only the parts of the columns that contain the data for this record must be read.
This example shows how compression can limit not only the amount of data that
must be read for the selection process, but can simplify the selection itself while
the columnar storage model further reduces the amount of data that is needed
for the selection process. Although the example is simplified, it illustrates the
benefits of dictionary compression and columnar storage.
Efficient Transaction Processing in SAP HANA Database - The End of a Column Store Myth, found
at http://dl.acm.org/citation.cfm?id=2213946.
11
Figure 1-6 shows the lifecycle management for database records in the
column-store.
L1 Delta
Merge
Bulk Insert
L2 Delta
Merge
Main store
Unified Table
Read
Figure 1-6 Lifetime management of a data record in the SAP HANA column-store
12
13
14
The individual database tables can be placed on different servers within the
cluster, or tables bigger than what a single server can hold can be split into
several partitions, either horizontally (a group of rows per partition) or vertically (a
group of columns per partition), with each partition on a separate server within
the cluster.
15
16
Chapter 2.
17
18
SAP HANA
Studio
SAP HANA
Studio Repository
SQL
MDX
SQL Script
Calculation Engine
Relational Engines
Software Update
Manager
Row
Store
Column
Store
SAP HANA
Client
Transaction
Manager
Authorization
Manager
Metadata
Manager
Page
Management
Persistency Layer
Logger
Data Volumes
Persistent Storage
Log Volumes
LM Structure
JVM
SAP CAR
19
20
21
22
XS
S and S+
M and M+
Compressed
data in
memory
64 GB
128 GB
256 GB
512 GB
Server main
memory
128 GB
256 GB
512 GB
1024 GB
Number of
CPUs
23
The T-shirt sizes, S+ and M+, denote upgradeable versions of the S and M sizes:
S+ delivers capacity that is equivalent to S, but the hardware is upgradeable
to an M size.
M+ delivers capacity that is equivalent to M, but the hardware is upgradeable
to an L size.
These T-shirt sizes are used when relevant growth of the data size is expected.
In addition to these standard T-shirt sizes, which apply to all use cases of SAP
HANA, there are configurations that are specific to and limited for use with SAP
Business Suite applications that are powered by SAP HANA. Table 2-2 shows
these additional T-shirt sizes.
Table 2-2 Additional T-shirt sizes for SAP Business Suite that are powered by SAP HANA
SAP T-shirt size
XL
XXL
Compressed data in
memory
512 GB
1 TB
2 TB
1 TB
2 TB
4 TB
Number of CPUs
The workload for the SAP Business Suite applications has different
characteristics. It is less CPU-bound and more memory-intensive than a
standard SAP HANA workload. Therefore, the memory per core ratio is different
than for the standard T-shirt sizes. All workloads can be used on the T-shirt sizes
in Table 2-1 on page 23, including SAP Business Suite applications with SAP
HANA as the primary database. The T-shirt sizes in Table 2-2 are specific to and
limited for use with SAP Business Suite applications that are powered by SAP
HANA only.
Section 4.5.3, SAP Business Suite powered by SAP HANA on page 76 has
more information about which of the SAP Business Suite applications are
supported by SAP HANA as the primary database.
For more information about T-shirt size mappings to IBM eX5 based building
blocks, see 6.1, IBM eX5 based environments on page 146.
24
To simplify things and to avoid confusion, IBM decided to introduce the following
generic naming schema that replaces T-shirt sizes. Instead of S, M, L, and so on,
names for different types of building blocks with the following naming convention
are used:
XXX-YS-NNNN-Z (dashes can also be left out)
XXX means servers model type, YS is the number of installed processors (or
sockets), NNNN is the memory amount in gigabytes, and Z is stand-alone (S) or
cluster (C) node.
For example, model AC3-2S-256-S represents a stand-alone 4S server
(AC3-models have four sockets) with two processors that are installed and 256
GB of memory. This model replaces former T-shirt size S.
A model AC4-8S-1024-C represents an 8-socket cluster node (AC4-models have
eight sockets) with eight processors that are installed and 1 TB of main memory.
This new naming schema is used in this book for all the X6 based SAP HANA
landscapes that are described 6.2, IBM X6 based environments on page 157.
25
26
Although the SAP HANA software comes packaged in two different editions
(platform and enterprise edition), the SAP HANA software licensing depends on
the use case. Chapter 4, SAP HANA integration scenarios on page 51
describes different use cases for SAP HANA.
Tip: Licensing SAP HANA is handled by SAP and there are different options
depending on the use case. Because metrics can always change, discuss
licensing options for your particular use case with SAP.
27
28
Chapter 3.
29
User
workstation
SAP HANA
studio
(optional)
SAP HANA
client
(optional)
Application
Server
SAP HANA
client
SAP host
agent
Data Modeling
Row Store
SAP HANA
client
SAP HANA
LM structure
SAP HANA
studio
repository
SAP HANA
Lifecycle
Manager
Column Store
SAP HANA
database
Application
Function
Library
SAP
liveCache
Applications
Figure 3-1 Distribution of software components that are related to SAP HANA
30
31
The SAP HANA client libraries are delivered in 32-bit and 64-bit editions. It is
important always to use the correct edition based on the architecture of the
application that uses this client. 32-bit applications cannot use 64-bit client
libraries and vice versa.
To access the SAP HANA database from Microsoft Excel, you can also use a
special 32-bit edition of the SAP HANA client that is called SAP HANA client
package for Microsoft Excel.
The SAP HANA client is compatible with earlier versions, meaning that the
revision of the client must be the same or higher than the revision of the SAP
HANA database.
The SAP HANA client libraries must be installed on every machine where
connectivity to the SAP HANA database is required. This includes not only all
servers, but also user workstations that are hosting applications that are directly
connecting to the SAP HANA database (for example, SAP BusinessObjects
Client Tools or Microsoft Excel).
Whenever the SAP HANA database is updated to a more recent revision, all
clients that are associated with this database also must be upgraded.
For more information about how to install the SAP HANA client, see the official
SAP guide SAP HANA Database - Client Installation Guide, which can be
downloaded at the following website:
http://help.sap.com/hana_platform
32
The following main function areas are provided by the SAP HANA studio (each
function area is also illustrated by a corresponding figure of the user interface):
Database administration
The key functions are stopping and starting the SAP HANA databases, status
overview, monitoring, performance analysis, parameter configuration, tracing,
and log analysis.
Figure 3-2 shows the SAP HANA studio user interface for database
administration.
33
Security management
This function area provides tools that are required to create users, to define
and assign roles, and to grant database privileges.
Figure 3-3 shows an example of the user definition window.
34
Data management
These functions create, change, or delete database objects (such as tables,
indexes, and views), and include commands to manipulate data (for example,
insert, update, delete, or perform a bulk load).
Figure 3-4 shows an example of the table definition window.
35
Modeling
This is the user interface that is used to work with models (metadata
descriptions about how source data is transformed in resulting views),
including the possibility to define new custom models, and to adjust or delete
existing models.
Figure 3-5 shows a simple analytic model.
36
Content management
These are the functions offering the possibility to organize models in
packages, to define delivery units for transport into a subsequent SAP HANA
system, or to export and import individual models or whole packages.
Content management functions are accessible from the main window in the
modeler perspective, as shown in Figure 3-6.
Figure 3-6 SAP HANA studio - content functions on the main window of modeler perspective
37
Replication management
Data replication into the SAP HANA database is controlled from the data
provisioning window in the SAP HANA studio, where new tables can be
scheduled for replication, suspended, or replication for a particular table can
be interrupted.
Figure 3-7 shows an example of a data provisioning window.
38
Lifecycle Management
The SAP HANA solution offers the possibility to download and install
automatically updates to SAP HANA software components. This function is
controlled from the Lifecycle Manager window in the SAP HANA studio.
Figure 3-8 shows an example of such a window.
The SAP HANA database queries are consumed indirectly by using front-end
components, such as SAP BusinessObjects BI 4.0 clients. Therefore, the SAP
HANA studio is required only for administration or development and is not
needed for users.
The SAP HANA studio runs on the Eclipse platform; therefore, every user must
have a recent version of the Java Runtime Environment (JRE) installed that uses
the same architecture (64-bit SAP HANA studio has 64-bit JRE as a
prerequisite).
Currently supported platforms are Windows 32-bit, Windows 64-bit, and Linux
64-bit.
39
The SAP HANA studio is also compatible with earlier versions, meaning that the
revision level of the SAP HANA studio must be the same or higher than the
revision level of the SAP HANA database. However, based on practical
experience, the preferred approach is to keep SAP HANA studio on same
revision level as the SAP HANA database whenever possible. Installation and
parallel usage of multiple revisions of SAP HANA studio on one workstation is
possible. When using one SAP HANA studio instance for multiple SAP HANA
databases, the revision level of the SAP HANA studio must be the same or
higher revision level than the highest revision level of the SAP HANA databases
being connected to.
SAP HANA studio must be updated to a more recent version on all workstations
whenever the SAP HANA database is updated. This can be automated by using
Update Server, which can be configured by using SAP HANA Lifecycle Manager
(HLM). There are more details about HLM in 3.1.7, SAP HANA Lifecycle
Manager on page 41. Using HLM is the best way to keep installations of SAP
HANA studio synchronized with the SAP HANA database revision.
For more information about how to install the SAP HANA studio, see the official
SAP guide, SAP HANA Database - Studio Installation Guide, which is available
for download at the following website:
http://help.sap.com/hana_platform
40
41
The SAP HLM is a tool that automates many installation, configuration, and
deployment tasks that are related to SAP HANA environments. It offers the
following functions:
42
Like Unified Installer, these two utilities are mainly intended for hardware vendors
to automate initial deployment of SAP HANA on their appliance models. System
administrators are expected to use SAP HANA Lifecycle Manager (HLM)
described in 3.1.7, SAP HANA Lifecycle Manager on page 41.
Main features of the SAP HANA Lifecycle Management tools are:
Install SAP HANA system (single-node or scale-out)
Add host to scale-out SAP HANA system
Enable or disable auto-start of SAP HANA database instance
Regeneration of SSL certificates that are used by SAP HANA Lifecycle
Manager
Installation and update of additional SAP HANA components:
Application Function Library
SAP HANA Client
SAP HANA Studio
SAP HANA Database server
SAP HANA Lifecycle Manager (HLM)
SAP liveCache applications
Update of server-side studio repository
Install and update SAP host agent
For more information about SAP HANA Lifecycle Management tools, see the
official SAP guide, SAP HANA Server Installation Guide, which is available for
download at the following location:
http://help.sap.com/hana_platform
With monitor content update and additional SAP notes for SP02
43
The SMD agent is an optional component, which can be installed on the SAP
HANA appliance. It enables diagnostic tests of the SAP HANA appliance through
SAP Solution Manager. The SMD agent provides access to the database logs
and the file system, and collects information about the systems CPU and
memory consumption through the SAP host agent.
More information about how to deploy SMD agent is provided in the official SAP
guide, SAP HANA Update and Configuration Guide, which is available for
download at the following location:
http://help.sap.com/hana_platform
44
SAP HANA
database
Source System
SAP ERP
Trigger-Based Replication
Application Layer
ETL-Based Replication
Embedded BW
Database
Log
File
Extractor-Based Replication
Log-Based Replication
The following sections describe these replication methods for SAP HANA in
more detail.
45
46
SAP BusinessObjects Data Services provides several kinds of data quality and
data transformation functions. Due to the rich feature set that is available,
implementation time for the ETL-based replication is longer than for the other
replication methods. SAP BusinessObjects Data Services offers integration with
SAP HANA. SAP HANA is available as a predefined data target for the load
process.
The ETL-based replication server is the ideal solution for all SAP HANA
customers who need data replication from non-SAP data sources.
47
48
Only certain versions of IBM DB2 on AIX, Linux, and HP-UX are supported by this replication
method.
To set up replication with the Sybase Replication Server, the definition and
content of tables that are chosen to be replicated must be copied initially from the
source database to the SAP HANA database. This initial load is done with the
R3Load program, which is also used for database imports and exports. Changes
in tables during initial copy operation are captured by the Sybase Replication
Server; therefore, no system downtime is required.
This replication method is recommended only for SAP customers who were
invited to use it during the ramp-up of SAP HANA 1.0. It was part of the SAP
HANA Enterprise Extended Edition, which was discontinued with the introduction
of SAP HANA 1.0 SPS05 in November 2012.3
SAP recommends that you use trigger-based data replication using the SAP
Landscape Transformation Replicator.
49
Real-Time
Near Real-Time
Real-Time Capabilities
Preferred by SAP
SAP LT System
Direct Extractor
Connection
Unicode only
Very limited DB support
Data Conversion Capabilities
Sybase
Replication Server
Real Real-Time
The replication method that you choose depends on the requirements. When
real-lime replication is needed to provide benefit to the business, and the
replication source is an SAP system, then the trigger-based replication is the
best choice. Extractor-based replication might keep project costs down by
reusing existing transformations. ETL-based replication provides the most
flexibility regarding data source, data transformation, and data cleansing options,
but does not provide real-time replication.
50
Chapter 4.
51
Technology platform
Operational reporting
Accelerator
In-memory products
Next generation applications
Accelerator
Operational
Reporting
In-Memory
Products
Data Modeling
Technology
platform
Column Store
Row Store
Next
Generation
Applications
SAP HANA
Figure 4-1 Basic use case scenarios that are defined by SAP in session EIM205
These five basic use case scenarios describe the elementary ways that SAP
HANA can be integrated. This chapter covers each of these use case scenarios
in dedicated sections.
SAP maintains a SAP HANA Use Case Repository with specific examples for
how SAP HANA can be integrated. This repository is online at the following
website:
http://www.experiencesaphana.com/community/resources/use-cases
52
The use cases in this repository are divided into categories that are based on
their relevance to a specific industry sector. It is a good idea to review this
repository to find inspiration about how SAP HANA can be used in various
scenarios.
Data Modeling
Non-SAP
or SAP
data source
Non-SAP
application
Column Store
Row Store
SAP HANA
SAP
Reporting
and Analytics
SAP HANA is not technologically dependent on other SAP products and can be
used independently as the only one SAP component in the clients information
technology (IT) landscape. However, SAP HANA can be easily integrated with
other SAP products, such as SAP BusinessObjects BI platform for reporting or
SAP BusinessObjects Data Services for ETL replication, which gives clients the
possibility to use only the components that are needed.
There are many ways that SAP HANA can be integrated into a client landscape,
and it is not possible to describe all combinations. Software components around
the SAP HANA offering can be seen as building blocks, and every solution must
be assembled from the blocks that are needed in a particular situation.
This approach is versatile and the number of possible combinations is growing
because SAP constantly keeps adding new components to their SAP
HANA-related portfolio.
53
IBM offers consulting services that help clients to choose the correct solution for
their business needs.
Current situation
Non-SAP
application
Data replication
Non-SAP
application
Non-SAP
application
SAP HANA
Non-SAP
application
Custom
database
Data replication
SAP HANA
SAP HANA
Figure 4-3 Examples of SAP HANA deployment options with regard to data acquisition
The initial situation is displayed schematically in the upper left of Figure 4-3. In
this example, a client-specific non-SAP application writes data to a custom
database that is slow and is not meeting client needs.
The other three examples in Figure 4-3 show that SAP HANA can be deployed in
such a scenario. These examples show that there is no single solution that is
best for every client, but that each situation must be considered independently.
54
Each of these three solutions has both advantages and disadvantages, which we
highlight to show aspects of a given solution that might need more detailed
consideration:
Replacing the existing database with SAP HANA
The advantage of this solution is that the overall architecture is not going to be
significantly changed. The solution remains simple without the need to
include additional components. Customers might also save on license costs
for the original database.
A disadvantage to this solution is that the custom application must be
adjusted to work with the SAP HANA database. If ODBC or JDBS is used for
database access, this is not a significant problem. Also, the whole setup must
be tested properly. Because the original database is being replaced, a certain
amount of downtime is inevitable.
Clients considering this approach must be familiar with the features and
characteristics of SAP HANA, especially when certain requirements must be
met by the database that is used (for example, special purpose databases).
Populating SAP HANA with data replicated from the existing database
The second option is to integrate SAP HANA as a side-car database to the
primary database and to replicate required data using one of the available
replication techniques.
An advantage of this approach is that the original solution is not touched and
no downtime is required. Also, only the required subset of data must be
replicated from the source database, which might allow customers to
minimize acquisition costs because SAP HANA acquisition costs are linked
directly to the volume of stored data.
The need for implementing replication technology can be seen as the only
disadvantage of this solution. Because data is delivered only into SAP HANA
through replication, this component is a vital part of the whole solution.
Customers considering this approach must be familiar with various replication
technologies, including their advantages and disadvantages, as described in
3.2, Data replication methods for SAP HANA on page 44.
Clients must also be aware that replication might cause additional load on the
existing database because modified records must be extracted and then
transported to the SAP HANA database. This aspect is highly dependent on
the specific situation and can be addressed by choosing the proper replication
technology.
Adding SAP HANA as a second database in parallel to the existing one
This third option keeps the existing database in place while adding SAP
HANA as a secondary database. The custom application then stores data in
both the original database and in the SAP HANA database.
55
56
Current situation
Custom
database
Possible scenario
Non-SAP
application
Custom
database
Non-SAP
application
SAP HANA
SAP analytic
tools
SAP BOBJ
reporting
Figure 4-4 An example of SAP HANA as a source for other applications
The initial situation is visualized schematically in the left part of Figure 4-4. A
customer-specific application runs queries against a custom database, which is a
function that we must preserve.
A potential solution is in the right part of Figure 4-4. A customer-specific
application runs problematic queries against the SAP HANA database. If the
existing database is still part of the solution, specific queries that do not need
acceleration can still be run against the original database.
Specialized analytic tools, such as SAP BusinessObjects Predictive Analysis,
can be used to run statistical analysis on data that is stored in the SAP HANA
database. This tool can run analysis directly inside the SAP HANA database,
which helps avoid expensive transfers of massive volumes of data between the
application and the database. The result of this analysis can be stored in SAP
HANA, and the custom application can use these results for further processing,
for example, to facilitate decision making.
SAP HANA can be easily integrated with products from the SAP
BusinessObjects family. Therefore, these products can be part of the solution,
responsible for reporting, monitoring critical key performance indicators (KPIs)
using dashboards, or for data analysis.
These tools can also be used without SAP HANA; however, SAP HANA is
enabling these tools to process much larger volumes of data and still provide
results in reasonable time.
57
58
Corporate BI
Database
Local BI
Data
Mart
SAP ERP 1
SAP ERP n
...
BI
Data
Mart
ETL
DB
Non-SAP
Business
Application
BI
Data
Mart
ETL
Database
DB
Database
Database
DB
With the huge amount of data that is collected in an enterprise data warehouse,
response times of queries for reports or navigation through data can become an
issue, generating new requirements for the performance of such an environment.
To address these requirements, SAP introduced the SAP NetWeaver Business
Warehouse Accelerator (SAP NetWeaver BW Accelerator), which is built for this
use case by speeding up queries and reports in the SAP NetWeaver BW by
using in-memory technology. Although being a perfect fit for an enterprise data
warehouse holding huge amounts of data, the combination of SAP NetWeaver
BW and SAP NetWeaver BW Accelerator is not always a viable solution for
relatively small data marts.
59
With the introduction of SAP HANA 1.0, SAP provided an in-memory technology
aiming to support Business Intelligence (BI) at a business unit level. SAP HANA
combined with business intelligence tools, such as the SAP BusinessObjects
tools and data replication mechanisms feeding data from the operational system
into SAP HANA in real time, brought in-memory computing to the business unit
level. Figure 4-6 shows such a landscape with the local data marts replaced by
SAP HANA.
Corporate BI
Accelerator
Local BI
SAP
HANA
1.0
SAP ERP 1
SAP ERP n
...
Sync
Database
SAP
HANA
1.0
Database
Non-SAP
Business
Application
Database
SAP
HANA
1.0
Figure 4-6 SAP vision after the introduction of SAP HANA 1.0
60
SAP
Business
Suite
Data Modeling
Column Store
repl.
RDBMS
Row Store
SAP
Reporting
and Analytics
SAP HANA
Usually, the first step is the replication of data into the SAP HANA database,
which usually originates from the SAP Business Suite. However, some solution
packages are also built for non-SAP data sources.
Sometimes, source systems must be adjusted by implementing modifications or
by performing specific configuration changes.
Data typically is replicated by using the SAP Landscape Transformation
replication; however, other options, such as replication by using SAP
BusinessObjects Data Services or SAP HANA Direct Extractor Connection
(DXC), are also possible. The replication technology usually is chosen as part of
the package design and cannot be changed easily during implementation.
A list of tables to replicate (for SAP Landscape Transformation replication) or
transformation models (for replication using Data Services) are part of the
package.
SAP HANA is loaded with models (views) that are either static (designed by SAP
and packaged) or automatically generated based on customized criteria. These
models describe the transformation of source data into the resulting column
views. These views are then consumed by SAP BusinessObjects BI 4.0 reports
or dashboards that are either delivered as final products or pre-made templates
that can be finished as part of implementation process.
Some solution packages are based on additional components (for example, SAP
BusinessObjects Event Insight). If required, additional content that is specific to
these components can also be part of the solution package.
Individual use cases, required software components, prerequisites, and
configuration changes (including overall implementation processes) are properly
documented and attached as part of the delivery.
61
62
SAP HANA can also be used to accelerate existing processes in SAP Business
Suite systems, even for those systems that are not yet released to be running
directly on the SAP HANA database.
Some SAP systems are processing large amounts of records that must be
filtered or aggregated based on specific criteria. Results are then used as inputs
for all dependent activities in a given system.
In the case of large data volumes, the running time can be unacceptable (in
number of hours). Such workloads can easily run several hours, which can cause
unnecessary delays. Currently, these tasks typically are being processed
overnight as batch jobs.
SAP HANA as an accelerator can help decrease this running time.
Figure 4-8 illustrates this use case scenario.
SAP UI
SAP
Business
Suite
read
Data Modeling
repl.
RDBMS
SAP
Reporting
and Analytics
Column Store
Row Store
SAP HANA
The accelerated SAP system must meet specific prerequisites. Before this
solution can be implemented, installation of specific support packages or
implementation of SAP Notes might be required, which introduces the necessary
code changes in the source system.
The SAP HANA client must be installed on a given server, and the SAP kernel
must be adjusted to support direct connectivity to the SAP HANA database.
63
As a next step, replication of data from the source system is configured. Each
specific use case has a defined replication method and a list of tables that must
be replicated. The most common method is the SAP Landscape Transformation
replication. However, some solutions offer alternatives. For example, for the SAP
CO-PA Accelerator, replication can also be performed by an SAP CO-PA
Accelerator-specific ABAP report in the source system.
The source system is configured to have direct connectivity into SAP HANA as
the secondary database. The required scenario is configured according to the
specifications and then activated. During activation, the source system
automatically deploys the required column views into SAP HANA and activates
new ABAP code that was installed in the source system as the solution
prerequisite. This new code can run, consuming queries against the SAP HANA
database, which leads to shorter execution times.
Because SAP HANA is populated with valuable data, it is easy to extend the
accelerator use case by adding operational reporting functions. Additional
(usually optional) content is delivered for SAP HANA and for SAP
BusinessObjects BI 4.0 client tools, such as reports or dashboards.
SAP HANA as the accelerator and SAP HANA for operational reporting use case
scenarios can be combined in to a single package. Here is a list of SAP RDSs
implementing SAP HANA as an accelerator:
SAP Bank Analyzer Rapid-Deployment Solution for Financial Reporting with
SAP HANA (see SAP Note 1626729):
http://service.sap.com/rds-hana-finrep
SAP rapid-deployment solution for customer segmentation with SAP HANA
(see SAP Note 1637115):
http://service.sap.com/rds-cust-seg
SAP ERP rapid-deployment solution for profitability analysis with SAP HANA
(see SAP Note 1632506):
http://service.sap.com/rds-hana-copa
SAP ERP rapid-deployment solution for accelerated finance and controlling
with SAP HANA (see SAP Note 1656499):
http://service.sap.com/rds-hana-fin
SAP Global Trade Services rapid-deployment solution for sanctioned-party
list screening with SAP HANA (see SAP Note 1689708):
http://service.sap.com/rds-gts
64
SAP
Business
Suite
RDBMS
SAP BW
SAP ECC
Column Store
Column Store
Row Store
Row Store
SAP HANA
SAP HANA
Figure 4-9 SAP products running on SAP HANA - SAP NetWeaver BW and SAP ECC
65
Corporate BI
Virtual
Data Mart
Virtual
Data Mart
Virtual
Data Mart
Local BI
Local BI
SAP ERP 1
SAP ERP n
...
SAP
HANA
Database
Non-SAP
Business
Application
SAP
HANA
Database
Database
66
Company Code
Master Data Table:
/BI0/PCOMP_CODE
DIMID
SID_0CHNGID
SID_0RECORDTP
Fact Table:
/BI0/F0COPC_C08
SID_0REQUID
KEY_0COPC_C08P
KEY_0COPC_C08T
KEY_0COPC_C08U
Enterprise Structure
Dimension Table:
/BI0/D0COPC_C081
KEY_0COPC_C081
DIMID
KEY_0COPC_C082
SID_0COMP_CODE
KEY_0COPC_C083
SID_0PLANT
KEY_0COPC_C084
KEY_0COPC_C085
AMOUNTFX
Material
Dimension Table:
/BI0/D0COPC_C082
COMP_CODE
Company Code
SID Table:
/BI0/SCOMP_CODE
OBJVERS
CHANGED
COMP_CODE
CHRT_ACCTS
SID
COMPANY
CHCKFL
COUNTRY
DATAFL
...
INCFL
Plant
SID Table:
/BI0/SPLANT
Plant
Master Data Table :
/BI0/PPLANT
PLANT
PLANT
OBJVERS
SID
CHANGED
AMOUNTVR
DIMID
CHCKFL
ALTITUDE
PRDPLN_QTY
SID_0MATERIAL
DATAFL
BPARTNER
LOTSIZE_CM
SID_0MAT_PLANT
INCFL
...
The core part of every InfoCube is the fact table. This table contains dimension
identifiers (IDs) and corresponding key figures (measures). This table is
surrounded by dimension tables that are linked to fact tables using the dimension
IDs.
67
Dimension tables are small tables that group logically connected combinations of
characteristics, usually representing master data. Logically connected means
that the characteristics are highly related to each other, for example, company
code and plant. Combining unrelated characteristics leads to many possible
combinations, which can have a negative impact on the performance.
Because master data records are in separate tables outside of the InfoCube, an
additional table is required to connect these master data records to dimensions.
These additional tables contain a mapping of auto-generated Surrogate IDs
(SIDs) to the real master data.
This complex structure is required on classical databases; however, with SAP
HANA, this requirement is obsolete. So, SAP introduced the SAP HANA
Optimized Star Schema, which is illustrated in Figure 4-12.
Fact Table:
/BI0/F0COPC_C08
Data Package
Dimension Table:
/BI0/D0COPC_C08P
KEY_0COPC_C08P
DIMID
SID_0CALDAY
SID_0CHNGID
SID_0FISCPER
SID_0RECORDTP
SID_0FISCVARNT
SID_0REQUID
Company Code
Master Data Table:
/BI0/PCOMP_CODE
COMP_CODE
Company Code
SID Table:
/BI0/SCOMP_CODE
CHANGED
COMP_CODE
CHRT_ACCTS
SID_0CURRENCY
SID
COMPANY
SID_0UNIT
CHCKFL
COUNTRY
SID_0COMP_CODE
DATAFL
SID_0PLANT
INCFL
SID_0MATERIAL
SID_0MAT_PLANT
SID_0CURTYPE
Plant
SID Table:
/BI0/SPLANT
Plant
Master Data Table :
/BI0/PPLANT
PLANT
...
PLANT
OBJVERS
AMOUNTFX
SID
CHANGED
AMOUNTVR
CHCKFL
ALTITUDE
PRDPLN_QTY
DATAFL
BPARTNER
LOTSIZE_CM
INCFL
Figure 4-12 SAP HANA Optimized Star Schema in an SAP NetWeaver BW system
68
OBJVERS
The content of all dimensions (except for the Data Package dimension) is
incorporated into the fact table. This modification brings several advantages:
Simplified modeling
Poorly designed dimensions (wrong combinations of characteristics) cannot
affect performance anymore. Moving characteristics from one dimension to
another is not a physical operation anymore; instead, it is just a metadata
update.
Faster loading
Because dimension tables do not exist, all overhead workload that is related
to identification of existing combinations or creating combinations in the
dimension tables is not required anymore. Instead, the required SID values
are inserted directly into the fact table.
The SAP HANA Optimized Star Schema is used automatically for all newly
created InfoCubes on the SAP NetWeaver BW system running on the SAP
HANA database.
Existing InfoCubes are not automatically converted to this new schema during
the SAP HANA conversion of the SAP NetWeaver BW system. The conversion of
standard InfoCubes to in-memory optimized InfoCubes must be done manually
as a follow-up task after the database migration.
69
SAP HANA can calculate all aggregations in real time. Therefore, aggregates are
no longer required, and roll-up activity that is related to aggregate updates is
obsolete. This also reduces the overall run time of update operations.
If SAP NetWeaver BW Accelerator is used, the update of its indexes is also no
longer needed. Because SAP HANA is based on technology similar to SAP
NetWeaver BW Accelerator, all queries are accelerated. Query performance with
SAP HANA can be compared to situations where all cubes are indexed by the
SAP NetWeaver BW Accelerator. In reality, query performance can be even
faster than with SAP NetWeaver BW Accelerator because additional features are
available for SAP NetWeaver BW running on SAP HANA, for example, the
possibility to remove an InfoCube and instead run reports against in-memory
optimized DataStore Objects (DSOs).
70
Not all SAP NetWeaver BW add-ons are supported to run on the SAP
HANA-based system. For the latest information, see following SAP Notes:
Note 1600929 - SAP NetWeaver BW powered by SAP HANA DB: Information
Note 1532805 - Add-On Compatibility of SAP NetWeaver 7.3
Note 1652738 - Add-on compatibility for SAP NetWeaver EHP 1 for NW 7.30
Unless your system already meets the minimal release requirements, the first
step before converting SAP NetWeaver BW is to upgrade the system to the latest
available release and to the latest available support package level.
A database upgrade might be required as part of the release upgrade or as a
prerequisite before database migration to SAP HANA. For a list of supported
databases, see SAP Note 1600929.
Table 4-1 lists the databases that were approved as source databases for the
migration to SAP HANA at the time of writing.
Table 4-1 Supported source databases for a migration to the SAP HANA database
Database
SAP NetWeaver
BW 7.30
SAP NetWeaver
BW 7.31
SAP NetWeaver
BW 7.40
Oracle
11.2
11.2
11.2
MaxDB
7.8
7.9
7.9
MS SQL server
2008
2008
2008, 2012
9.7
9.7
10.1
6.1, 7.1
6.1, 7.1
7.1
9, 10
9, 10
10
SybaseASE
N/A
15.7
15.7, 16
SAP HANA is not a supported database for any SAP NetWeaver Java stack.
Therefore, dual-stack installations (ABAP plus Java) must be separated into two
individual stacks by using the Dual-Stack Split Tool from SAP.
Because some existing installations are still non-Unicode installations, another
important prerequisite step might be a conversion of the database to Unicode
encoding. This Unicode conversion can be done as a separate step or as part of
the conversion to the SAP HANA database.
71
All InfoCubes with data persistency in the SAP NetWeaver BW Accelerator are
set as inactive during conversion, and their content in SAP NetWeaver BW
Accelerator is deleted. These InfoCubes must be reloaded again from the
original primary persistence; therefore, required steps must be incorporated into
the project plan.
A migration to the SAP HANA database follows the same process as any other
database migration. All activity in the SAP NetWeaver BW system is suspended
after all preparation activities are finished. A special report is run to generate
database-specific statements for the target database that is used during import.
Next, the content of the SAP system is exported to a platform-independent
format and stored in files on disk.
These files are then transported to the primary application server of the SAP
NetWeaver BW system. The application part of SAP NetWeaver BW is not
allowed to run on the SAP HANA appliance. Therefore, a minimal installation
must have two servers:
SAP HANA appliance hosting the SAP HANA database
The SAP HANA appliance is delivered by IBM with the SAP HANA database
preinstalled. However, the database is empty.
Primary application server hosting ABAP instance of SAP NetWeaver BW
There are minimal restrictions regarding the operating system of the primary
application server. For available combinations, see the Product Availability
Matrix (PAM) (search for SAP NetWeaver 7.3 and download the overview
presentation), found at the following website:
http://service.sap.com/pam
At the time of writing, the operating systems that are shown in Table 4-2 on
page 73 are available to host the ABAP part of the SAP NetWeaver BW
system.
72
HP-UX 11.31
IA64 (64-bit)
Solaris 10
SPARC (64-bit)
Solaris 10
x86_64 (64-bit)
Linux SLES 11
x86_64 (64-bit)
Linux RHEL 5
x86_64 (64-bit)
Linux RHEL 6
x86_64 (64-bit)
IBM i 7.1
Power (64-bit)
SAP
NetWeaver
BW 7.30
No
Yes
Yes
Yes
Yes
Yes
Yes
Yes
Yes
No
SAP
NetWeaver
BW 7.31
No
Yes
Yes
Yes
Yes
Yes
Yes
No
Yes
Yes
SAP
NetWeaver
BW 7.40
Yes
Yes
Yes
Yes
Yes
Yes
Yes
No
Yes
Yes
The next step is the database import. It contains the installation of the SAP
NetWeaver BW on the primary application server and the import of data into the
SAP HANA database. The import occurs remotely from the primary application
server as part of the installation process.
Parallel export/import by using socket connection and File Transfer Protocol
(FTP) and Network File System (NFS) exchange modes are not supported.
Currently, only the asynchronous file-based export/import method is available.
After mandatory post-activities, conversion of InfoCubes and DataStore objects
to their in-memory optimized form must be initiated to use all the benefits that the
SAP HANA database can offer. This task can be done either manually for each
object or as a mass operation by using a special report.
Clients must plan enough time to perform this conversion. This step can be
time-consuming because the content of all InfoCubes must be copied into
temporary tables that have the new structure.
After all post-activities are finished, the system is ready to be tested.
73
The possibility to transport content across different releases can reduce the
amount of effort that is required to build a proof of concept (POC) system
because most of the prerequisite activities, such as the release upgrade,
database upgrade, and dual-stack split, are not needed.
After transporting the available objects (metadata definitions), their content must
also be transported from the source to the target system. The SAP NetWeaver
BW consultant must assess which available options are most suitable for this
purpose.
This approach is not recommended for production migration where a
conventional database migration is used. Therefore, additional effort that is
invested in building a POC system in the same way as the production system is
treated, is a valuable test. This test can help customers create a realistic effort
estimation for the project, estimate required runtimes, and define detailed
planning of all actions that are required. All involved project team members
become familiar with the system and can solve and document all specific
problems.
74
75
The Product Lifecycle Management (PLM) application is not available with SAP
HANA, but there is a plan to make these available to run with SAP HANA.
SAP Business Suite, which is powered by SAP HANA, does not induce any
functional changes. Configuration, customization, the ABAP Workbench,
connectivity, security, transports, and monitoring stay unchanged. For
modifications, the same upgrade requirements as with any other upgrade apply.
Customized code can stay unchanged, or can be adjusted to use additional
performance.
SAP Business Suite applications can benefit in various ways from using the
in-memory technology of SAP HANA:
Running dialog processes instead of batch
Integration of unstructured data and machine to machine data (M2M) with
ERP processes
Integration of predictive analysis with ERP processes
Running operational reports in real time, directly on the source data
Removing the need for operational data stores
Eliminating the need for data replication or transfers to improve operational
report performance
76
http://www.news-sap.com/sap-business-suite-on-sap-hana-launch
SAP is enabling the existing functions in SAP applications to use the in-memory
technology with the following versions:
SAP enhancement package 6 for SAP ERP 6.0, version for SAP HANA
SAP enhancement package 2 for SAP CRM 7.0, version for SAP HANA
SAP enhancement package 2 for SAP SCM 7.0, version for SAP HANA
SAP enhancement package 2 for SAP SRM 7.0, version for SAP HANA
Restrictions
There are certain restrictions4 in effect regarding running SAP Business Suite
with SAP HANA.
Currently, multi-node support for SAP Business Suite with SAP HANA is limited.5
SAP HANA multi-node configurations can serve different purposes:
Achieving high-availability by the use of standby nodes
Scaling the main memory to accommodate larger databases (scale-out)
Scale-out scenarios with multiple worker nodes (as described in 5.3.3,
Scaling-out SAP HANA using GPFS on page 128) are not yet supported for
SAP Business Suite with SAP HANA.
High availability (HA) scenarios for SAP Business Suite with SAP HANA are
supported, but restricted to the simplest case of two servers, one being the
worker node and one acting as a standby node. In this case, the database is not
partitioned, but the entire database is on a single node. This is why this
configuration is sometimes also referred to as a single-node HA configuration.
Because of these restrictions with regards to scalability, SAP decided to allow
configurations with a higher memory per core ratio, specifically for this use case.
Sections 6.1.2, Single-node eX5 solution for SAP Business Suite on HANA on
page 151 and 6.2.2, Single-node X6 solution for SAP Business Suite on HANA
on page 165 describe available configurations that are dedicated to SAP
Business Suite, which is powered by SAP HANA.
77
A new software component can be integrated with SAP HANA either directly or it
can be built on top of the SAP NetWeaver stack, which can work with the SAP
HANA database using client libraries.
Because of its breadth and depth, this use case scenario is not described in
detail as part of this publication.
78
Chapter 5.
79
80
The x3850 X5 and the workload-optimized x3950 X5 are the logical successors
to the x3850 M2 and x3950 M2, featuring the IBM eX4 chipset. Compared with
previous generation servers, the x3850 X5 offers the following features:
High memory capacity
Up to 64 dual inline memory modules (DIMMs) standard and 96 DIMMs with
the MAX5 memory expansion per four-socket server
Intel Xeon processor E7 family
Exceptional scalable performance with advanced reliability for your most
data-demanding applications
Extended SAS capacity with eight HDDs and 900 GB 2.5-inch SAS drives or
1.6 TB of hot-swappable Redundant Array of Independent Disks 5 (RAID 5)
with eXFlash technology
Standard dual-port Emulex 10 GB Virtual Fabric adapter
Ten-core, 8-core, and 6-core processor options with up to 2.4 GHz (10-core),
2.13 GHz (8-core), and 1.86 GHz (6-core) speeds with up to 30 MB L3 cache
Scalable to a two-node system with eight processor sockets and 128 dual
inline memory module (DIMM) sockets
Seven PCIe x8 high-performance I/O expansion slots to support hot-swap
capabilities
Optional embedded hypervisor
81
The x3850 X5 and x3950 X5 both scale to four processors and 2 TB of RAM.
With the MAX5 attached, the system can scale to four processors and 3 TB of
RAM. Two x3850 X5 servers can be connected together for a single system
image with eight processors and 4 TB of RAM.
5.1.2 x3690 X5
The x3690 X5 (Figure 5-2) is a 2U rack-optimized server that brings new features
and performance to the mid tier.
This machine is a two-socket, scalable system that offers up to four times the
memory capacity of current two-socket servers. It supports the following
specifications:
Up to two sockets for Intel Xeon E7 processors. Depending on the processor
model, processors have six, eight, or ten cores.
Scalable 32 - 64 DIMM sockets with the addition of an MAX5 memory
expansion unit.
Advanced networking capabilities with a Broadcom 5709 dual Gb Ethernet
controller standard in all models and an Emulex 10 Gb dual-port Ethernet
adapter standard on some models, optional on all others.
Up to 16 hot-swap 2.5-inch SAS HDDs, up to 16 TB of maximum internal
storage with RAID 0, 1, or 10 to maximize throughput and ease installation.
RAID 5 is optional. The system comes standard with one HDD backplane that
can hold four drives. Second and third backplanes are optional for an
additional 12 drives.
New eXFlash high-input/output operations per second (IOPS) solid-state
device (SSD) storage technology.
82
83
84
Normal
operation
Hardware
OS
Error
detected
Error
passed to
OS
Hardware
correctable
error
Error
corrected
Memory
Page
unused
Memory Page
unmapped
and marked
SAP HANA
Application
signaled
Application
terminates
Page identified
and data can be
reconstructed
Reconstruct
data in
corrupted page
85
5.1.4 Memory
For an in-memory appliance, such as SAP HANA, a systems main memory, its
capacity, and its performance play an important role. The Intel Xeon processor
E7 family, which is shown in Figure 5-4, has a memory architecture that is suited
to the requirements of such an appliance.
The E7 processors have two SMIs. Therefore, memory must be installed in
matched pairs. For better performance, or for systems that are connected
together, memory must be installed in sets of four. The memory that is used in
the eX5 systems is DDR3 SDRAM registered DIMMs. All of the memory runs at
1066 MHz or less, depending on the processor.
Processor
Memory
controller
Buffer
Buffer
Memory
controller
Buffer
Buffer
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
86
Figure 5-5 shows eight possible memory configurations for the two memory
cards and 16 DIMMs connected to each processor socket in an x3850 X5.
Similar configurations apply to the x3690 X5 and HX5. Each configuration has a
relative performance score.
1
2
3
4
5
6
7
8
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 1
Mem Ctrl 2
Relative
performance
Each processor:
2 memory controllers
2 DIMMs per channel
8 DIMMs per MC
1.0
Each processor:
2 memory controllers
1 DIMM per channel
4 DIMMs per MC
0.94
0.61
Mem Ctrl 2
Each processor:
2 memory controllers
2 DIMMs per channel
4 DIMMs per MC
0.58
Mem Ctrl 2
Each processor:
2 memory controllers
1 DIMM per channel
2 DIMMs per MC
Mem Ctrl 2
Each processor:
1 memory controller
2 DIMMs per channel
8 DIMMs per MC
0.51
Mem Ctrl 2
Each processor:
1 memory controller
1 DIMM per channel
4 DIMMs per MC
Mem Ctrl 2
Each processor:
1 memory controller
2 DIMMs per channel
4 DIMMs per MC
Mem Ctrl 2
Mem Ctrl 2
Each processor:
1 memory controller
1 DIMM per channel
2 DIMMs per MC
Memory card
DIMMs
Channel
Memory buffer
SMI link
Memory controller
Mem Ctrl 1
0.47
0.31
Memory configurations
1
0.94
0.9
0.8
0.7
0.61
0.6
0.58
0.51
0.5
0.47
0.4
0.31
0.29
0.3
0.2
0.1
0
1
Configuration
0.29
Figure 5-5 Relative memory performance based on DIMM placement (one processor and two memory
cards shown)
87
Hemisphere mode
Hemisphere mode is an important performance optimization of the Intel Xeon
processor E7, 6500, and 7500 product families. Hemisphere mode is
automatically enabled by the system if the memory configuration allows it. This
mode interleaves memory requests between the two memory controllers within
each processor, enabling reduced latency and increased throughput. It also
allows the processor to optimize its internal buffers to maximize memory
throughput.
Hemisphere mode is enabled only when the memory configuration behind each
memory controller on a processor is identical. In addition, because eight DIMMs
per processor are required for using all memory channels, eight DIMMs per
processor must be installed at a time for optimized memory performance.
88
89
The x3950 X5-based models of the IBM Systems Solution for SAP HANA come
with IBM High IOPS adapters with 1.2 TB storage capacity (7143-HAx, -HBx, and
-HCx).
Figure 5-6 shows the IBM 1.2 TB High IOPS MLC Mono adapter.
eXFlash
IBM eXFlash is the name of the eight 1.8-inch SSDs, the backplanes, SSD
hot-swap carriers, and indicator lights that are available for the x3690 X5. Each
eXFlash can be put in place of four SAS or SATA disks. The eXFlash units
connect to the same types of ServeRAID disk controllers as the SAS/SATA disks.
Figure 5-7 shows an eXFlash unit, with the status light assembly on the left side.
Status lights
Solid-state drives
(SSDs)
90
In addition to using less power than rotating magnetic media, the SSDs are more
reliable and can service many more IOPS. These attributes make them suited to
I/O-intensive applications, such as transaction processing, logging, backup and
recovery, and Business Intelligence (BI). Built on enterprise-grade MLC NAND
flash memory, the SSD drives that are used in eXFlash for SAP HANA deliver up
to 60,000 read IOPS per single drive. Combined into an eXFlash unit, these
drives potentially can deliver up to 480,000 read IOPS and up to 4 GBps of
sustained read throughput per eXFlash unit.
In addition to its superior performance, eXFlash offers superior uptime with three
times the reliability of mechanical disk drives. SSDs have no moving parts to fail.
Each drive has its own backup power circuitry, error correction, data protection,
and thermal monitoring circuitry. They use Enterprise Wear-Leveling to extend
their use even longer.
A single eXFlash unit accommodates up to eight hot-swap SSDs and can be
connected to up to two performance-optimized controllers. The x3690 X5-based
models for SAP HANA enable RAID protection for the SSD drives by using two
ServeRAID M5015 controllers with the ServeRAID M5000 Performance
Accelerator Key for the eXFlash units.
91
Storage
Each SAP HANA model must provide a certain amount of traditional storage
capacity so that savepoints of the in-memory database can be written out to
persistent storage in regular intervals. In addition, a log entry must be stored for
every change in the database.
All systems come with storage for both the data volume and the log volume, as
shown in Figure 5-8 on page 93. Savepoints are stored on a RAID protected
array of 10 K SAS hard disk drives (HDDs), and optimized for data throughput.
Savepoints consist of a consistent backup of all database tables that are kept in
main memory. SAP HANA database logs are stored on flash-based High IOPS
PCIe adapters.
92
Time
Data savepoint
to persistent
storage
SAS Drives
Log written
to persistent storage
(committed transactions)
HighIOPS SSD
optimized for
optimized for
throughput
Server
local
storage
This flash technology storage device is optimized for high IOPS performance and
low latency to provide the SAP HANA database with a log storage that allows the
highest possible performance. Because a transaction in the SAP HANA
database can return only after the corresponding log entry is written to the log
storage, overall database performance is limited to how quickly those log entries
can be persisted.
93
Figure 5-9 shows the two storage devices, HDD and flash, together with the SAP
HANA software stack for a single node that is based on the x3950 X5 model.
Both the savepoints (data01) and the logs (log01) are stored once, denoted as
the first replica in the figure.1
x3950 X5
SAP HANA DB
DB partition 1
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
Shared file system - GPFS
First replica
HDD
Flash
data01
log01
Figure 5-9 Storage architecture of a x3950 workload-optimized solution for SAP HANA
(SAP HANA DB mentioned twice to be consistent with scale-out figures later on in this
chapter)
The HDD storage in Figure 5-9 consists of the local drives that are internal to the
x3950 server, and is enhanced with additional drives in a locally attached SAS
enclosure for bigger T-shirt sizes that require more storage capacity. The amount
of flash memory can be enhanced with additional HighIOPS adapters that are
installed in the server.
Despite GPFS being a cluster file system, single node IBM SAP HANA solutions
also can use it. From a GPFS point of view, single node configuration is a specific
form of a cluster, consisting of just one node. This single-node solution does not
use the cluster features of the GPFS, but takes advantages of GPFS by using
different types of storage on eX5 systems in one file system and placing files
onto proper storage media according to the IO type (that is, either data or log).
1
94
Previous editions of this book used the term primary data for the first replica. To be consistent with
official documentation and to emphasize that there is no difference between multiple replicas of a
single file, this edition uses the term first replica (and second and third replica later on in the book).
Network
The standard x3950 X5 building blocks are used to build single node SAP HANA
database environments and scale-out, or clustered, environments.
In scale-out installations, the participating nodes are interconnected by 10 Gb
Ethernet in a redundant fashion. There are two redundant 10 Gb Ethernet
networks for the communication within the solution:
A fully redundant 10 Gb Ethernet network for cluster-internal communication
of the SAP HANA software
A fully redundant 10 Gb Ethernet network for cluster-internal communication
of GPFS, including replication
These networks are internal to the scale-out solution and have no connection to
the client network. IBM uses Emulex 10 Gigabit Ethernet adapters for both
internal networks. Those adapters and the networking switches for these
networks are part of the appliance and cannot be substituted with other than the
validated switch models.
Uplinks to the client network are handled through either 1 Gigabit or 10 Gigabit
Ethernet connection depending on the existing client infrastructure.
95
Figure 5-10 illustrates the networking architecture for a scale-out solution and
shows the SAP HANA scale-out solution that is connected to an SAP NetWeaver
Business Warehouse (SAP NetWeaver BW) system as an example. Port
numbers reference the physical ports in Figure 5-11 on page 97.
SAP
BW
system
switch
switch
10 GbE switch
node 1
node 2
node 3
...
node n
10 GbE switch
Figure 5-10 Networking architecture of the x3950 X5 workload-optimized solution for SAP HANA
All network connections within the scale-out solution are fully redundant. Both
the internal GPFS network and the internal SAP HANA network are connected to
two 10 Gb Ethernet switches, which are interconnected for full redundancy. The
switch model that is used here is the IBM System Networking RackSwitch
G8264. It delivers exceptional performance, being both lossless and low latency.
With 1.2 Tbps throughput, the G8264 provides massive scalability and low
latency that is ideal for latency-sensitive applications, such as SAP HANA.
The scale-out solution for SAP HANA makes intensive use of the advanced
capabilities of this switch, such as virtual link aggregation groups (vLAGs). For
smaller scale-out deployments, the smaller IBM Systems Networking
RackSwitch G8124 can be used instead of the G8264. For details about the
switch options, see 5.4, IBM System Networking options on page 136.
96
To illustrate the network connectivity, Figure 5-11 shows the back of an x3950 X5
building block with the network interfaces available. The letters denoting the
interfaces correspond to the letters that are used in Figure 5-10 on page 96.
GPFS
IMM
SAP HANA
Figure 5-11 The back of an x3950 X5 building block with the network interfaces available
Each eX5 building block comes with four 10 Gb Ethernet and six 1 GbE network
ports. The available 1 Gb Ethernet interfaces available (a.b.e.f.g.h) on the system
can be used to connect the systems to other networks or systems, for example,
for client access, application management, systems management, and data
management. The interface that is denoted with the letter i is used to connect
the integrated management module (IMM) of the server to the management
network.
Integrated virtualization
The VMware ESXi embedded hypervisor software is a virtualization platform that
allows multiple operating systems to run on a host system concurrently. Its
compact design allows it to be embedded in physical servers.
97
IBM offers versions of VMware vSphere Hypervisor (ESXi) that are customized
for select IBM hardware to give you online platform management, including
updating and configuring firmware, platform diagnostic tests, and enhanced
hardware alerts. All models support several USB keys as options, which are
listed in Table 5-1.
Table 5-1 VMware ESXi memory keys
Part number
Feature code
Description
41Y8298
A2G0
41Y8300
A2VC
41Y8307
A383
41Y8311
A2R3
41Y8385
A584
The x3850 X5 and x3950 X5 have two internal USB connectors that are available
for the embedded hypervisor USB key. The location of these USB connectors is
shown in Figure 5-12.
Embedded
hypervisor key
installed
Figure 5-12 Location of internal USB ports for embedded hypervisor on the x3850 X5
and x3950 X5
98
For more information about the USB keys, and to download the IBM customized
version of VMware ESXi, go to the following website:
http://www.ibm.com/systems/x/os/vmware/esxi
In addition to the USB key, you also must license vSphere with VMware.
Depending on the physical server, you need different licenses. For more
information, see 6.4, SAP HANA on VMware vSphere on page 177. That
section also contains more information about virtualizing an SAP HANA
instance.
99
Storage
All systems come with storage for both the data volume and the log volume. The
building blocks that are based on the x3690 X5 come with combined data and log
storage on an array of RAID-protected, hot-swap eXFlash SSD drives.
These flash technology storage devices are optimized for high IOPS
performance and low latency to provide the SAP HANA database with a log
storage that allows the highest possible performance. Because a transaction in
the SAP HANA database can return only after the corresponding log entry is
written to the log storage, high IOPS performance and low latency are key to
database performance.
Figure 5-13 shows the storage architecture of workload-optimized solution that is
based on the x3690 X5 model.
x3690 X5
SAP HANA DB
DB partition 1
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
Shared file system - GPFS
SSD
First replica
SSD
data + log
Figure 5-13 Storage architecture of the x3690 X5 workload-optimized solution for SAP
HANA
IBM uses General Parallel File System (GPFS) to provide a robust and
high-performance file system that allows you to grow an SAP HANA environment
without needing to reformat the storage arrays. GPFS also takes all available
storage arrays and combines them into one file system for use with SAP HANA.
Additional hot spare SSDs are supported. They can be added without impacting
the performance of the overall solution.
100
Network
The standard x3690 X5 building blocks are used to build single node SAP HANA
database environments and scale-out, or clustered, environments. In scale-out
installations, the participating nodes are interconnected by 10 Gb Ethernet in a
redundant fashion.
The networking architecture for scale-out SAP HANA environments that are
based on the x3690 X5 follows the same approach as the x3950 X5 solution.
There are two redundant 10 Gb Ethernet networks for the communication within
the solution:
A fully redundant 10 Gb Ethernet network for cluster-internal communication
of the SAP HANA software
A fully redundant 10 Gb Ethernet network for cluster-internal communication
of GPFS, including replication
These networks are internal to the scale-out solution and have no connection to
the client network. IBM uses Emulex 10 Gigabit Ethernet adapters for both
internal networks. Those adapters and the networking switches for these
networks are part of the appliance and cannot be substituted with other than the
validated switch models.
Uplinks to the client network are handled through either a 1 Gigabit or 10 Gigabit
Ethernet connection, depending on the existing client infrastructure.
101
The x3950 X6 looks like two x3850 X6 servers where one is placed on top of the
other. However, unlike eX5 servers, x3950 X6 employs single chassis with a
single midplane design without any external connectors and cables.
Figure 5-15 shows the x3950 X6.
102
The X6 systems offer a new bookshelf design concept that is based on a fixed
chassis that is mounted in a standard rack cabinet. There is no need to pull the
chassis in or out of the rack to access components because all components can
be accessed either from the front or from the rear like pulling books from a
bookshelf.
Figure 5-16 shows the x3850 X6 server with one of the four Compute Books
partially removed.
Figure 5-16 IBM x3850 X6 server with a Compute Book partially removed
The modular component that can be installed in a chassis is called a Book. There
are several types of books available:
Compute Books
A Compute Book contains one processor, 24 DIMM slots, and 2 hot-swap fan
modules. It is accessible from the front of the server.
The x3850 X6 supports up to four Compute Books. The x3950 X6 supports
up to eight Compute Books.
103
Storage Books
The Storage Book contains standard 2.5-inch drives or IBM eXFlash 1.8-inch
hot-swap SSD drives. It also provides front USB and video ports, and it has
two PCIe slots that are reserved for internal storage adapters. The Storage
Book is accessible from the front of the server.
The x3850 X6 has one Storage Book. The x3950 X6 has two Storage Books.
I/O Books
I/O Book is a container that provides PCIe expansion capabilities. I/O Books
are accessible from the rear of the server.
There are three types of I/O Books:
Primary I/O Book. This book provides core I/O connectivity, including the
ML2 unique slot for an onboard network, three standard PCIe 3.0 slots,
Integrated Management Module II, hot-swap fan modules and USB, video,
serial, and systems management ports.
Full-length I/O Book. This hot-swap Book provides three optional
full-length PCIe slots.
Half-length I/O Book. This hot-swap Book provides three optional
half-length PCIe slots.
The x3850 X6 has one Primary I/O Book and supports one or two of the full or
half-length I/O Books (one of each or two of either). The x3950 X6 has two
Primary I/O Books and supports up to four of the full or half-length I/O Books
(any combination).
You can find more information about X6 servers and the sixth generation of IBM
Enterprise X-Architecture technology in IBM X6 Servers: Technical Overview,
REDP-5059.
The next sections introduce the technology components that are used to build
the IBM X6-based workload-optimized solution for SAP HANA and explain the
architecture of x3850 X6 and x3950 X6-based solutions. They share a common
concept so that you can start with a x3850 X6-based installation and later on
upgrade to an x3950 X6 installation without leaving parts on the floor.
104
The Intel Xeon processor E7 v2 family CPUs are the latest Intel scalable
processors and can be used to scale up to four processors in the x3850 X6 or up
to eight processors in the x3950 X6.
The current models of the X6 systems use processors from the Intel Xeon
processor E7 v2 product family. The Intel Xeon processors that are used in the
X6 systems are follow-ons to the Intel Xeon processor E7 product family. New
processors feature the new Intel microarchitecture (formerly codenamed
IvyBridge-EX) and a new 22 nm manufacturing process that provides higher
core count, larger cache sizes, higher core frequencies, and higher memory
speeds. In addition, these new processors support more memory with up to
24 DIMMs per processor and faster low-latency I/O with integrated PCIe 3.0
controllers.
The Intel Xeon processor E7 v2 product family offers the following key features:
Up to 15 cores and 30 threads (using Hyper-Threading feature) per processor
Up to 37.5 MB of L3 cache
Up to 3.4 GHz core frequencies
Up to 8 GTps bandwidth of QPI links
Integrated memory controller with four SMI2 channels that support up to
24 DDR3 DIMMs
Up to 1600 MHz DDR3 memory speeds and up to 2667 MHz SMI link speeds
Integrated PCIe 3.0 controller with 32 lanes per processor
Intel Virtualization Technology (VT-x and VT-d)
Intel Turbo Boost Technology 2.0
Intel Advanced Vector Extensions (AVT)
Intel AES-NI instructions for accelerating of encryption
Advanced QPI and memory reliability, availability, and serviceability (RAS)
features
Machine Check Architecture recovery (non-execution and execution paths)
Enhanced Machine Check Architecture Gen1
Machine Check Architecture I/O
Security technologies: OS Guard, Secure Key, and Intel TXT
105
The Intel Xeon E7 v2 processors have many new features that improve
performance of SAP HANA workloads. A white paper called Infuse your
business with real-time, data-driven intelligence shows the benefits of the new
generation Xeon processors for SAP HANA. You can find this white paper at the
Intel website:
http://www.intel.com/content/www/us/en/big-data/big-data-xeon-e7-sap-ha
na-real-time-business-platform-brief.html
106
PCIe 3.0 uses the 128b/130b encoding scheme, which is more efficient than the
8b/10b encoding that is used in the PCIe 2.0 protocol. This approach reduces the
processing impact to less that 2% compared to the 20% of PCIe 2.0, and allows
almost double bandwidth at 8 GTps speed.
Up to 32 PCIe 3.0 lanes are available per processor. These 32 lanes can be split
into any combination of x4, x8, and x16.
For more information about Data Direct I/O, go to the following website:
http://www.intel.com/content/www/us/en/io/direct-data-i-o.html
QuickPath Interconnect
The Intel Xeon E7 processors that are implemented in IBM X6 servers include
two integrated memory controllers in each processor. Processor-to-processor
communication is carried over shared-clock or coherent quick path interconnect
(QPI) links. Each processor has three QPI links to connect to other processors.
Figure 5-17 shows the QPI configurations. On the left side is how the four
sockets of the x3850 X6 are connected together. On the right side is how all eight
sockets of the x3950 X6 are connected together.
x3850 X6
4 sockets
x3950 X6 - 8 sockets
Each processor has some amount of memory, which is connected directly to the
processor. To access memory that is connected to another processor, each
processor uses QPI links through another processors. This design creates a
non-uniform memory access (NUMA) system. Similarly, I/O can be local to a
processor or remote through another processor.
For QPI usage, Intel modified the MESI cache coherence protocol to include a
forwarding state. Therefore, when a processor asks to copy a shared cache line,
only one other processor responds.
107
5.2.2 Memory
The x3850 X6 and x3950 X6 support DDR3 memory with ECC protection. The
x3850 X6 supports up to 96 DIMMs when all processors are installed (24 DIMMs
per processor), and the x3950 X6 supports up to 192 DIMMs. The processor, the
memory buffers, and the corresponding memory DIMM slots are on the Compute
Book.
Each processor has two integrated memory controllers, and each memory
controller has two Scalable Memory Interconnect generation 2 (SMI2) links that
are connected to two scalable memory buffers. Each memory buffer has two
DDR3 channels, and each channel supports three DIMMs, for a total of 24
DIMMs per processors.
108
Memory
controller
SMI2 links
Memory
buffer
Memory
buffer
Memory
buffer
Memory
buffer
DDR3 links
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
DIMM
The SMI2 links runs at a frequency of up to 1333 MHz and supports two transfers
per clock cycle, leading to an effective transfer rate of 2666 megatransfers per
second (MTps). Using a data width of 64-bit, this results in a channel bandwidth
of 21.3 GBps per SMI2 link.
Note: The Intel Xeon processor E7 v2 family allows for an alternative
operational mode that is called RAS mode or Lockstep mode. Although RAS
mode is supported on the IBM X6 architecture, you cannot enable it for any of
the workload-optimized solutions for SAP HANA.
Chipkill
Chipkill memory technology, an advanced form of ECC from IBM, is available for
the X6 servers. Chipkill (also known as Single Device Data Correction (SDDC))
protects the memory in the system from any single memory chip failure. It also
protects against multi-bit errors from any portion of a single memory chip.
Chipkill on its own can provide 99.94% memory availability to the applications
without sacrificing performance and with standard ECC DIMMs.
109
110
The SSD caching feature should not be confused with IBM FlashCache Storage Accelerator.
SSD caching works in a transparent manner to the operating system, that is, that
the SSDs are not visible as block devices by Linux. They can be seen only with
the controller configuration utility.
SSD caching of the local HDD arrays means that you can accelerate all types of
I/O operations that are sent to the local RAID devices. Because all storage
controllers in the X6 workload-optimized solutions for SAP HANA are configured
with SSD caching, all data that SAP HANA writes to disk is accelerated.
SSD caching is the most flexible way to speed up any kind of disk-based
operation in an SAP HANA solution.
111
Intel Xeon Processor E7-8880 v2: 2.5 GHz, for up to eight sockets (default)
Intel Xeon Processor E7-8890 v2: 2.8 GHz, for up to eight sockets3
Intel Xeon Processor E7-4880 v2: 2.5 GHz, for up to four sockets4
Intel Xeon Processor E7-4890 v2: 2.8 GHz, for up to four sockets4
112
Available upon request for compute-bound SAP HANA installations (such as SAP Business Suite
powered by SAP HANA).
Available upon special request only because CPU Books with E7-48x0 v2 processors cannot be
reused when scaling up to an x3950 X6 installation.
For an overview of which DIMMs are used with which memory configuration, see
6.2, IBM X6 based environments on page 157. DIMM placement is crucial for
best performance. It is not supported to change module placement or memory
configuration without consulting IBM first.
Storage
SAP requires a certain amount of persistent storage capacity in each SAP HANA
server to save main memory data onto a non-volatile storage media regularly.
This storage space is divided into a log volume and a data volume. For binary
and configuration files, SAP requires an additional volume. The exact storage
requirements depend on the nodes memory size. Bigger nodes must provide
more storage capacity.
IBM has the following storage devices in the x3850 X6 solution for SAP HANA:
3.6 TB (four 1.2 TB 2.5-inch SAS drives in a RAID 5) using four internal
storage bays in the Storage Book.
Two additional 400 GB 2.5-inch SAS SSDs are used for the SSD caching
feature of the ServeRAID adapter. They occupy two bays in the Storage Book.
9.6 TB (nine 1.2 TB 2.5-inch SAS drives in a RAID 5 or ten 1.2 TB 2.5-inch
SAS drives in a RAID 6) using storage expansion enclosure EXP2524,
connected to an IBM ServeRAID M5120 adapter.
Two additional 400 GB 2.5-inch SAS SSDs are used for the SSD caching
feature of the ServeRAID adapter. They occupy two bays in the EXP2524.
IBM uses General Parallel File System (GPFS) to provide a robust and
high-performant file system that you can use to non-disruptively grow your
environment as your needs increase. If you add additional drives to your node,
GPFS transparently includes this additional storage and starts using it. No data
is lost during this upgrade.
113
Figure 5-19 shows the storage architecture on an x3850 solution. The optional
second RAID array is shown in dashed lines. GPFS takes care of the different
sizes of the block devices. It balances I/O operations to maximize the usage of
both devices.
x3850 X6 node
SAP HANA DB
DB partition 1
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
Shared file system - GPFS
3.6 TB
First replica
9.6 TB
data + log
Additional hot spare drives are supported. They can be added without impacting
the overall performance of the solution.
Network
IBM solutions for SAP HANA have several different network interfaces that can
be grouped into the following modes:
Internal communication (HANA communication and GPFS communication).
Redundancy is required.
External communication (SAP data management, SAP client access, data
replication, appliance management, and others, depending on the customer
landscape). Redundancy is optional.
Internal communication remains internal within scale-out SAP HANA solutions.
They have no connection to the customer network. The networking switches for
these networks are part of the appliance and cannot be substituted with other
than the validated switch models.
114
IBM uses two dual-port Mellanox ConnectX-3 VPI adapters running in 10 Gbit
Ethernet mode for both internal communication networks. Each of the two
networks requires its own connection on two physical interfaces to allow for
redundancy in case of a hardware failure in the network. The IBM System
Networking RackSwitch G8264 is used as the switch for internal communication.
One switch handles both traffic, HANA, and GPFS communication. To allow for a
switch failure, a second G8264 is required in scale-out solutions. Smaller
scale-out installations can also use the G8124 switch. For more information
about the switches, see 5.4.1, IBM System Networking RackSwitch G8264 on
page 137 and 5.4.2, IBM System Networking RackSwitch G8124 on page 140.
For the uplink into the client network, an Intel quad-port 1 Gigabit Ethernet
adapter is included as the default adapter. A second adapter can be added if
more ports are required (either for access to more networks or for redundancy
reasons) or a different vendor can be chosen if, for example, 10 Gigabit is also
required towards the customer networks.
Figure 5-20 shows the back side of an x3850 X6 workload-optimized solution
with one quad-port 1 GbE card installed (the right-most PCIe slot). The network
names are examples only.
GPFS
SAP Appl.
SAP Data M.
HANA
SAP Data M
IMM
115
Figure 5-21 shows how the different network interfaces are connected.
SAP
BW
system
switch
B E
10 GbE switch
node 1
switch
node 2
...
node n
10 GbE switch
G H
D G H
Integrated virtualization
The VMware ESXi embedded hypervisor software is a virtualization platform that
allows multiple operating systems to run on a host system at the same time. Its
compact design allows it to be embedded in physical servers.
IBM offers versions of VMware vSphere Hypervisor (ESXi) customized for select
IBM hardware to give you online platform management, including updating and
configuring firmware, platform diagnostic tests, and enhanced hardware alerts.
All models support the USB keys as options, which are listed in Table 5-2.
Table 5-2 VMware ESXi memory keys for the x3850 X6
Part number
Feature code
Description
41Y8298
A2G0
41Y8382
A4WZ
41Y8385
A584
116
The x3850 X6 has one internal USB connector on the primary I/O book for the
embedded hypervisor USB key. The location of this USB connector is shown in
Figure 5-22.
Figure 5-22 Location of internal USB port for embedded hypervisor on the x3850 X6
For more information about the USB keys, and to download the IBM customized
version of VMware ESXi, go to the following website:
http://www.ibm.com/systems/x/os/vmware/esxi
In addition to the USB key, you also need to license vSphere with VMware.
Depending on the configuration of the physical server, you must buy different
licenses. For more information, see 6.4, SAP HANA on VMware vSphere on
page 177 This section also contains more information about virtualizing an SAP
HANA instance on x3850 X6 servers.
117
118
For an overview of which DIMMs are used with which memory configuration, see
6.2, IBM X6 based environments on page 157. DIMM placement is crucial for
best performance. Changing the module placement or memory configuration
without consulting IBM first is not supported.
Storage
SAP requires a certain amount of persistent storage capacity in each SAP HANA
server to save main memory data onto a non-volatile storage media regularly.
This storage space is divided into a log volume and a data volume. For binary
and configuration files, SAP requires an additional volume. The exact storage
requirements depend on the nodes memory size. Bigger nodes must provide
more storage capacity.
IBM has the following storage devices in the x3950 X6 solution for SAP HANA:
3.6 TB (four 1.2 TB 2.5-inch SAS drives in a RAID 5) using four internal
storage bays in the lower Storage Book.
Two additional 400 GB 2.5-inch SAS SSDs are used for the SSD caching
feature of the ServeRAID adapter. They occupy two bays in the lower Storage
Book.
6 TB (six 1.2 TB 2.5-inch SAS drives in a RAID 56) using six internal storage
bays in the upper Storage Book, which requires an additional IBM ServeRAID
M5210 storage controller in the upper Storage Book to be installed.
5
6
Available upon request for compute-bound SAP HANA installations (like SAP Business Suite
powered by SAP HANA)
Installations requiring RAID 6 have only 4.8 TB usable storage space instead of 6 TB.
119
Two 400 GB 2.5-inch SAS SSDs are used for the SSD caching feature of the
upper ServeRAID adapter. They occupy two bays in the upper Storage Book.
9.6 TB (nine 1.2 TB 2.5-inch SAS drives in a RAID 5 or ten 1.2 TB 2.5-inch
SAS drives in a RAID 6) using storage expansion enclosure EXP2524,
connected to an IBM ServeRAID M5120 storage controller.
Two additional 400 GB 2.5-inch SAS SSDs are used for the SSD caching
feature of the ServeRAID adapter. They occupy two bays in the EXP2524.
9.6 TB (nine 1.2 TB 2.5-inch SAS drives in a RAID 5 or ten 1.2 TB 2.5-inch
SAS drives in a RAID 6) using empty bays in the EXP2524.
Two additional 400 GB 2.5-inch SAS SSDs are used for the SSD caching
feature of the ServeRAID adapter. They occupy two bays in the EXP2524.
IBM uses General Parallel File System (GPFS) to provide a robust and
high-performance file system that allows to non-disruptively grow your
environment as your needs increase. If you add additional drives to your node,
GPFS transparently includes this additional storage and starts using it. No data
is lost during this upgrade.
Figure 5-23 on page 121 shows the storage architecture on an x3950 X6
solution. The optional second, third, and fourth RAID arrays are shown in dashed
lines. GPFS takes care of the different sizes of the block devices. It balances I/O
operations to maximize the usage of all available devices.
120
x3950 X6 node
SAP HANA DB
DB partition 1
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
6 TB
9.6 TB
9.6 TB
data + log
Additional hot spare drives are supported. They can be added without impacting
the overall performance of the solution.
Network
IBM solutions for SAP HANA have several different network interfaces that can
be grouped into the following modes:
Internal communication (HANA communication and GPFS communication).
Redundancy is required.
External communication (SAP data management, SAP client access, data
replication, appliance management, and others, depending on the customer
landscape). Redundancy is optional.
Internal communication remains internal within scale-out SAP HANA solutions.
They have no connection to the customer network. The networking switches for
these networks are part of the appliance and cannot be substituted with other
than the validated switch models.
121
IBM uses two dual-port Mellanox ConnectX-3 VPI adapters running in 10 Gbit
Ethernet mode for both internal communication networks. Each of the two
networks requires its own connection on two physical interfaces to allow for
redundancy in case of a hardware failure in the network. The IBM System
Networking RackSwitch G8264 is used as the switch for internal communication.
One switch handles both traffic, HANA, and GPFS communication. To allow for a
switch failure, a second G8264 is required in scale-out solutions. Smaller
scale-out installations can also use the G8124 switch. For more information
about the switches, see 5.4.1, IBM System Networking RackSwitch G8264 on
page 137 and 5.4.2, IBM System Networking RackSwitch G8124 on page 140.
For the uplink into the customer network, two Intel quad-port 1 Gigabit Ethernet
adapters are included as the default adapters. Third and fourth adapters can be
added if more ports are required (to access more networks) or a different vendor
can be chosen if, for example, 10 Gigabit is required also towards the customer
networks.
Figure 5-24 shows the back side of an x3950 X6 workload-optimized solution
with two quad-port 1G cards installed (the right-most PCIe slots). The network
names are examples only.
SAP Appl.
SAP Data M.
HANA
GPFS
IMM
SAP Data M
122
Figure 5-25 shows how the different network interfaces are connected.
SAP
BW
system
switch
B E
10 GbE switch
node 1
switch
node 2
...
node n
10 GbE switch
G H
D G H
Integrated virtualization
The VMware ESXi embedded hypervisor software is a virtualization platform that
allows multiple operating systems to run on a host system at the same time. Its
compact design allows it to be embedded in physical servers.
IBM offers versions of VMware vSphere Hypervisor (ESXi) that are customized
for selected IBM hardware to give you online platform management, including
updating and configuring firmware, platform diagnostic tests, and enhanced
hardware alerts. All models support the USB keys options that are listed in
Table 5-3.
Table 5-3 VMware ESXi memory keys for the x3950 X6
Part number
Feature code
Description
41Y8298
A2G0
41Y8382
A4WZ
41Y8385
A584
123
The x3950 X6 has two internal USB connectors on each of the primary I/O Books
for the embedded hypervisor USB key. The location of these USB connectors is
shown in Figure 5-26.
Figure 5-26 Location of the internal USB port for the embedded hypervisor on the x3950
X6
Although the x3950 X6 has two primary I/O books, you need to equip only one of
them with the embedded hypervisor. Installing two hypervisors is supported only
when the x3950 X6 is configured to be partitioned, where the two halves of the
server operate as two independent four-socket servers.
For more information about the USB keys, and to download the IBM customized
version of VMware ESXi, see the following website:
http://www.ibm.com/systems/x/os/vmware/esxi
In addition to the USB key, you also must license vSphere with VMware.
Depending on the configuration of the physical server, you must buy different
licenses. For more information, see 6.4, SAP HANA on VMware vSphere on
page 177. This section also contains more information about virtualizing an SAP
HANA instance on x3950 X6 servers.
124
125
126
GPFS File Placement Optimizer (GPFS FPO) is the name for a set of features to
support big data applications on shared-nothing architectures. In such
scenarios, hundreds or even thousands of commodity servers compute certain
problems. They do not have shared storage to hold the data. The internal disks of
the nodes are used to store all data, which requires a new way of thinking to run
a cluster file system on top of a shared-nothing architecture.
Here are some of the features that are introduced with GPFS-FPO:
Write affinity: Provides control over the placement of new data. It can either
be written to the local node or wide striped across multiple nodes.
Locality awareness: Ability to obtain on which node certain data chunks are.
This allows the scheduling of jobs on the node holding the data, thus avoiding
costly transfer of data across the network.
Metablocks: Enable two block sizes within the same file system. MapReduce
workloads tend to have small files (below 1 MB, for example, for index files)
and large files (such as 128 MB, holding the actual data) in the same file
system. The concept of metablocks allows for an optimal usage of the
available physical blocks.
Pipelined replication: Makes the most effective use of the node interconnect
bandwidth. Data that is written on node A sends data to node B, which in turn
sends data to node C. In contrast to pipelined replication, the other replication
schema is star replication, where node A sends data to both node B and node
C. For bandwidth-intense operations or for servers with limited network
bandwidth, the outgoing link of node A can limit replication performance in
such a scenario. Choosing the correct replication schema is important when
running in a shared-nothing architecture because this almost always involves
replicating data over the network.
Fast recovery: An intelligent way to minimize recovery efforts after the cluster
is healthy again. After an error, GPFS tracks what updates are missing
through the failed disks. In addition, the load to recover the data is distributed
across multiple nodes. GPFS also allows two different recovery policies. After
a disk has failed, data can either be rebuilt when the disk is replaced or it can
immediately be rebuilt using other nodes or disks to hold the data.
GPFS offers reliability and is installed on thousands of nodes across industries,
from weather research to multimedia, retail, financial industry analytics, and web
service providers. GPFS also is the basis of many IBM cloud storage offerings.
127
The IBM Systems Solution for SAP HANA benefits in several ways from the
features of GPFS:
GPFS provides a stable, industry-proven, cluster-capable file system for SAP
HANA.
GPFS transparently works with multiple replicas (that is, copies) of a single
file to protect from disk failures.
GPFS adds extra performance to the storage devices by striping data across
devices.
With the new FPO extensions, GPFS enables the IBM Systems Solution for
SAP HANA to grow beyond the capabilities of a single system, into a
scale-out solution, without introducing the need for external storage.
GPFS adds high-availability and disaster recovery features to the solution.
All these features make GPFS the ideal file system for the IBM Systems Solution
for SAP HANA.
128
Each node of the cluster holds its own savepoints and database logs on the
local storage devices of the server.
The GPFS file system is a shared file system. Because GPFS spans all
nodes of the cluster, it makes the data of each node available to all other
nodes in the cluster despite using local storage devices only (for more
information about this technology, see 5.3.2, GPFS extensions for
shared-nothing architectures on page 126).
To an outside application connecting to the SAP HANA database, this looks like a
single instance of SAP HANA. The SAP HANA software distributes the requests
internally across the cluster to the individual worker nodes, which process the
data and exchange intermediate results, which are then combined and sent back
to the requester. Each node maintains its own set of data, persisting it with
savepoints and logging data changes to the database log that are stored on local
storage.
GPFS combines the storage devices of the individual nodes into one large file
system, making sure that the SAP HANA software has access to all data
regardless of its location in the cluster. GPFS also makes sure that savepoints
and database logs of an individual database partition are stored on the
appropriate storage device of the node on which the partition is located. Although
GPFS provides the SAP HANA software with the functionality of a shared
storage system, it ensures maximum performance and minimum latency by using
locally attached disks and flash devices.
In addition, because server-local storage devices are used, the total capacity and
performance of the storage within the cluster automatically increases with the
addition of nodes, maintaining the same per-node performance characteristics
regardless of the size of the cluster. This kind of scalability is not achievable with
external storage system.
Note: With eX5 and X6 nodes, SAP validates the IBM scale-out solution for up
to 56 nodes in a cluster. However, the building block approach of IBM makes
the solution scalable without any known limitations.
IBM has shown scalability for up to 224 nodes in a single SAP HANA scale-out
cluster. With the current X6 servers, this allows for SAP HANA database
instances of up to 448 TB.
Clients requiring scale-out configurations beyond the generally available
56 nodes can work with IBM and SAP to jointly validate such large clusters at
the client site.
129
Scaling out an IBM SAP HANA solution creates a cluster of nodes. SAP HANA
designates nodes in a scale-out configuration with a certain role. They can be
either a worker node or a standby node. Worker nodes actively process
workload, and standby nodes are only part of the cluster and do not process
workload while the cluster remains in a healthy state. Standby nodes take over
the role of a worker node when it fails. Standby nodes are required for scale-out
clusters with high availability.
SAP HANA
worker node
SAP HANA
worker node
SAP HANA
worker node
SAP HANA
worker node
Figure 5-27 Network architecture of a four node scale-out solution (eX5 and X6)
130
Going up the stack, Figure 5-28 gives insight into a three-node configuration and
shows how GPFS stripes data across the nodes. Local storage can either be a
single storage device (like for smaller X6 nodes) or multiple devices (for bigger
X6 nodes or x3690 X5 nodes), or even devices with different underlying
technologies (such as HDD and flash memory on x3950 X5 nodes).
node01
node02
node03
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- Index server
- Statistic server
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
131
132
Figure 5-29 shows a four-node cluster with the fourth node being a standby
node.
node01
node02
node03
node04
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- Index server
- Statistic server
- Index server
- Statistic server
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
Second replica
local storage
If a node has an unrecoverable hardware error, the storage devices holding the
nodes data might become unavailable or even destroyed. In contrast to
Scale-out solution without high-availability capabilities on page 130, when
high-availability features are implemented, the GPFS file system replicates the
data of each node to the other nodes, creating a second replica, to prevent data
loss in case one of the nodes goes down. Replication is done in a striping
fashion. Every node has a piece of data of all other nodes. In Figure 5-29, the
contents of the data storage (that is, the savepoints, here data01) and the log
storage (that is, the database logs, here log01) of node01 are replicated to
node02, node03, and node04.
Replication happens for all nodes generating data, so that all information is
available twice within the GPFS file system, which makes it tolerant to the loss of
a single node. Replication occurs synchronously. The write operation finishes
only when the data is both written locally and on a remote node. This ensures
consistency of the data at any point in time. Although GPFS replication is done
over the network and in a synchronous fashion, this solution still over-achieves
the performance requirements for validation by SAP.
133
The File Placement Optimizer (FPO), part of GPFS, ensures that the first replica
always is stored local to the node generating the data. In case SAP HANA data
must be read from disk (for example, for backups or restore activity), FPO always
prefers the replica that is available locally. This ensures the best read
performance of the cluster.
Using replication, GPFS provides the SAP HANA software with the functionality
and fault tolerance of a shared storage subsystem while maintaining its
performance characteristics. Again, because server-local storage devices are
used, the total capacity and performance of the storage within the cluster
automatically increases with the addition of nodes, maintaining the same
per-node performance characteristics regardless of the size of the cluster. This
kind of scalability is not achievable with external storage systems.
134
node01
node02
node03
node04
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- SAP HANA DB
Defunct node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- Index server
- Statistic server
- Index server
- Statistic server
First replica
data01 + log01
data02 + log02
local storage
local storage
data03 + log03
Second replica
local storage
local storage
The data that node04 must load into memory is the data of node03, which failed,
including its local storage devices. For that reason, GPFS had to deliver the data
to node04 from the second replica, which is spread across the cluster. GPFS
handles this transparently so that the application does not recognize from which
node the data was read. If data is available locally, GPFS prefers to read from
node04 and avoid going over the network.
Now, when node04 starts writing savepoints and database logs again during the
normal course of operations, these are not written over the network, but to the
local drives, again with a second replica striped across the other cluster nodes.
135
After fixing the cause for the failure of node03, it can be reintegrated into the
cluster as the new standby system. This situation is shown in Figure 5-31.
node01
node02
node03
node04
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
- SAP HANA DB
Worker node
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- Index server
- Statistic server
- Index server
- Statistic server
First replica
data01 + log01
data02 + log02
local storage
local storage
data03 + log03
Second replica
local storage
local storage
136
All IBM System x solutions for SAP HANA use network switches that meet these
requirements. There are three top-of-rack Ethernet switches that are part of the
scale-out solution of IBM:
IBM System Networking RackSwitch G8264
The G8264 switch is a 10 Gb Ethernet switch with 64 SFP+ ports. This switch
is used in scale-out solutions to provide internal cluster communication for
GPFS and SAP HANA networks.
IBM System Networking RackSwitch G8124
The G8124 switch is a 10 Gb Ethernet switch with 24 SFP+ ports. It is used
for smaller installations that do not need as many ports as the G8264 switch
provides.
IBM System Networking RackSwitch G8052
The G8052 switch is a 1 Gb Ethernet switch with 48 10/100/1000 BASE-T
RJ45 ports. It is used in scale-out environments for management and SAP
client networks.
137
The G8264 switch offers the following benefits with respect to SAP HANA
environments:
High performance: The 10 Gb/40 Gb Low Latency Switch provides the best
combination of low latency, non-blocking line-rate switching, and ease of
management. It also has a throughput of 1.2 Tbps.
Lower power and better cooling: The G8264 switch uses as little as 275 W of
power, which is a fraction of the power consumption of most competitive
offerings. Unlike side-cooled switches, which can cause heat recirculation and
reliability concerns, the G8264 switchs front-to-rear or rear-to-front cooling
design reduces data center air conditioning costs by having airflow match the
servers in the rack. In addition, variable speed fans assist in automatically
reducing power consumption.
Layer 3 functionality: IBM System Networking RackSwitch switches include
Layer 3 functionality, which provides security and performance benefits, as
inter-VLAN traffic stays within the switch. These switches also provide the full
range of Layer 3 protocols from static routes for technologies, such as Open
Shortest Path First (OSPF) and Border Gateway Protocol (BGP) for
enterprise customers.
Seamless interoperability: IBM System Networking RackSwitch switches
interoperate seamlessly with other vendors' upstream switches.
Fault tolerance: IBM System Networking RackSwitch switches learn
alternative routes automatically and perform faster convergence in the
unlikely case of a link, switch, or power failure. The switches use proven
technologies, such as L2 trunk failover, advanced VLAN-based failover,
VRRP, and Hot Link.
Multicast: This switch supports IGMP Snooping v1, v2, and v3 with 2 K IGMP
groups, and Protocol Independent Multicast, such as PIM Sparse Mode or
PIM Dense Mode.
Converged fabric: IBM System Networking RackSwitch switches are
designed to support CEE and connectivity to FCoE gateways. CEE helps
enable clients to combine storage, messaging traffic, VoIP, video, and other
data on a common data center Ethernet infrastructure. FCoE helps enable
highly efficient block storage over Ethernet for consolidating server network
connectivity. As a result, clients can deploy a single-server interface for
multiple data types, which can simplify both deployment and management of
server network connectivity, while maintaining the high availability and
robustness required for storage transactions.
138
139
FCoE/Lossless Ethernet:
802.1 Data Center Bridging
Priority Based Flow Control (PFC)
Enhanced Transmission Selection (ETS)
Data Center Bridge Exchange protocol (DCBX)
FIP Snooping
Fibre Channel over Ethernet (FCoE)
Converged Enhanced Ethernet (CEE)
Trunking:
LACP
Static Trunks (Etherchannel)
Configurable Trunk Hash algorithm
Spanning Tree:
Multiple Spanning Tree (802.1 s)
Rapid Spanning Tree (802.1 w)
PVRST+
Fast Uplink Convergence
BPDU guard
High availability:
Layer 2 failover
Hot Links
VRRP
140
The G8124 switch offers the following feature benefits with respect to SAP HANA
environments:
High performance: The 10G Low Latency (<700 ns) switch provides the best
combination of low latency, non-blocking line-rate switching and ease of
management.
Lower power and better cooling: The G8124 switch uses as little power as two
60 W light bulbs, which is a fraction of the power consumption of most
competitive offerings. Unlike side-cooled switches, which can cause heat
recirculation and reliability concerns, the G8124 switchs rear-to-front cooling
design reduces data center air conditioning costs by having airflow match the
servers in the rack. In addition, variable speed fans assist in automatically
reducing power consumption.
Layer 3 functionality: This IBM System Networking RackSwitch includes Layer
3 functionality, which provides security and performance benefits, as
inter-VLAN traffic stays within the switch. This switch also provides the full
range of Layer 3 protocols from static routes for technologies, such as Open
Shortest Path First (OSPF) and Border Gateway Protocol (BGP) for
enterprise customers.
Active MultiPath (AMP): Effectively doubles bandwidth by allowing all uplink
ports to be active/active, eliminating cross-stack traffic, and providing up to
900 Gbps aggregate bandwidth between servers. Built-in fault tolerance
constant health checking ensures maximum availability.
Seamless interoperability: IBM System Networking RackSwitch switches
interoperate seamlessly with other vendors' upstream switches.
Fault tolerance: IBM System Networking RackSwitch switches learn
alternative routes automatically and perform faster convergence in the
unlikely case of a link, switch, or power failure. The switch uses proven
technologies, such as L2 trunk failover, advanced VLAN-based failover,
VRRP, and Hot Link.
Converged fabric: The IBM System Networking RackSwitch is designed to
support CEE and connectivity to FCoE gateways. CEE helps enable clients to
combine storage, messaging traffic, VoIP, video, and other data on a common
data center Ethernet infrastructure. FCoE helps enable highly efficient block
storage over Ethernet for consolidating server network connectivity. As a
result, clients can deploy a single-server interface for multiple data types,
which can simplify both deployment and management of server network
connectivity, while maintaining the high availability and robustness that are
required for storage transactions.
141
The G8124 switch provides the best combination of low latency, non-blocking
line-rate switching and ease of management. The G8124 switch has the following
performance characteristics:
100% line rate performance
Latency under 700 ns (ultra-low latency, less than 1 microsecond)
480 Gbps non-blocking switching throughput (full duplex)
Interface options:
Twenty-four 10G SFP+ fiber connectors
In HA scale-out configurations, the first four ports of the switch should be
used for ISL between switches; the other 20 ports may be used for nodes
connections. Each cluster node uses two switch ports (one port for internal
GPFS network, one port for internal HANA network), so the G8124 switch
supports up to 10 nodes in the cluster.
2x 10/100/1000 Ethernet RJ45 ports for management
Two dedicated RJ45 ports are used as the management ports of the switch.
The IBM System Networking RackSwitch G8264 supports the following features:
Security:
RADIUS
TACACS+
SCP
Wire Speed Filtering: Allow and Deny
SSH v1 and v2
HTTPS Secure BBI
Secure interface login and password
MAC address move notification
Shift B Boot menu (Password Recovery/ Factory Default)
VLANs:
Port-based VLANs
4096 VLAN IDs supported
1024 Active VLANs (802.1Q)
Private VLAN Edge
FCoE/Lossless Ethernet:
802.1 Data Center Bridging
Priority Based Flow Control (PFC)
Enhanced Transmission Selection (ETS)
Data Center Bridge Exchange protocol (DCBX)
FIP Snooping
Fibre Channel over Ethernet (FCoE)
Converged Enhanced Ethernet (CEE)
142
Trunking:
LACP
Static Trunks (Etherchannel)
Configurable Trunk Hash algorithm
Spanning Tree:
Multiple Spanning Tree (802.1 s)
Rapid Spanning Tree (802.1 w)
PVRST+
Fast Uplink Convergence
BPDU guard
High availability:
Layer 2 failover
Hot Links
VRRP
The G8052 switch is a Top-of-Rack data center switch that delivers unmatched
line-rate Layer 2/3 performance at an attractive price. It has forty-eight
10/100/1000 BASE-T RJ45 ports and four 10 Gigabit Ethernet SFP+ ports, and
includes hot-swap redundant power supplies and fans standard, minimizing your
configuration requirements. Unlike most rack equipment that cools from side to
side, the G8052 switch has rear-to-front or front-to-rear airflow that matches
server airflow.
For 10 Gb uplinks, there is a choice of either SFP+ transceivers (SR or LR) for
longer distances or more cost-effective and lower-power-consuming options,
such as SFP+ direct-attached cables (DAC or Twinax cables), which can be
1 - 7 meters in length and are ideal for connecting to another TOR switch, or even
connecting to an adjacent rack.
143
The G8052 switch provides the following features with respect to SAP HANA
environments:
High performance: The G8052 switch provides up to 176 Gbps throughput
and supports four SFP+ 10 Gb uplink ports for a low oversubscription ratio, in
addition to a low latency of 1.7 ms.
Lower power and better cooling: The G8052 switch typically consumes just
120 W of power, a fraction of the power consumption of most competitive
offerings. Unlike side-cooled switches, which can cause heat recirculation and
reliability concerns, the G8052 switchs rear-to-front or front-to-rear cooling
design reduces data center air conditioning costs by matching airflow to the
servers configuration in the rack. Variable speed fans assist in automatically
reducing power consumption.
Layer 3 functionality: The G8052 switch includes Layer 3 functionality, which
provides security and performance benefits, as inter-VLAN traffic can be
processed at the access layer. This switch also provides the full range of
Layer 3 static and dynamic routing protocols, including Open Shortest Path
First (OSPF) and Border Gateway Protocol (BGP) for enterprise customers at
no additional cost.
Fault tolerance: These switches learn alternative routes automatically and
perform faster convergence in the unlikely case of a link, switch, or power
failure. The switch uses proven technologies, such as L2 trunk failover,
advanced VLAN-based failover, VRRP, Hot Link, Uplink Failure Detection
(UFD), IGMP v3 Snooping, and OSPF.
Seamless interoperability: IBM RackSwitch switches interoperate seamlessly
with other vendors' upstream switches.
Here are the performance features and specifications of the G8052 switch:
Interface options:
Forty-eight 10/100/1000BaseT ports (RJ-45)
Four 10 GbE SFP+ ports
G8052 switches are used in all System x offerings for SAP HANA as switches for
the management network (connecting IMM interfaces of the nodes) and as the
default choice for SAP HANA appliances to the client network. Clients that have
10 Gbps Ethernet backbone for their SAP environments can also choose the
G8264 switch for uplink connectivity.
144
Chapter 6.
145
146
L
M
M
M
XL
XXL
SoH
SoH
BW, SoH
M
M
M
M
XM
SoH
BW, SoH
XS
S
SSS
BW, SoH
S+
BW = Business Warehouse
SoH = Suite on HANA
BW, SoH
BW, SoH
128 GB
256 GB
256 GB
512 GB
1 TB
2 TB
4 TB
Memory size
Figure 6-1 Portfolio of eX5 based building blocks for SAP HANA
In some cases, there are several building blocks that are available for one T-shirt
size. In some cases, two-building blocks must be combined to build a specific
T-shirt size. Table 6-1 shows all eX5-based building blocks and their features.
Table 6-1 Overview of all eX5 workload-optimized models for SAP HANA
Building
block
Server
CPUs
Main
memory
Business
Warehouse
Suite on HANA
Upgrade
options
XS
x3690 X5
2x Intel Xeon
E7-2870
128 GB
Yes
Yes
XS S
x3690 X5
2x Intel Xeon
E7-2870
256 GB
Yes
Yes
None
S+
x3950 X5
2x Intel Xeon
E7-8870
256 GB
Yes
Yes
S+ M
x3950 X5
4x Intel Xeon
E7-8870
512 GB
Yes
Yes
M XM
ML
XM
x3950 X5
4x Intel Xeon
E7-8870
1 TB
No
Yes
XM XL
x3950 X5 +
x3950 X5
8x Intel Xeon
E7-8870
1 TB
Yes
Yes
L XL
147
Building
block
Server
CPUs
Main
memory
Business
Warehouse
Suite on HANA
Upgrade
options
XL
x3950 X5 +
x3950 X5
8x Intel Xeon
E7-8870
2 TB
No
Yes
XL
XXL
XXL
x3950 X5 +
x3950 X5
8x Intel Xeon
E7-8870
4 TB
No
Yes
None
All models come with preinstalled software that is composed of SUSE Linux
Enterprise Server for SAP Applications (SLES for SAP) 11, IBM GPFS, and the
SAP HANA software stack. Licenses and maintenance fees (for three years) for
SLES for SAP and GPFS are included. The section GPFS license information
on page 256 has an overview about which type of GPFS license comes with a
specific model.
XM, XL, and XXL building blocks are specific to and limited for use with SAP
Business Suite powered by SAP HANA. They have a different memory to core
ratio than the regular models, which is only suitable for this specific workload, as
described in 4.5.3, SAP Business Suite powered by SAP HANA on page 76.
148
Processors
L
M
M
M
M
M
M
M
XS
128 GB
S
SSS
256 GB
S+
256 GB
512 GB
1 TB
Memory size
Figure 6-2 Portfolio of single-node eX5 based solutions for Business Warehouse
There are two different servers on which these solutions are based on: the x3690
X5 and the IBM System 3850 X5 and x3950 X5. Table 6-2 lists the details of each
of the building blocks.
Table 6-2 IBM eX5 workload-optimized models for Business Warehouse on SAP HANA
Building
block
Server
(MTM)
CPUs
Main
memory
Log
storage
Data
storage
Upgrade
options
XS
x3690 X5
(7147-HAxa)
2x Intel Xeon
E7-2870
128 GB DDR3
(8x 16 GB)
x3690 X5
(7147-HBx)
2x Intel Xeon
E7-2870
256 GB DDR3
(16x 16 GB)
S+
x3950 X5
(7143-HAx)
2x Intel Xeon
E7-8870
256 GB DDR3
(16x 16 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS HDD
S+ M
x3950 X5
(7143-HBx)
4x Intel Xeon
E7-8870
512 GB DDR3
(32x 16 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS HDD
ML
XS S
149
Building
block
Server
(MTM)
CPUs
Main
memory
Log
storage
Data
storage
x3950 X5
(7143-HBx)
+
x3950 X5
(7143-HCx)
8x Intel Xeon
E7-8870
1 TB DDR3
(64x 16 GB)
1.2 TB High
IOPS adapter
14x 900 GB
10 K SAS HDD
Upgrade
options
a. x = Country-specific letter (for example, EMEA MTM is 7147-HAG, and the US MTM is 7147-HAU).
Contact your IBM representative for regional part numbers.
b. The two servers are connected with QPI wrap cards that leverage eX5 scalability.
You cannot upgrade from the x3690 X5 to the x3950 X5. Upgrades within one
server type are supported. The following upgrade options exist:
An XS building block can be upgraded to be an S-size SAP HANA system by
adding 128 GB of main memory to the system.
An S+ building block can be upgraded to be an M-size SAP HANA system by
adding two more processors plus another 256 GB of main memory.
M building blocks can be extended with the L option to resemble an L-size
SAP HANA system.
With the option to upgrade S+ to M, and M to L, IBM can provide an
unmatched upgrade path from a T-shirt size S+ up to a T-shirt size L, without
the need to retire a single piece of hardware.
You can also grow into a clustered scale-out environment without changing any
of the hardware of the single-node solution. The upgrade to scale-out is
supported for S, M, and L (as denoted by the multiple boxes in Figure 6-2 on
page 149). Scale-out of XS and S+ is not supported because adding more
memory first, that is, scaling up, gives better performance than adding additional
nodes.
Upgrading the server requires downtime of the system. However, because of the
capability of GPFS to add storage capacity to an existing GPFS file system by
just adding devices, data that is on the system remains intact. Do a backup of the
data before changing the systems configuration.
150
XS
XL
XXL
XM
S+
Memory size
128 GB
256 GB
256 GB
512 GB
1 TB
2 TB
4 TB
Figure 6-3 Portfolio of eX5 based building blocks for SAP Business Suite on SAP HANA
151
The XS and S building blocks are based on the x3690 X5 server; all other T-shirt
sizes are based on the x3950 X5 building blocks. Table 6-3 lists the configuration
details of each building block for a Business Suite environment.
Table 6-3 IBM eX5 workload-optimized models for SAP Business Suite on SAP HANA
Building
block
Server
(MTM)
CPUs
Main
memory
Log
storage
Data
storage
Upgrade
options
XS
x3690 X5
(7147-HAxa)
2x Intel Xeon
E7-2870
128 GB DDR3
(8x 16 GB)
XS S
x3690 X5
(7147-HBx)
2x Intel Xeon
E7-2870
256 GB DDR3
(16x 16 GB)
None
S+
x3950 X5
(7143-HAx)
2x Intel Xeon
E7-8870
256 GB DDR3
(16x 16 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS HDD
S+ M
x3950 X5
(7143-HBx)
4x Intel Xeon
E7-8870
512 GB DDR3
(32x 16 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS HDD
M XM
ML
XM
x3950 X5
(7143-HDx)
4x Intel Xeon
E7-8870
1 TB DDR3
(32x 32 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS HDD
XM XL
x3950 X5
(7143-HBx)
+
x3950 X5
(7143-HCx)
8x Intel Xeon
E7-8870
1 TB DDR3
(64x 16 GB)
1.2 TB High
IOPS adapter
14x 900 GB
10 K SAS HDD
L XL
8x Intel Xeon
E7-8870
2 TB DDR3
(64x 32 GB)
2x 1.2 TB
High IOPS
adapter
14x 900 GB
10 K SAS HDD
XL
XXL
8x Intel Xeon
E7-8870
4 TB DDR3
(128x 32 GB)
4x 1.2 TB
High IOPS
adapter
14x 900 GB
10 K SAS HDD
+ 8x 900 GB
10 K SAS HDD
through
EXP2524
None
XL
x3950 X5
(7143-HDx)
+
x3950 X5
(7143-HEx)
b
XXL
x3950 X5
(7143-HDx)
+
x3950 X5
(7143-HEx)
b
a. x = Country-specific letter (for example, EMEA MTM is 7147-HAG, and the US MTM is 7147-HAU).
Contact your IBM representative for regional part numbers.
b. The two servers are connected with QPI wrap cards that leverage eX5 scalability.
152
You cannot upgrade from the x3690 X5 to the x3950 X5 building blocks.
Upgrades within one server type are supported. The following upgrade options
exist:
An XS building block can be upgraded to be an S-size system by adding
128 GB of main memory to the system.
An S+ building block can be upgraded to be an M-size system by adding two
more processors plus another 256 GB of main memory.
M building blocks can be extended with the L option to resemble an L-size
system.
With the option to upgrade S+ to M, and M to L, IBM can provide an
unmatched upgrade path from a T-shirt size S+ up to a T-shirt size L, without
the need to retire a single piece of hardware.
Customers can start with an S+ configuration, and then upgrade M and L, and
finally to XL only by adding new components. Further growth to XXL is
possible, but requires an exchange of the memory DIMMs.
Clients starting at a 1 TB configuration can upgrade from XM to XL and then
XXL without the need to retire a single piece of hardware, which quadruples
the memory capacity.
Upgrading the server requires downtime of the system. However, because of the
capability of GPFS to add storage capacity to an existing GPFS file system by
just adding devices, data that is on the system remains intact. Do a backup of the
data before changing the systems configuration.
153
All HA and DR solutions can be implemented using either GPFS based storage
replication or SAP HANA System Replication. You also can use the standby
nodes in HA and DR solutions to run a second SAP HANA instance for
non-productive purposes. For more information about these architectures, see
7.2, HA and DR for single-node SAP HANA on page 196.
Adding additional business continuity features after the initial implementation is
easily possible. Only additional hardware must be bought, and no parts must be
retired.
Scale-out SAP HANA environments are not supported when running SAP
Business Suite on top.
Processors
L
M
M
M
M
M
M
M
S
SSS
Memory size
256 GB
512 GB
1 TB
Figure 6-4 Portfolio of eX5 based solutions for scale-out Business Warehouse
The S building block is based on the x3690 X5, and M and L are based on the
x3950 X5 building blocks. Table 6-4 on page 155 lists the configuration details of
each building block.
154
Table 6-4 IBM eX5 workload-optimized models for Business Warehouse on SAP HANA
Building
block
Server
(MTM)
CPUs
Main
memory
Log
storage
Data
storage
Maximum
cluster
nodes
x3690 X5
(7147-HBxa)
2x Intel Xeon
E7-2870
256 GB DDR3
(16x 16 GB)
16
x3950 X5
(7143-HBx)
4x Intel Xeon
E7-8870
512 GB DDR3
(32x 16 GB)
1.2 TB High
IOPS adapter
6x 900 GB
10 K SAS
HDD
56
x3950 X5
(7143-HBx)
+
x3950 X5
(7143-HCx)b
8x Intel Xeon
E7-8870
1 TB DDR3
(64x 16 GB)
1.2 TB High
IOPS adapter
14x 900 GB
10 K SAS
HDD
56
a. x = Country-specific letter (for example, EMEA MTM is 7147-HAG, and the US MTM is 7147-HAU).
Contact your IBM representative for regional part numbers.
b. The two servers are connected with QPI wrap cards that leverage eX5 scalability.
An SAP HANA cluster environment can consist of only building blocks with the
same memory size. Mixing M and L nodes in a cluster is not supported. When
upgrading from a cluster with M nodes to L nodes, you must add the additional
hardware to every node.
The minimum number of nodes in any SAP HANA scale-out environment is three
nodes. A cluster of two nodes is not supported. The maximum number of nodes
that is generally available for implementation is 56. However, IBM validates
feasibility for up to 224 nodes in its labs.1
If you are interested in solutions beyond 56 nodes, contact your IBM representative or email the
IBM SAP International Competence Center at [email protected].
155
156
For a scale-out solution built upon the SSD-only building blocks that are based
on x3690 X5, additional 200 GB 1.8 MLC SSD drives are required to
accommodate the additional storage capacity that is required for GPFS
replication. The total number of SSD drives that are required is documented in
the SAP Product Availability Matrix (PAM) for SAP HANA, which is available
online at (search for HANA) the following website:
http://service.sap.com/pam
157
BW, SoH
BW, SoH
BW, SoH
BW, SoH
SoH
SoH
BW = Business Warehouse
SoH = Suite on HANA
Scale-out for BW only
2
BW, SoH
128 GB
BW, SoH
256 GB
BW, SoH
384 GB
BW, SoH
512 GB
768 GB
1 TB
1.5 TB
2 TB
Memory size
Figure 6-5 Portfolio of x3850 X6 based building blocks for SAP HANA
Figure 6-6 on page 159 shows all the building blocks for SAP HANA solutions
that are based on the x3950 X6 server, which covers the mid-range to high-end
requirements.
158
Processors
BW, SoH
BW, SoH
BW, SoH
BW, SoH
SoH
SoH
SoH
BW = Business Warehouse
SoH = Suite on HANA
Scale-out for BW only
4
BW, SoH
256 GB
BW, SoH
BW, SoH
512 GB
768 GB
BW, SoH
1 TB
SoH
1.5 TB
SoH
2 TB
3 TB
4 TB
6 TB
Memory size
Figure 6-6 Portfolio of x3950 X6 based building blocks for SAP HANA
Table 6-5 shows the technical details of all x3850 X6-based workload-optimized
models for SAP HANA. They cover the range of two and four socket solutions.
Table 6-5 Overview of x3850 X6 workload-optimized models for SAP HANA
Model
CPUsa
Main
memory
Business
Warehouse
Scale-out
BW
Suite on
HANA
Total rack
space
AC3-2S-128
2x Intel Xeon
E7-8880 v2
128 GB DDR3
Yes
No
Yes
4U
AC3-2S-256
2x Intel Xeon
E7-8880 v2
256 GB DDR3
Yes
Yes, up to
4 nodes
Yes
4U
AC3-2S-384
2x Intel Xeon
E7-8880 v2
384 GB DDR3
Yes
No
Yes
4U
AC3-2S-512
2x Intel Xeon
E7-8880 v2
512 GB DDR3
Yes
No
Yes
4U
AC3-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
Yes
No
Yes
4U
AC3-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
Yes
Yes, up to
56 nodesb
Yes
4Uc
159
Model
CPUsa
Main
memory
Business
Warehouse
Scale-out
BW
Suite on
HANA
Total rack
space
AC3-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
Yes
Yes, up to
56 nodesb
Yes
6U
AC3-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
Yes
Yes, up to
56 nodesb
Yes
6U
AC3-4S-1536
4x Intel Xeon
E7-8880 v2
1.5 TB DDR3
No
N/A
Yes
6U
AC3-4S-2048
4x Intel Xeon
E7-8880 v2
2 TB DDR3
No
N/A
Yes
6U
a. Alternative CPU types E7-4880 v2, E7-4890 v2 (both support no upgrade to eight sockets), and
E7-8890 v2 are available upon request. For more information, see CPU and memory on
page 112.
b. Support for up to 224 nodes is verified by IBM development and presented to SAP.
c. 6U when used in a scale-out environment.
Table 6-6 shows the technical details of all x3950 X6 based workload-optimized
models for SAP HANA. They cover the range of four and eight socket solutions.
Table 6-6 Overview of IBM System x3950 X6 workload-optimized models for SAP HANA
Model
CPUsa
Main
memory
Business
Warehouse
Scale-out
BW
Suite on
HANA
Total rack
space
AC4-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
AC4-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
AC4-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
AC4-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
AC4-4S-1536
4x Intel Xeon
E7-8880 v2
1.5 TB DDR3
No
N/A
Yes
8U
AC4-4S-2048
4x Intel Xeon
E7-8880 v2
2 TB DDR3
No
N/A
Yes
8U
AC4-8S-512
8x Intel Xeon
E7-8880 v2
512 GB DDR3
Yes
No
Yes
8U
AC4-8S-1024
8x Intel Xeon
E7-8880 v2
1 TB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
160
Model
CPUsa
Main
memory
Business
Warehouse
Scale-out
BW
Suite on
HANA
Total rack
space
AC4-8S-1536
8x Intel Xeon
E7-8880 v2
1.5 TB DDR3
Yes
Yes, up to
56 nodesb
Yes
8U
AC4-8S-2048
8x Intel Xeon
E7-8880 v2
2 TB DDR3
Yes
Yes, up to
56 nodesb
Yes
8Uc
AC4-8S-3072
8x Intel Xeon
E7-8880 v2
3 TB DDR3
No
N/A
Yes
10U
AC4-8S-4096
8x Intel Xeon
E7-8880 v2
4 TB DDR3
No
N/A
Yes
10U
AC4-8S-6144
8x Intel Xeon
E7-8880 v2
6 TB DDR3
No
N/A
Yes
10U
a. Alternative CPU type E7-8890 v2 is supported for enhanced performance. Available upon request.
b. Support for up to 224 nodes is verified by IBM development and presented to SAP.
c. 10U when used in a scale-out environment.
161
Figure 6-7 gives an overview of all X6 based models that are available for a
single-node SAP BW environment.
Processors
x3950 X6
Upgrade supported by
replacing mechanical chassis
4
x3850 X6
Memory size
128 GB
256 GB
384 GB
512 GB
768 GB
1 TB
1.5 TB
2 TB
Figure 6-7 Portfolio of single-node X6 based solutions for Business Warehouse (x3850 X6 and x3950 X6)
The lower half of Figure 6-7 (below the dashed line) shows the available models
that are based on the x3850 X6 server, and the upper half shows the models that
are based on the x3950 X6 server. There are four socket models for both of
them. You physically can upgrade an x3850 X6 to an x3950 X6 server by
replacing the mechanical enclosure that houses the CPU Books, Storage Books,
and I/O Books. The Books can be used in both server types.
Table 6-5 on page 159 shows the technical details of all x3850 X6
workload-optimized models for SAP HANA.
162
Table 6-7 IBM System x3850 X6 workload-optimized models for single-node Business Warehouse
Model)
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
AC3-2S-128
2x Intel Xeon
E7-8880 v2
128 GB DDR3
(16x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
128 256
AC3-2S-256
2x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB or
16x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 384
256 512
2S 4S
AC3-2S-384
2x Intel Xeon
E7-8880 v2
384 GB DDR3
(48x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
2S 4S-512
AC3-2S-512
2x Intel Xeon
E7-8880 v2
512 GB DDR3
(32x 16 GB or
16x 32 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
2S 4S
AC3-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
AC3-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 768
512 1 TB
AC3-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
None
AC3-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
1 TB 1.5 TB
1 TB 2 TB
a. Alternative CPU types E7-4880 v2, E7-4890 v2 (both support no upgrade to eight sockets),
and E7-8890 v2 are available upon request. For more information, see CPU and memory
on page 112.
b. Additional drives are installed in the EXP2524 expansion unit.
163
Table 6-6 on page 160 shows the technical details of all x3950 X6
workload-optimized models for SAP HANA.
Table 6-8 Overview of x3950 X6 workload-optimized models for SAP HANA
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
AC4-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
AC4-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 768
512 1 TB
4S 8S
AC4-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
4S 8S-1 TB
AC4-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
1 TB 1.5 TB
1 TB 2 TB
4S 8S
AC4-4S-1536
4x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(96x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
4S 8S
AC4-4S-2048
4x Intel Xeon
E7-8880 v2
2 TB DDR3
(64x 32 GB or
32x 64 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
None
AC4-8S-512
8x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 1 TB
AC4-8S-1024
8x Intel Xeon
E7-8880 v2
1 TB DDR3
(128x 8 GB or
64x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
1 TB 1.5 TB
1 TB 2 TB
AC4-8S-1536
8x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(192x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
None
AC4-8S-2048
8x Intel Xeon
E7-8880 v2
2 TB DDR3
(128x 16 GB or
64x 32 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
2 TB 3 TB
2 TB 4 TB
a. Alternative CPU type E7-8890 v2 is supported for enhanced performance. Available upon request.
164
You can also grow into a clustered scale-out environment without changing any
of the hardware of the single-node solution. The upgrade to scale-out is
supported for memory configurations with 256 GB, 512 GB, 768 GB, 1 TB,
1.5 TB, or 2 TB (as denoted by the multiple boxes in Figure 6-7 on page 162).
Scale-out of other memory sizes is not supported because adding more memory
first (to scale up) to reach one of the supported memory sizes gives better
performance than adding additional nodes.
Upgrading the server requires downtime of the system. However, because of the
capability of GPFS to add storage capacity to an existing GPFS file system by
just adding devices, data that is on the system remains intact. Do a backup of the
data before changing the systems configuration.
All HA and DR solutions can be implemented using either GPFS based storage
replication or SAP HANA System Replication. You also can use the standby
nodes in HA and DR solutions to run a second SAP HANA instance for
non-productive purposes. For more information about these architectures, see
7.2, HA and DR for single-node SAP HANA on page 196.
Adding additional business continuity features after the initial implementation is
easily possible. Only additional hardware must be bought, and no parts must be
retired.
165
Figure 6-8 shows all the building blocks that you can choose from when you plan
for SAP Business Suite powered by SAP HANA.
Processors
x3950 X6
Upgrade supported
by replacing
mechanical chassis
4
x3850 X6
Memory size
128 GB
256 GB
384 GB
512 GB
768 GB
1 TB
1.5 TB
2 TB
3 TB
4 TB
6 TB
Figure 6-8 Portfolio of single-node X6 based models for SAP Business Suite powered by SAP HANA
The lower half of Figure 6-8 (below the dashed line at four processors) shows the
available models that are based on the x3850 X6 server and the upper half
shows the models that are based on the x3950 X6 server. There are four socket
models for both of them. You physically can upgrade an x3850 X6 to an x3950 X6
server by replacing the mechanical enclosure that houses the CPU Books,
Storage Books, and I/O Books. The Books can be used in both server types.
166
Table 6-9 lists the technical details of all x3850 X6-based models for SAP HANA
when running SAP Business Suite on top.
Table 6-9 Overview of x3850 X6 workload-optimized models for Business Suite on SAP HANA
Model)
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
AC3-2S-128
2x Intel Xeon
E7-8880 v2
128 GB DDR3
(16x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
128 256
AC3-2S-256
2x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB or
16x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 384
256 512
2S 4S
AC3-2S-384
2x Intel Xeon
E7-8880 v2
384 GB DDR3
(48x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
2S 4S-512
AC3-2S-512
2x Intel Xeon
E7-8880 v2
512 GB DDR3
(32x 16 GB or
16x 32 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
2S 4S
AC3-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
AC3-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 768
512 1T
AC3-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
None
AC3-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
1T 1.5T
1T 2T
AC3-4S-1536
4x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(96x 16 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
None
AC3-4S-2048
4x Intel Xeon
E7-8880 v2
2 TB DDR3
(64x 32 GB or
32x 64 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB 10K SAS HDD)b
None
a. Alternative CPU types E7-4880 v2, E7-4890 v2 (both support no upgrade to eight sockets), and
E7-8890 v2 are available upon request. For more information, see CPU and memory on
page 112.
b. Additional drives are installed in the EXP2524 expansion unit.
167
Table 6-10 shows the technical details of all x3950 X6-based models for SAP
Business Suite powered by SAP HANA.
Table 6-10 Overview of x3950 X6 workload-optimized models for Business Suite on SAP HANA
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
AC4-4S-256
4x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
256 512
AC4-4S-512
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 768
512 1 TB
4S 8S
AC4-4S-768
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
4S 8S- 1 TB
AC4-4S-1024
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
1 TB 1.5 TB
1 TB 2 TB
4S 8S
AC4-4S-1536
4x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(96x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
4S 8S
AC4-4S-2048
4x Intel Xeon
E7-8880 v2
2 TB DDR3
(64x 32 GB or
32x 64 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
None
AC4-8S-512
8x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB 10K SAS HDD)
512 1 TB
AC4-8S-1024
8x Intel Xeon
E7-8880 v2
1 TB DDR3
(128x 8 GB or
64x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
1 TB 1.5 TB
1 TB 2 TB
AC4-8S-1536
8x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(192x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
None
AC4-8S-2048
8x Intel Xeon
E7-8880 v2
2 TB DDR3
(128x 16 GB or
64x 32 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB 10K SAS HDD)
2 TB 3 TB
2 TB 4 TB
AC4-8S-3072
8x Intel Xeon
E7-8880 v2
3 TB DDR3
(192x 16 GB)
19.2 TB
(6x 400 GB SAS SSD,
19x 1.2 TB 10K SAS HDD)b
None
168
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
AC4-8S-4096
8x Intel Xeon
E7-8880 v2
4 TB DDR3
(128x 32 GB or
64x 64 GB)
19.2 TB
(6x 400 GB SAS SSD,
19x 1.2 TB 10K SAS HDD)b
4 TB 6 TB
AC4-8S-6144
8x Intel Xeon
E7-8880 v2
6 TB DDR3
(192x 32 GB)
28.8 TB
(8x 400 GB SAS SSD,
28x 1.2 TB 10K SAS HDD)b
None
a. Alternative CPU type E7-8890 v2 is supported for enhanced performance. Available upon request.
b. Additional drives are installed in the EXP2524 expansion unit.
Upgrading the server requires downtime of the system. However, because of the
capability of GPFS to add storage capacity to an existing GPFS file system by
just adding devices, data that is on the system remains intact. Do a backup of the
data before changing the systems configuration.
All HA and DR solutions can be implemented using either GPFS based storage
replication or SAP HANA System Replication. You also can use the standby
nodes in HA and DR solutions to run a second SAP HANA instance for
non-productive purposes. For more information about these architectures, see
7.2, HA and DR for single-node SAP HANA on page 196.
Adding additional business continuity features after the initial implementation is
easily possible. Only additional hardware must be bought, and no parts must be
retired.
Scale-out SAP HANA environments are not supported at the time of writing when
running SAP Business Suite on top.
169
Processors
8
x3950 X6
Upgrade supported by
replacing mechanical chassis
4
x3850 X6
2
Some upgrades require to
replace memory modules with
higher capacity modules
Memory size
256 GB
384 GB
512 GB
768 GB
1 TB
1.5 TB
2 TB
Figure 6-9 Portfolio of all X6 based solutions for scale-out Business Warehouse
The lower half of the figure shows models that are based on the x3850 X6 server
and the upper half shows models that are based on the x3950 X6 server. Models
with four sockets exist using both servers. You physically can upgrade a x3850
X6 server to a x3950 X6 server by replacing the x3850 4U mechanical chassis
with an 8U version for the x3950 X6. All CPU Books, I/O Books, and Storage
Books can be reused in the 8U chassis.
Table 6-11 on page 171 lists the technical details of all workload-optimized x3850
X6 models in a scale-out configuration.
170
Table 6-11 Overview of x3850 X6 models for a scale-out SAP Business Warehouse
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
Maximum
cluster
size
AC3-2S-256-C
2x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB or
16x 16 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB SAS HDD)
2S
4S+512
4 nodes
AC3-4S-512-C
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB SAS HDDb)
512 768
512 1
TB
56 nodes
AC3-4S-768-C
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB or
48x 16 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB SAS HDDb )
768 1
TB
56 nodes
AC3-4S-1024-C
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
13.2 TB
(4x 400 GB SAS SSD,
13x 1.2 TB SAS HDDb )
56 nodes
a. Alternative CPU types E7-4880 v2, E7-4890 v2 (both support no upgrade to eight sockets), and
E7-8890 v2 are available upon request. For more information, seeCPU and memory on page 112.
b. Additional drives are installed in the EXP2524 expansion unit.
Table 6-12 lists the technical details of all workload-optimized x3950 X6 models
in a scale-out configuration.
Table 6-12 Overview of x3950 X6 models for scale-out SAP Business Warehouse
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
Maximum
cluster
size
AC4-4S-256-C
4x Intel Xeon
E7-8880 v2
256 GB DDR3
(32x 8 GB)
3.6 TB
(2x 400 GB SAS SSD,
4x 1.2 TB SAS HDD)
256 512
56 nodes
AC4-4S-512-C
4x Intel Xeon
E7-8880 v2
512 GB DDR3
(64x 8 GB or
32x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB SAS HDD)
512 768
512 1 TB
56 nodes
AC4-4S-768-C
4x Intel Xeon
E7-8880 v2
768 GB DDR3
(96x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB SAS HDD)
768 1 TB
4S 8S
56 nodes
AC4-4S-1024-C
4x Intel Xeon
E7-8880 v2
1 TB DDR3
(64x 16 GB or
32x 32 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB SAS HDD)
4S 8S
56 nodes
171
Model
CPUsa
Main
memory
Storage
(data and log)
Upgrade
options
Maximum
cluster
size
AC4-8S-1024-C
8x Intel Xeon
E7-8880 v2
1 TB DDR3
(128x 8 GB or
64x 16 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB SAS HDD)
1 TB
1.5 TB
1 TB
2 TB
56 nodes
AC4-8S-1536-C
8x Intel Xeon
E7-8880 v2
1.5 TB DDR3
(192x 8 GB)
9.6 TB
(4x 400 GB SAS SSD,
10x 1.2 TB SAS HDD)
1.5 TB
2 TB
56 nodes
AC4-8S-2048-C
8x Intel Xeon
E7-8880 v2
2 TB DDR3
(128x 16 GB
or 64x 32 GB)
19.2 TB
(6x 400 GB SAS SSD,
19x 1.2 TB SAS HDDb)
56 nodes
a. Alternative CPU type E7-8890 v2 is supported for enhanced performance. Available upon request.
b. Additional drives are installed in the EXP2524 expansion unit.
An SAP HANA cluster environment can consist of only building blocks with the
same memory size. Mixing different memory sizes in a cluster is not supported.
For example, when upgrading from a cluster with 512 GB nodes to 1 TB nodes,
you must add the additional memory to every node.
Note: The minimum number of nodes in any SAP HANA scale-out
environment is three worker nodes. A cluster of two worker nodes is not
supported. Adding one standby node for business continuity results in a
four-node cluster being the minimum that is supported.
The maximum number of nodes that is available for implementation is 56
(except for model AC3-2S-256-C, which is limited to four nodes). However,
IBM validates feasibility for up to 224 nodes in its labs to accommodate for
growing customer demand.
If you are interested in solutions beyond 56 nodes, contact your IBM
representative or email the IBM SAP International Competence Center at
[email protected].
172
173
174
Although the latest releases of SAP HANA support a change in the number of nodes during a
backup and restore procedure, this approach does not give the best performance. The change in
landscape is mimicked by running multiple index server processes on one target machine. Avoid
this scenario for optimal performance.
Note: At the time of writing, SAP required the exact same number of cores
and amount of memory on every eX5 and X6 model. Following the new
memory per core ratio on the X6 nodes is not supported in hybrid clusters.
Therefore, hybrid scale-out environments are supported by the following
models only:
Two sockets: 256 GB memory (T-shirt size S with AC3-2S-256-C)
Four sockets: 512 GB memory (M with AC3-4S-512-C or AC4-4S-512-C)
Eight sockets: 1 TB memory (L with AC4-8S-1024-C)
A hybrid SAP HANA scale-out environment must have more than 50% of the
nodes still being eX5 models. When the majority of the nodes are X6-based, all
eX5 servers must be excluded from the cluster. A hybrid cluster with an equal
number of eX5 and X6 nodes is not supported. Node designation (SAP HANA
worker or standby) is not important for this situation. As an example, if your
cluster has five worker and one standby nodes running on eX5 technology, you
can add up to five X6 building blocks to it.
This situation leads to different upgrade possibilities. Slow-growing environments
can add X6 nodes to their cluster without hitting the 50% rule anytime soon.
Faster growing environments must take a migration scenario into account.
175
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB X6
node
1TB X6
node
1TB X6
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB eX5
node
1TB X6
node
8 or 10 TB total memory
1TB eX5
node
2TB X6
node
2TB X6
node
2TB X6
node
2TB X6
node
2TB X6
node
Figure 6-10 Example transition of an eX5 based cluster into an X6 based cluster
The database then grows and the customer decides to add X6 models to prepare
for a migration at some point. Up to four X6 based workload-optimized solutions
can be added to the existing five-node cluster. The new X6 nodes must be eight
sockets and 1 TB memory each (see the middle part of Figure 6-10), which
allows the cluster to grow up to 9 TB total.
Adding additional X6 models breaks the 50% rule. It is the correct point to
exclude all eX5 models from the cluster, leaving the database instance on X6
nodes only. To accommodate for the missing 5 TB of the eX5 nodes, you add
additional memory to each of the X6 servers and use the new memory per core
ratio on X6. You add 1 TB per X6 node. Depending on the fill level of the
database before the split, you might need one additional X6 node (see the lower
part of Figure 6-10). The total memory size of the cluster is then 8 TB (four X6
nodes only) or 10 TB (with one new X6 node).
176
3
4
One SAP HANA system, as referred to in this section, can consist of one single server or multiple
servers in a clustered configuration.
See http://global12.sap.com/news-reader/index.epx?articleID=22775 and
http://www.saphana.com/community/blogs/blog/2014/05/06/at-the-seasside-with-sap-hana-on
-vmware
177
Additionally, SAP supports VMware vSphere 5.1 since the release of SAP HANA
1.0 SPS 05 (that is, revision 45 and higher) for non-production use only. Relaxed
configuration rules apply to non-production environments. Consolidation of
multiple SAP HANA VMs onto one physical server is allowed for non-production
environments and eight socket servers are supported as host systems.
VMware Tools is approved by SAP for installation inside the VM operating
system.
Table 6-13 summarizes the overview of VMware scenarios with SAP HANA.
Table 6-13 Overview of VMware scenarios with SAP HANA
Feature
Non-production
SAP HANA instance
Production
SAP HANA instance
SPS05 or later
SPS07 or later
vSphere 5.1
vSphere 5.5
Yes
Yes
Scale-out deployment
No
No
Multiple
CPU overprovisioning
No
No
Memory overprovisioning
No
No
178
10
11
12
13
14
15
Figure 6-11 SAP HANA possible slots per T-shirt size (x = reserved) on eX5 servers
179
Table 6-14 shows the configuration parameters for virtual machines from slot
sizes 1 - 8. Sizes 7 and 8 are not supported by VMware because of the restriction
of a maximum number of vCPUs per guest of 64 in their latest VMware ESX
server software. You can install one or more non-production virtual machines of
slot sizes 1 - 6 on any x3950 X5 workload-optimized solution for SAP HANA
appliance.
Table 6-14 SAP HANA virtual machine sizes when virtualizing eX5 models
SAP
T-shirt
SAP HANA
support pack
IBM
name
vCPUs
(HT on)
Virtual
memory
Required
no. of slots
Total
HDD
Total
SSD
XXS
SPS 05
VM1
10
64 GB
352 GB
64 GB
XS
SPS 05
VM2
20
128 GB
608 GB
128 GB
None
Manually
VM3
30
192 GB
864 GB
192 GB
SPS 05
VM4
40
256 GB
1120 GB
256 GB
None
Manually
VM5
50
320 GB
1376 GB
320 GB
None
Manually
VM6
60
384 GB
1632 GB
384 GB
None
N/A
VM7a
70
448 GB
1888 GB
448 GB
N/A
VM8a
80
512 GB
2144 GB
512 GB
a. This slot size is not possible due to limitations of the VMware ESXi 5 hypervisor.
For production environments on vSphere on eX5, only one single virtual machine
is allowed per physical server.
180
For a complete overview of the available editions and their feature sets, see
http://www.vmware.com/files/pdf/vsphere_pricing.pdf.
181
Table 6-15 lists the order numbers for both the Enterprise and Enterprise Plus
editions of VMware vSphere 5.
Table 6-15 VMware vSphere 5 ordering P/Ns
Description
VMware P/N
License / Subscription
IBM P/N
VS5-ENT-C /
VS5-ENT-PSUB-C
4817SE3
VS5-ENT-PL-C /
VS5-ENT-PL-PSUB-C
4817SE5
VS5-ENT-C /
VS5-ENT-3PSUB-C
4817TE3
VS5-ENT-PL-C /
VS5-ENT-PL-3PSUB-C
4817TE5
VS5-ENT-C /
VS5-ENT-5PSUB-C
4817UE3
VS5-ENT-PL-C /
VS5-ENT-PL-5PSUB-C
4817UE5
As an example, if you want to deploy SAP HANA on an eX5 M building block (for
example, 7143-HBx), will need four licenses. An X6 model AC48S1024 needs
eight licenses.
Sizing
The sizing is the same for virtualized and non-virtualized SAP HANA
deployments. Although there is a small performance impact because of the
virtualization, the database size and the required memory size are not affected.
Support
As with any other deployment type of SAP HANA, clients are asked to open an
SAP support ticket by using the integrated support model that is outlined in 8.7.1,
IBM and SAP integrated support on page 252. Any non-SAP related issue is
routed to VMware first, and it eventually is forwarded to the hardware partner. In
certain, but rare situations, SAP or its partners might need to reproduce the
workload on bare metal.
182
183
184
Side-by-side accelerators
This use case has the following attributes:
SAP HANA serves as a secondary database for SAP Business Suite
applications.
Data is replicated from the SAP application into the in-memory SAP HANA
appliance, for example, by using the SAP LT Replicator server.
The SAP Business Suite application is accelerated by retrieving results of
complex calculations on mass data directly from the in-memory database.
The user interface for users remains unchanged to ensure nondisruptive
acceleration.
Additional analytics that are based on the replicated data in SAP HANA can
provide new insights for users.
185
The difference to the reporting scenario is that the consumer of the data that
is replicated to SAP HANA is not a BI tool, but the source system itself.
Also for this scenario, several SAP RDSs are provided by SAP.
In release R1.4, IBM will support all existing SAP HANA use cases, including
SAP BW running on SAP HANA and SAP Business Suite running on SAP
HANA. All certified IBM configurations will be supported, including both
single-node and scale-out configurations with high availability (HA) features.
Options to share one appliance (MCOD, MCOS, or MCOC), which are described
in 6.5, Sharing an SAP HANA system on page 183, will be supported as part of
the offering.
Virtualization and disaster recovery (DR) features will be supported as part of
subsequent releases and are not available in release R1.4.
186
Chapter 7.
187
188
This publication focuses on SAP HANA, that is, the database layer only. For that reason, this
chapter does not describe RCO any further.
189
Technology
RTO
RPO
Suitable for
HA
Suitable for
DR
Infrastructure
level
GPFS based
synchronous
replication
HA: Zero
DR: Minutes
Zero
Yes
Yes
GPFS based
asynchronous
replicationa
N/A
N/A
N/A
N/A
Backup - Restore
Usually hours
Hours to days
No
Yes
SSR
(synchronous)
Minutes
Zero
Limitedb
Yes
SSR
(asynchronous)
Minutes
Seconds
No
Yes
Application
level
190
Note: SSR does not support automatic failover to a standby system. As the
name implies, it only replicates data to a standby system. Manual intervention
is required for the failover to happen.
SAP uses the term Host Auto-Failover to describe the capability of an
automatic failover to a standby system. At the time of writing, Host
Auto-Failover is not available with the SAP HANA product.
The next two sections explain GPFS based storage replication and SSR in more
detail.
The rest of this chapter then describes all supported scenarios for HA and DR of
single-node and scale-out SAP HANA environments that are based on IBM
System x solutions. Conventional backup methods are described in 7.4, Backup
and restore on page 236.
191
192
In GPFS terminology, each data copy is referred to as a replica. This term also applies to the
primary data, which is called the first replica. This term indicates that all data copies are equal.
Among other criteria, the following criteria must be met to enable the SSR
feature:
The SAP HANA software revision of the target environment must be the same
or higher than the source environment.
Both systems must use the same SID and instance numbers.
On both systems, instance number + 1" must be free because it is used for
replication purposes.
Replication modes
SSR can be set to one of three modes:
Synchronous mode: Makes the primary site system wait until the change is
committed and persisted on the secondary site system.
Synchronous in-memory mode: Makes the primary site system acknowledge
the change after it is committed in main memory on the secondary site
system, but not yet persisted on disk.
Asynchronous mode (available since SAP HANA SPS06): Makes the primary
site system commit transaction when the replicated log is sent to the DR site.
The system does not wait for an acknowledgment from the remote site.
Asynchronous replication allows for greater distances because a high latency
between the primary and secondary system does not prevent a production
workload from running at maximum performance as with synchronous
replication.
In both synchronous modes, the impact is defined by the transmission time from
the primary to its corresponding secondary system process. When running SSR
in synchronous mode, you must add the time that it takes to persist the change
on a disk in addition to the transmission delay.
In case the connection between the two data centers is lost, live-replication
stops. Then, after a (configurable) timer expires on the primary site system, it
resumes work without replication.
When the connection is restored, the secondary site system requests a delta
snapshot of what changes were done since the connection was lost. Live
replication can then continue after this delta is received on the secondary site
system.
193
System failover
If there is a failover to the secondary system (system takeover), manual
intervention is required to change the secondary site system from recovery mode
to active mode. SAP HANA automatically loads all row-based tables into memory
and rebuilds the row store indexes. In the next step, all logs since the last
received snapshot are replayed. After this step finishes, the system can accept
incoming database connections. A restart of the SAP HANA database instance is
not required.
An optional feature (called table preload) enables the primary system to share
information about which columns are loaded into main memory. The secondary
system can use this information to preload those tables in main memory.
Preloading reduces the duration of a system takeover operation.
194
asynchronous
replication
synchronous
replication
SAP HANA
node01
node02
DB partition 1
- SAP HANA DB
Worker node
- Index server
- Statistic server
- SAP HANA studio
node01
DB partition 2
DB partition 1
- SAP HANA DB
Standby node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- Index server
- Statistic server
- Index server
- Statistic server
- SAP HANA studio
Flash
HDD
Flash
data01
log01
data02
log02
node02
node01
node03
DB partition 1
DB partition 2
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- Index server
- Statistic server
- Index server
- Statistic server
- SAP HANA studio
Flash
Primary data
HDD
Flash
HDD
Flash
data01
log01
data02
log02
node03
DB partition 2
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- Index server
- Statistic server
node02
SAP HANA DB
SAP HANA DB
- SAP HANA DB
Worker node
HDD
SAP HANA
SAP HANA
node03
SAP HANA DB
HDD
Flash
Replica
Primary data
HDD
Flash
HDD
Flash
data01
log01
data02
log02
HDD
Flash
Replica
195
Attention: When talking about latency, make sure to specify the layer at which
you are measuring it. Network engineers talk about network latency, but SAP
prefers to use application latency.
196
Table 7-2 Overview of HA and DR options for single-node SAP HANA solutions
Characteristic
Stretched
HA
DR
(using
GPFS)
DR
(using
SSR)
HA and DR
(using
GPFS only)
HA and DR
(using
GPFS and
SSR)
Required data
centers
2 or 3
(metro
distance)
2 or 3
(metro
distance)
2
(metro
distance or
higher)
2 or 3
metro
distance)
2 or 3
(metro
distance or
higher)
RTO
Seconds
Seconds
Minutes
Minutes
Seconds,
for HA,
minutes for
DR
Seconds,
for HA,
minutes for
DR
RPO
Zero
Zero
Zero
Zero or
higher
Zero
Zero or
higher
Replication
method
GPFS
(sync)
GPFS
(sync)
GPFS
(sync)
SAP HANA
(sync or
async)
GPFS
(sync)
GPFS
(sync) plus
SSR (sync
or async)
Automatic
failover
Yes
Yes
No
No
Yes, for HA
node
Yes, for HA
node
Can host
non-production
No
No
Yes
Yes
Yes
Yes
Number of SAP
HANA server
nodes
Number of
GPFS quorum
servers
Tolerated node
failures
The following sections describe each of the solutions in more detail and show
architectural diagrams for each of them.
197
198
The quorum node is not running any SAP HANA processes. It runs only a small
GPFS process, so it does not need connectivity to the SAP HANA network. Only
GPFS access is required. Quorum functionality can be implemented on different
servers (you do not need to use an IBM SAP HANA building block for this task).
In most cases, a smaller system like the IBM System x3550 M4 is suitable for this
task. Without this quorum node, if only the communication between the SAP
HANA nodes is interrupted but otherwise the nodes are running fine, GPFS
cannot tell which side should continue to operate, and it unmounts the file system
on both nodes to protect from inconsistent data (this situation is called a split
brain). When GPFS is down, SAP HANA cannot continue to operate because
data and logs are not accessible. Administrative intervention is required in that
case.
Figure 7-2 shows the detailed conceptual view of a single-node HA solution.
While the GPFS quorum node is part of the GPFS cluster, it does not contribute
disk capacity to it. For that reason, in this example, we extend the shared file
system only half into the quorum node.
node02
node01
quorum node
SAP HANA DB
Worker
Stand-by
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
First replica
data01 + log01
Second replica
data01 + log01
local storage
local storage
199
For a fully redundant solution, the network also must be built using two redundant
Ethernet switches. Figure 7-3 shows the network setup of such a single-node HA
installation.
Worker Node
Standby Node
Quorum Node
GPFS Links
SAP HANA Links
Inter-Switch Link (ISL)
G8264 switches
200
This solution requires two identical SAP HANA building blocks plus a GPFS
quorum node. One SAP HANA building block is installed at each site. The node
at the primary site is installed as a worker node and the node at the secondary
site is installed as a standby node.
GPFS ensures that all data that is written on the active node is replicated over to
the server in the second data center, which is running SAP HANA in standby
mode.
GPFS needs a quorum node to act as tie-breaker if the network between the
servers breaks. This quorum node should be in a separate third location to
ensure maximum reliability of the solution.
Figure 7-4 outlines the conceptual view of a stretched HA implementation for a
single-node SAP HANA solution. The GPFS quorum node is placed at a
dedicated third site C.
Site C
Site A
Site B
quorum node
node01
node02
SAP HANA DB
Worker
Stand-by
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
First replica
data01 + log01
data01 + log01
local storage
Second replica
local storage
201
The GPFS quorum node does not contribute disk space to the file system; it is
connected to the GPFS network only to decide on the surviving site in split-brain
situations. This is why the yellow box only partially spans to the quorum node.
The network view that is outlined in Figure 7-5 shows that the quorum node
needs to be connected only to the GPFS network.
Site C
Site B
Quorum Node
Worker Node
GPFS Links
Standby Node
G8264 switches
Figure 7-5 Network setup of single node with stretched HA (three-site approach with dual inter-site links)
202
Site C
Site B
Quorum Node
Worker Node
Standby Node
MPLS
or other campus/metro switching
GPFS Links
G8264 switches
Figure 7-6 Network setup of single node with stretched HA (three-site approach with one inter-site link)
If a separate third location is not available, the quorum node must be placed in
the primary data center with the active worker node. This gives the active site a
higher weight so that GPFS continues to operate even if the server hosting the
SAP HANA standby process at the second site is no longer reachable. The
standby server can become unreachable, for example, because of a server
hardware failure or because of a broken link between the two data centers.
203
Figure 7-7 shows a stretched HA solution with the quorum node in the primary
site A together with the active worker node.
Site A
quorum node
Site B
node02
node01
SAP HANA DB
Worker
Stand-by
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
First replica
data01 + log01
Second replica
data01 + log01
local storage
local storage
Site B
Worker Node
Standby Node
Quorum Node
GPFS Links
SAP HANA Links
Inter-Switch Link (ISL)
G8264 switches
Figure 7-8 Network setup of single node with stretched HA (two-site approach with dual inter-site links)
204
This scenario assumes that the customer has two links between the primary and
secondary sites that are available for SAP HANA usage. If not, then a
consolidated single network link can also be used, as shown in Figure 7-9.
Site B
Worker Node
Standby Node
Quorum Node
GPFS Links
SAP HANA Links
MPLS
or other
campus/metro
switching
G8264 switches
Figure 7-9 Network setup of single node with stretched HA (two-site approach with one inter-site link)
The disadvantage of having the GPFS quorum node in the primary data center
and not in a separate third location comes into play only if the primary data
center experiences an outage.
If the GPFS quorum node is at a third independent site, then it can constitute,
together with the standby node at site B, the majority of the cluster nodes, so the
shared file systems stays up at the standby node and SAP HANA can
automatically fail over.
If the GPFS quorum node is in the same failure domain as the SAP HANA worker
node, then upon a failure in this domain, the file system on the standby node
unmounts and manual intervention is required to bring it back up. In this situation,
the GPFS cluster loses the majority of the nodes. Automatic failover is still
possible in any failure situation that does not affect the network subsystem. If the
quorum node can still communicate to the standby node, automatic failover
works. This situation protects against any failure of the worker node, such as
hardware failures, firmware faults, or operating system and user errors.
205
Only if there is a total site outage (when the SAP HANA worker node and the
GPFS quorum node are unavailable) is manual intervention required on site B to
bring the database instance back up.
Because the node at site B is not idling (it runs SAP HANA standby processes
that monitor the worker node), it is not possible to host non-production database
instances on this node.
206
Figure 7-10 shows a single-node solution for DR using GPFS based storage
replication.
Site C
Site A
Site B
quorum node
node01
node02
SAP HANA DB
SAP HANA DB
Worker
Worker
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- SAP HANA Studio
First replica
data01 + log01
data01 + log01
local storage
Second replica
local storage
Figure 7-10 Detailed view of single-node DR solution using GPFS (three-site approach)
207
The network can be implemented in several different ways. The basic version is
shown in Figure 7-11, with fully redundant network connections towards both
satellite sites B and C.
Site B
Site C
Worker Node
DR Node
Quorum Node
GPFS Links
G8264 switches
Figure 7-11 Network setup of a single node with DR installation (three-site approach with dual links)
208
Site B
Site C
Worker Node
DR Node
Quorum Node
MPLS
or other
campus/metro
switching
GPFS Links
SAP HANA Links
Inter-Switch Link (ISL)
G8264 switches
(usually MPLS LER or
other dist switches)
Figure 7-12 Network setup of a single node with DR (three-site approach with alternative inter-site
connectivity)
209
In the absence of a third site that hosts the GPFS quorum server this machine
must be installed at the primary site A, as shown in Figure 7-13, to ensure
continuous availability of the file system if the link between sites A and B is
broken.
Site A
quorum node
Site B
node01
node02
SAP HANA DB
SAP HANA DB
Worker
Worker
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- SAP HANA Studio
First replica
data01 + log01
data01 + log01
local storage
Second replica
local storage
Figure 7-13 Detailed view of single-node DR solution using GPFS (two site approach)
Similar to the approach with three data centers, there is more than one way of
how networking can be implemented with two data centers depending on
customer requirements. Figure 7-14 on page 211 shows the fully redundant
networking architecture with two links between site A and B.
210
Site B
Storage expansion for
non-prod DB instance
Worker Node
DR Node
Quorum Node
GPFS Links
SAP HANA Links
Inter-Switch Link (ISL)
G8264 switches
Figure 7-14 Network setup of single node with DR installation (two-site approach with dual links)
Different approaches, similar to what is shown in Figure 7-12 on page 209, can
be implemented for a two-site solution. Details must be agreed upon by the
client.
During normal operations, SAP HANA is running only on site As node. No SAP
HANA processes are running on the DR node at site B. The only GPFS
processes that are running are the ones that receive replicated data from the
active worker node from site A.
If the worker node at site A goes down, the file system stays up on the DR node if
the GPFS quorum node is still healthy and available. If the quorum node also
goes down, then GPFS unmounts the file system on the DR node because the
majority of the GPFS cluster nodes are down. If this happens, human
intervention is required. An administrator manually can override GPFS cluster
logic and mount the file system just on the DR node by using the second replica.
After the file system is available on the DR node, SAP HANA processes can be
started that load database data and replay log files so that the database instance
can resume operation with no data loss (RPO of zero).
Because during normal operation there are no SAP HANA processes running on
the secondary node (not even in standby mode), there is an option to host a
non-productive SAP HANA instance (like development or test) there. In such a
scenario, if there is a disaster, this non-productive SAP HANA instance must be
shut down before a takeover can happen from the primary system.
211
This additional SAP HANA instance needs its own storage space for persistency
and logs. IBM uses the EXP2524 unit to provide this additional space. The
EXP2524 adds up to twenty-four 2.5-inch HDDs that are connected directly to
the server through an SAS interface. A second file system is created over those
drives for the additional SAP HANA instance. This second file system is visible
only to this node. Data is not replicated to a second node.
If a disaster occurs and the productive SAP HANA instance is switched to run on
the DR server at site B, the non-productive instance must be shut down first.
After the primary server at site A is repaired, customers can choose to either
switch back their productive instance to site A or they can let it continue to run on
the DR node at site B.
If customers choose to keep their production instance on site B, this means that
the non-production instance must now be started on site A, that is, the former
primary server. For that reason, the site A server must have an EXP2524
attached as well (not just site Bs machine). Non-production data on the
EXP2524 must be copied manually to the other side before you can start the
non-production SAP HANA instance.
212
Site A
Site B
node01
node01
SAP HANA DB
SAP HANA DB
Worker
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
First replica
Worker
SAP HANA
System
Replication
- SAP HANA DB
- Index server
- Statistic server
- SAP HANA studio
data01 + log01
data01 + log01
local storage
First replica
local storage
Figure 7-15 Detailed view of single-node DR solution using SAP HANA System
Replication
A failover to site B always requires manual intervention and does not happen
automatically.
SSR can be configured to work either in synchronous or asynchronous mode.
Different physical distances can be realized in each mode. For more information,
see Replication modes on page 193.
213
Different architectures are possible for the network connecting the two SAP
HANA building blocks. Figure 7-16 shows a setup with redundant switches for the
SAP HANA replication network.
Site B
Storage expansion for
non-prod DB instance
Worker Node
DR Node
G8264 switches
Figure 7-16 Network setup of single node with SAP HANA System Replication (dual inter-site links)
Because SAP HANA processes are not in an active standby mode on the DR
node, it is possible to host a non-production instance by using dedicated
additional storage. IBM uses the EXP2524 unit to provide this additional space.
The EXP2524 adds up to twenty-four 2.5-inch HDDs that are directly connected
to the server through an SAS interface. A second file system is created over
those drives for the additional SAP HANA instance. This second file system is
visible only to this node. Data is not replicated to a second node.
Designs without the network switches are also supported. Figure 7-17 on
page 215 shows such an approach. It is possible to have these inter-site links
going across VPNs (or other forms of overlay networks) if the latency and
throughput requirements are met.
214
Site B
Storage expansion for
non-prod DB instance
Worker Node
DR Node
Figure 7-17 Network setup of single node with SAP HANA System Replication (no
switches)
215
Figure 7-18 shows the architectural view of such a single node with a HA and DR
solution. No additional GPFS quorum node is required because the configuration
includes three nodes running GPFS processes. No split-brain situation can occur
in a cluster with three members.
Site A
Site B
node02
node01
SAP HANA DB
Worker
node03
SAP HANA DB
Stand-by
Worker
- SAP HANA DB
Worker node
- SAP HANA DB
Standby node
- SAP HANA DB
Worker node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
- Index server
- Statistic server
- SAP HANA studio
First replica
data01 + log01
Second replica
data01 + log01
data01 + log01
local storage
local storage
Third replica
local storage
During normal operation, when the setup is healthy, SAP HANA is running only
on the nodes at the primary data center. From these two nodes, only the worker
node responds to user queries. The HA node does not respond to requests. The
DR node at site B has no active SAP HANA processes. Only GPFS processes
are running that receive data (third replica) and write it to local disk.
If the primary site worker node experiences an outage, the standby node detects
it and SAP HANA fails over and resumes operation on this node. This node is
promoted from a standby to a worker designation and responds to user requests.
Now, GPFS has only two replicas that are left in the cluster. When s the first node
is repaired, it rejoins the cluster and its local replica is restored.
216
If there is a disaster where both nodes at the primary site are lost, manual
intervention is required to get SAP HANA running at the DR site B. Because
GPFS has lost the majority of its cluster nodes, it must be told that it is safe to
operate using only the surviving node (this temporary procedure is called
relaxing the node quorum). The file system can then be mounted again and SAP
HANA can be started. It reads data and log files from local disk, loads it into
memory, and can then resume database operation. Because GPFS replication
happens synchronously, no data is lost, and the DR node continues exactly with
the last committed transaction before the disaster happened.
Figure 7-19 shows the networking architecture with redundant Ethernet switches.
Site B
Storage expansion for
non-prod DB instance
Worker Node
Standby Node
DR Node
GPFS Links
SAP HANA Links
G8264 switches
Figure 7-19 Network setup of a single node with HA plus DR using GPFS
Because during normal operation the node at the secondary site is not hosting
any running instance, there is an option to host a non-productive SAP HANA
instance (such as development or test) on this DR site node. If there is a disaster,
this non-productive SAP HANA instance must be stopped before the production
database instance can be started on this node.
This additional SAP HANA instance needs its own storage space for persistency
and logs. IBM uses the EXP2524 unit to provide this additional space. The
EXP2524 adds up to twenty-four 2.5-inch HDDs that are directly connected to
the server through an SAS interface. A second file system is created over those
drives for the additional SAP HANA instance. This second file system is visible
only to this node and is not replicated to another node.
217
Site A
node02
node01
Site B
quorum node
node01
SAP HANA DB
SAP HANA DB
Worker
- SAP HANA DB
Standby node
- Index server
- Statistic server
- SAP HANA studio
- Index server
- Statistic server
First replica
Worker
Stand-by
- SAP HANA DB
Worker node
- SAP HANA DB
SAP HANA
System
Replication
- Index server
- Statistic server
- SAP HANA studio
Shared file system - GPFS
data01 + log01
data01 + log01
Second replica
First replica
data01 + log01
local storage
local storage
local storage
Figure 7-20 Detailed view of single-node solution with HA using GPFS and DR using SSR
At any point, there are two data replicas available at site A. Synchronous GPFS
replication ensures that I/O operations are marked as successfully complete if
and the operation is written to disk on node01 and node 02 of site A.
218
Depending on the mode in which SSR runs, the third data copy on site B either is
a synchronous or an asynchronous copy. If it is configured to be a synchronous
copy, then restrictions on the distance of the DR site apply. If SSR runs in
asynchronous mode, then longer distances can be covered (as described in
7.1.3, Special considerations for DR and long-distance HA setups on
page 195). Data replication to the DR site happens on SAP HANA level. The DR
node feeds GPFS with the incoming replication data. GPFS stores this data on
its local disk (indicated by first replica on the right in Figure 7-20 on page 218).
The networking architecture for such a single-node setup is shown in
Figure 7-21. Redundant switches are used at both sites. At the DR site B, there is
no need to connect the GPFS interfaces to the Ethernet switch because from a
GPFS perspective the DR node is creating its own single node GPFS cluster.
This server is not a member of site As GPFS cluster.
Quorum Node
Site B
Storage expansion for
non-prod DB instance
Worker Node
Standby Node
DR Node
GPFS Links
SAP HANA Links
G8264 switches
Figure 7-21 Network setup of single-node HA (using GPFS) plus DR (using SSR)
219
The DR node does not need GPFS connectivity to site A, which allows you to
leave out the Ethernet switches at site B and connect the node directly to the
primary site switches. This setup is shown in Figure 7-22.
Depending on the type of inter-site link that connects the primary and secondary
data centers, different scenarios are possible. For example, if only one inter-site
link is available, only one network interface of the DR node can be cabled. The
different options and their implications to redundancy and availability must be
discussed with the customer.
During normal operation, the node at the DR site B is running in recovery mode,
which allows you to host a non-productive SAP HANA instance (such as
development or test) on this idling node. If there is a disaster, this non-productive
SAP HANA instance must be stopped before the production database instance
can be started on this node.
Quorum Node
Site B
Storage expansion for
non-prod DB instance
Worker Node
Standby Node
DR Node
GPFS Links
SAP HANA Links
G8264 switches
Figure 7-22 Network setup of single-node HA (using GPFS) plus DR (using SSR) (fewer switches at site B)
This additional SAP HANA instance needs its own storage space for persistency
and logs. IBM uses the EXP2524 unit to provide this additional space. The
EXP2524 adds up to twenty-four 2.5-inch HDDs that are directly connected to
the server through an SAS interface. A second file system is created over those
drives for the additional SAP HANA instance. This second file system is visible
only to this node and is not replicated to another node.
220
HA and DR
(using GPFS)
HA and DR
(using GPFS and
SSR)
2 or 3a
Geographical distance
Not applicable
Metro distance
Metro distance or
higher
RTO
Seconds
RPO
Zero
Zero
Zero or higher
Replication method
GPFS
(synchronous)
GPFS
(synchronous)
GPFS (synchronous)
plus SSR
(synchronous or
asynchronous)
Automatic HA failover
Yes
Yes
Yes
Automatic DR failover
Not applicable
Yes
No
Can host
non-production
No
Yes, at DR site
Yes, at DR site
a. A third data center is required only for automatic DR failover. If no third site is
available, manual failover can be implemented instead.
221
222
Basic architecture
During normal operation, there is one active SAP HANA instance running. The
SAP HANA instance on the secondary site is not active. The architecture on
each site is identical to a standard scale-out cluster with HA, as described in
7.3.1, High availability using GPFS storage replication on page 222. The
architecture must include standby servers for HA. A server failure is handled
completely within one site and does not enforce a site failover. Figure 7-23 shows
this setup.
Site A
node01
node02
node03
Site B
node05
node04
Partition 2
Partition 3
node06
node07
node08
Standby
Partition 2
Partition 3
Standby
synchronous
replication
Third replica
GPFS
quorum
node
Site C
Figure 7-23 Basic setup of the disaster recovery solution using GPFS synchronous replication
223
The connection between the two main sites, A and B, depends on the clients
network infrastructure. Use a dual link dark Fibre Connection to allow for
redundancy in the network switch at each site. For full redundancy, an additional
link pair is required to fully mesh the four switches. Figure 7-24 shows a
connection with one link pair in between. It also shows that only the GPFS
network must span both sites because they make up one single stretched GPFS
cluster. The SAP HANA internal network is kept within each site because SAP
HANA does not need to communicate to the other site.
Site A
Site B
GPFS VLAN
ISL
GPFS VLAN
GPFS VLAN
GPFS VLAN
ISL
Figure 7-24 Networking details for SAP HANA GPFS based DR solutions
Within each site, the 10 Gb Ethernet network connections for both the internal
SAP HANA and the internal GPFS network are implemented in a redundant
layout.
Depending on where exactly the demarcation point is between the SAP HANA
installation, the client network, and the inter-site link, different architectures can
be used. In the preferred case, there is a dedicated 10 Gb Ethernet connection
going out of each of the two SAP HANA switches at each site towards the
demarcation point. There is no requirement with which technology the data
centers are connected if the technology can route IP traffic across the link. In
general, low-latency interconnect technologies are the preferred choice. For
more information about latency, see 7.1.3, Special considerations for DR and
long-distance HA setups on page 195.
Depending on the client infrastructure, the inter-site link might be the weakest link
and no full 10 Gb Ethernet can be carried across it. SAP has validated the IBM
solution by using a 10 Gb interconnect, but depending on the workload, a HANA
cluster might not generate that much traffic and a value much smaller than 10 Gb
is sufficient. This choice must be decided upon by each individual client. For
example, the initial database load might take hours using a 1 Gbit connection, or
minutes when using a 10 Gbit network connection to the remote site. During
normal operation, latency is more critical than bandwidth for the overall
application performance.
224
225
Depending on the GPFS configuration, either three, five, or six copies are kept
up-to-date.3 Most valid file system descriptors are required for the file system to
be accessible.4 If disks fail over time, GPFS updates another copy on different
disks to ensure that all copies are alive. If multiple disks fail concurrently and
each was holding a valid copy of the file system descriptor, then a cluster loses
the file system descriptor quorum and the file system automatically is
unmounted.
Site failover
During normal operation, there is a running SAP HANA instance active on the
primary site. The secondary site has an installed SAP HANA instance that is
inactive. A failover to the remote SAP HANA installation is a manual procedure;
however, it is possible to automate the steps in a script. Depending on the reason
for the site failover, you can decide whether the secondary site becomes the new
production site or a failback must happen after the error in the primary site is
fixed.
To ensure the highest level of safety, during normal operation the GPFS file
system is not mounted on the secondary site. This action ensures that there is no
read nor write access to the file system. If you want, however, the file system can
be mounted, for example, read-only, to allow for backup operations on the DR
site.
A failover is defined as anything that brings the primary site down. Single errors
are handled within the site by using the fully redundant local hardware (such as a
spare HA node and second network interface).
Events that are handled within the primary site include the following ones:
Single-server outage
Single-switch outage
Accidentally pulled network cable
Local disk failure
Events that cause a failover to the DR site include the following ones:
Power outage in the primary data center causing all nodes to be down
Two servers going down at the same time (not necessarily because of the
same problem)
3
4
226
For environments with just one disk, GPFS uses only one file system descriptor. This scenario does
not apply to SAP HANA setups.
Even if all disks fail, if there is at least one valid file system descriptor that is accessible, you still
have a chance to recover data manually from the file system.
The time for the failover procedure depends on how long it takes to open SAP
HANA on the backup site. The data already is in the data center and ready to be
used. Any switch from one site to the other includes a downtime of SAP HANA
operations because the two independent instances on either site must not run
concurrently because of the sharing of the persistency and log files on the file
system.
GPFS provides the means to restore HA capabilities on the backup site. During
normal operation, only one replica is pushed to the backup site. However, clients
that consider their data centers to be equal might choose to not gracefully fail
back to the primary data center after it is repaired, but instead continue to run
production from the backup site. For this scenario, a GPFS restripe is triggered
that creates a second replica on the backup site out of the one available replica.
This procedure is called restoring HA capabilities. The exact commands are
documented in the IBM SAP HANA Operations Guide.5 The duration of this
restriping depends on the amount of data in the file system.
When the SAP HANA instance is started, the data is loaded into main memory.
The SAP HANA database is restored to the latest savepoint and the available
logs are recovered. This procedure can be automated, but it is dependent on the
client environment. The commands in the Operations Guide provide a template
for such automation.
Clients choosing to continue running production out of the former backup data
center can easily add the former primary site after it is restored. The nodes are
integrated back into the GPFS cluster and a resynchronization of the most recent
version of the data occurs between the new primary data center and the new
secondary data center. One replica is held in the new secondary site. The overall
picture looks exactly like before a failover only with the data centers having
switched their designation.
Site failback
A site failback is defined as a graceful switch of the production SAP HANA
instance from the secondary data center back to the primary data center.
To understand the procedure for a failback, it is important to know the initial state
of the DR environment. There are two possibilities:
You have two replicas that are local in the data center and one replica in the
remote sites data center. Here are some examples:
A disaster happened and you restored HA on the backup site; now you are
ready to fail back production to the primary site again.
227
During normal operations, you want to switch production to the backup site
for maintenance reasons.
You have one replica that is local in the data center and two replicas are in the
remote sites data center. For example, a disaster happened and you are
running production from the backup data center without having HA restored.
Environments with only one working replica must restore HA first before being
able to fail back gracefully, which ensures the highest level of safety for your data
during the failback procedure.
When SAP HANA is running from a file system with two local replicas, the
failback procedure is identical to a controlled failover procedure. The data center
is assumed to be down, the active site now is the remote site, HA is restored
(using GPFS restriping), and the second site is attached again with one single
replica of the data. SAP HANA can be started when it is shut down on the other
site, but it experiences a performance impact during HA restore (that is, GPFS
restriping). GPFS restriping is an I/O-heavy operation.
Site C
Worker
Node
Worker
Node
Worker
Node
Site B
DR
Node
Standby
Node
DR
Node
DR
Node
DR
Node
Quorum
Node
GPFS Links
SAP HANA Links
Inter-Switch Link (ISL)
G8264 switches
Figure 7-25 Network setup of GPFS based DR for a scale-out system with a dedicated quorum node
228
The quorum node is configured to act as a GPFS quorum server without storing
any SAP HANA data. The only requirement is to have a small disk or partition
available that can hold a file system descriptor. The quorum node needs network
access to only the GPFS network. SAP HANA network access is not required.
In a standard setup, two nodes from the primary data center act as quorum
nodes, two nodes from the secondary data center act as quorum nodes, plus the
additional quorum node at the third site. The number of quorum nodes is shown
as the weight of a data center in Figure 7-26.
Primary site
Secondary site
Weight = 2
Weight = 2
Tertiary site
GPFS quorum node only
Weight = 1
Figure 7-26 Outline of a three-site environment with a dedicated GPFS quorum node
In the unlikely event that two inter-site links are interrupted at the same time,
there is still the majority of quorum nodes that are available to communicate with
each other over the one connection that is left to decide from which data center
to run production. In terms of weight, this means that in any of these situations a
minimum of three can be ensured to always be up.
If the links between the primary and secondary and between the secondary and
tertiary data centers go down, SAP HANA keeps running out of the primary site
without any downtime. If the links between the primary and tertiary and between
the secondary and tertiary data centers go down, the primary and secondary
data center can still communicate and SAP HANA keeps running out of the
primary data center.
229
If the links between the primary and secondary and the primary and tertiary data
centers go down, it means that the production SAP HANA instance in the primary
data center is isolated and loses GPFS quorum. As a safety measure, GPFS
prevents any writing to the file system on the primary site and SAP HANA stops.
GPFS stays up and running on the secondary site because the quorum node still
has a connection to it. Depending on client requirements, HA must be restored
first before SAP HANA can be started and production continues to run out of the
secondary site data center.
It is a valid use case to set up an existing server to act as a GPFS quorum node
for the SAP HANA DR installation.6 GPFS must have root access on this
machine to run. Three GPFS server licenses are required on the primary and on
the secondary sites for the first three servers.7 Additional servers need GPFS
FPO licenses.
The main advantage of having a dedicated quorum node is that the file system
always is available during failover and failback without any manual intervention if
there is at least one site and the quorum nodes can communicate with each
other.
6
7
230
The applicability of this statement must be verified for each installation by IBM.
The currently validated minimum number of servers on each site for DR is three. This is required by
SAP to set up a scale-out environment with HA. Hence, the requirement is for at least three GPFS
server licenses per site.
Site B
Worker
Node
Worker
Node
Worker
Node
DR
Node
Standby
Node
DR
Node
DR
Node
DR
Node
GPFS Links
SAP HANA Links
G8264 switches
Figure 7-27 Network setup of GPFS based DR for a scale-out system without a dedicated quorum node
A dedicated quorum node at a third site allows the file system to remain
accessible even if the primary or secondary site goes down because GPFS has
the majority of its nodes still available. Without a dedicated quorum node, the
quorum nodes functionality must be put on the primary site to ensure that GPFS
in the primary data center continues to operate even if the inter-site link is
interrupted.
Figure 7-28 shows the GPFS view of an environment without a dedicated
quorum node. The weight symbolizes quorum designation. There are three
servers with a quorum designation in the primary data center and two servers in
the backup data center.
Primary site
Secondary site
Weight = 3
Weight = 2
Figure 7-28 Outline of a two-site environment without a dedicated GPFS quorum node
231
If a disaster happens at the primary site, the GPFS cluster loses its quorum
because the two quorum nodes at the backup site do not meet the minimum of
three. An additional procedure is required to relax GPFS quorum on the surviving
secondary site before GPFS comes up again. The exact procedure is
documented in the Operations Guide for IBM Systems Solution for SAP HANA.
After GPFS is running again, the procedure is identical to a failover with a
dedicated quorum node. It is optional to restore HA capabilities within the
secondary site. SAP HANA can be started already while this restore procedure is
still running, but a performance impact must be expected because this is an
I/O-intensive operation.
You need three GPFS server licenses on the primary and on the backup sites
even though during normal operation only two of them are required on the
backup site. If a disaster happens and you must fail over the SAP HANA
production instance to the backup data center, this data center becomes the
main SAP HANA cluster and requires three GPFS server licenses. Additional
servers on either site get GPFS FPO licenses.8
232
The currently validated minimum number of servers on each site for DR is three. This is required by
SAP to set up a scale-out environment with HA. Hence, the requirement is for at least three GPFS
server licenses per site.
Primary site
Secondary site
node2
node3
node4
node5
node6
node7
node8
Local
storage
Local
storage
Local
storage
Local
storage
Local
storage
Local
storage
Local
storage
Local
storage
third
replica
Production
file
system
second
replica
first
replica
node1
Non-production
file system
First replica
Second replica
...
...
...
...
Second file system spanning only expansion unit drives (metadata and data)
Figure 7-29 Overview of running non-production SAP HANA instances on the idling backup site
If a failover happens from the primary site to the secondary site, and you are
planning to keep running production from the secondary data center, it means
that you must be able to host non-production instances in the primary data
center. To accommodate for this additional storage space on the primary site
servers, you must connect EXP2524 expansions units to them as well.
If you plan to fail back gracefully your production instance from the backup site to
the primary site after it is repaired, you do not need to have EXP2524 expansion
units on the primary site servers. There might be unforeseen outages that take a
long time to repair. You cannot run your non-production site instances during this
outage.
There is exactly one new file system that spans all expansion units of the backup
site servers. This new file system runs with a GPFS replication factor of two,
meaning that there are always two copies of each data block. The first replica is
stored local in the node writing the data. The second replica is stored in a striped
round-robin fashion over the other nodes. This is identical to a scale-out HA
environment. One server can fail and the data is still available on the other nodes
expansion units. Figure 7-29 shows this from node5s perspective. If node5
writes data into the non-production file system, it stores the first replica on local
disks in the expansion unit. The second replica is striped across node6, node7,
and node8s expansion unit drives (symbolized as a long blue box in the figure).
233
Summary
The DR solution for the IBM Systems Solution for SAP HANA uses the advanced
replication features of GPFS, creating a cross-site cluster that ensures availability
and consistency of data across two sites. It does not impose the need for
additional storage systems, but instead builds upon the scale-out solution for
SAP HANA. This simple architecture reduces the complexity in maintaining such
a solution while keeping the possibility of adding more nodes over time if the
database grows.
234
Site A
node01
node02
node03
node04
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
Standby node
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
local storage
node03
node04
Second replica
SAP HANA
System Replication
Site B
node01
node02
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
Standby node
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
Second replica
local storage
235
With the release of SAP HANA 1.0 SPS 08 in June 2014, it is possible to use
different networks for SSR traffic. You can use either the front-end client network,
the back-end SAP HANA network, or a new dedicated replication network
spanning both sites. Having the choice between multiple networks allows you to
better adapt to different customer situations.
During normal operation, the nodes at the DR site A are running in recovery
mode. This mode allows you to host a non-productive SAP HANA instance (such
as development or test) on these idling nodes. If there is a disaster, this
non-productive SAP HANA instance must be stopped before the production
instance can be started at site B.
This additional SAP HANA instance needs its own storage space for persistency
and logs. The EXP2524 unit is used to provide this additional space. The
EXP2524 adds up to twenty-four 2.5-inch HDDs that are connected directly to
the server through a SAS interface. A second file system is created over those
drives for the additional SAP HANA instance. Depending on the size of this
additional database instance, one or more DR site nodes must be used. Every
node in scope requires an EXP2524 storage expansion. This second file system
is created with a replication factor of two to implement HA in the non-production
instance.
236
Backing up
A backup of the SAP HANA database must be triggered through the SAP HANA
Studio or through the SAP HANA SQL interface. SAP HANA then creates a
consistent backup, consisting of one file per SAP HANA service on each cluster
node. SAP HANA always performs a full backup. Incremental backups are not
supported by SAP HANA.
SAP HANA internally maintains transaction numbers, which are unique within a
database instance, especially in a scale-out configuration. To create a consistent
backup across a scale-out configuration, SAP HANA chooses a specific
transaction number, and all nodes of the database instance write their own
backup files, including all transactions up to this transaction number.
The backup files are saved to a defined staging area that might be on the internal
disks, an external disk on an NFS share,9 or a directly attached SAN subsystem.
In addition to the data backup files, the SAP HANA configuration files and backup
catalog files must be saved to be recovered. For point in time recovery, the log
area also must be backed up.
With the System x solution for SAP HANA, one of the 1 Gbit network interfaces of
the server can be used for NFS connectivity, or an additional 10 Gbit network
interface must be installed (if a PCIe slot is available). You can add a Fibre
Channel host bus adapter (HBA) for SAN connectivity. The Quick Start Guide for
the IBM Systems Solution for SAP HANA lists supported hardware additions to
provide additional connectivity. This guide can be found at the following website:
http://www-947.ibm.com/support/entry/myportal/docdisplay?lndocid=MIGR-5
087035
Restoring a backup
It might be necessary to recover the SAP HANA database from a backup in the
following situations:
The data area is damaged.
If the data area is unusable, the SAP HANA database can be recovered up to
the last committed transaction if all the data changes after the last complete
data backup are still available in the log backups and log area. After the data
and log backups are restored, the SAP HANA databases uses the data and
log backups and the log entries in the log area to restore the data and replay
the logs to recover. It also is possible to recover the database using an older
data backup and log backups if all relevant log backups that are made after
the data backup are available.10
9
10
SAP Note 1820529 lists network file systems that are unsuitable for backup and recovery.
For help with determining the files that are needed for a recovery, see SAP Note 1705945.
237
238
More detailed information about the backup and recovery processes for the SAP
HANA database is provided in the SAP HANA Backup and Recovery Guide,
found at the following website:
http://help.sap.com/hana_platform
239
Using snapshots has much less impact on the performance of the running
database than to trigger a file-based backup. Triggering a GPFS snapshot works
in single-node and scale-out environments. The time that it takes to activate a
snapshot depends on the amount of data in the file system and the current load
on it.
11
240
For more information, see SAP Note 1730932 Using backup tools with Backint.
Backint certification
Backup tools using the Backint for SAP HANA interface are subject to
certification by SAP. The certification process is documented at
http://scn.sap.com/docs/DOC-34483. To determine which backup tools are
certified for Backint for SAP HANA, search the Partner Information Center and
select the SAP-defined integration scenario HANA-BRINT 1.1 - HANA Backint
Interface. The search function of the Partner Information Center is available at
http://www.sap.com/partners/directories/SearchSolution.epx.
As of April 2014, the following tools are certified by SAP:
The following sections give a short overview about the first two backup tools in
this list.
241
In addition to the Backint integration, Tivoli Storage Manager for ERP 6.4
continues to feature a file-based integration with SAP HANA for pre-SPS05 SAP
HANA deployments on IBM hardware. Section A.2, File-based backup with IBM
Tivoli Storage Manager for ERP on page 258 describes such a file-based
integration of Tivoli Storage Manager for ERP V6.4 with SAP HANA in detail.
242
With a NetBackup agent Version 7.5.0.6 or later installed on the SAP HANA
appliance,13 SAP HANA can be integrated into an existing NetBackup
deployment. This allows SAP HANA to send streamed backup data to
NetBackup, which uses an SAP policy to manage the backed up data sets and
provide destination targets, retention, duplication, and replication for DR
purposes. This seamless integration facilitates the following benefits:
13
243
Figure 7-31 shows the concept of using backup and restore as a basic DR
solution.
Primary Site
node01
node02
node03
node04
backup
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
Standby node
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
Storage
Second replica
local storage
copy or otherwise
transfer the backup files
Secondary Site
node01
node02
node03
node04
restore
SAP HANA DB
DB partition 1
DB partition 2
DB partition 3
Standby node
First replica
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
Storage
Second replica
local storage
244
Chapter 8.
Installation services
IBM SAP HANA Operations Guide
Interoperability with other platforms
Monitoring SAP HANA
Installing additional agents
Software and firmware levels
Support process
245
246
247
248
The SAP tool for the administration of and monitoring the SAP HANA appliance
is the SAP HANA Studio. It allows you to monitor the overall system state:
General system information (such as software versions).
A warning section shows the latest warnings that are generated by the
statistics server. Detailed information about these warnings is available as a
tooltip.
Bar views provide an overview of important system resources. The amount of
available memory, CPUs, and storage space is displayed, in addition to the
used amount of these resources.
In a distributed landscape, the amount of available resources is aggregated over
all servers.
Note: For more information about the administration and monitoring of SAP
HANA, see the SAP HANA Administration Guide, found at the following
website:
http://help.sap.com/hana_platform
Tolerated
249
Prohibited
Solutions that must not be used on the IBM Systems Solution for
SAP HANA. Using these solutions might compromise the
performance, stability, or data integrity of SAP HANA.
Do not install additional software on the SAP HANA appliance that is classified
as prohibited for use on the SAP HANA appliance. As an example, initial tests
show that some agents can decrease performance or even possibly corrupt the
SAP HANA database (for example, virus scanners).
In general, all additionally installed software must be configured to not interfere
with the functions or performance of the SAP HANA appliance. If any issues with
the SAP HANA appliance occur, you might be asked by SAP to remove all
additional software and to reproduce the issue.
The list of agents that are supported, tolerated, or prohibited for use on the SAP
HANA appliance are published in the Quick Start Guide for the IBM Systems
Solution for SAP HANA appliance, which is available at this website:
http://www-947.ibm.com/support/entry/myportal/docdisplay?lndocid=MIGR-5
087035
Firmware
Operating system
Hardware drivers
Software
The IBM System x SAP HANA support team reserves the right to perform basic
system tests on these levels when they are deemed to have a direct impact on
the SAP HANA appliance. In general, IBM does not give specific
recommendations to which levels are allowed for the SAP HANA appliance.
The IBM System x SAP HANA development team provides, at regular intervals,
new images for the SAP HANA appliance. Because these images have
dependencies regarding the hardware, operating system, and drivers, use the
latest image for maintenance and installation of SAP HANA systems. These
images can be obtained through IBM Support. Part number information is
contained in the Quick Start Guide.
250
If the firmware level recommendations for the IBM components of the SAP HANA
appliance are given through the individual System x support teams that fix known
code bugs, it is the clients responsibility to upgrade or downgrade to the
recommended levels as instructed by IBM Support.
If the operating system recommendations for the SUSE Linux components of the
SAP HANA appliance are given through the SAP, SUSE, or IBM support teams
that fix known code bugs, it is the clients responsibility to upgrade or downgrade
to the recommended levels, as instructed by SAP through an explicit SAP Note
or allowed through an OSS Customer Message. SAP describes their operational
concept, including updating of the operating system components, in SAP Note
1599888 - SAP HANA: Operational Concept. If the Linux kernel is updated, take
extra care to recompile the IBM High IOPS drivers and IBM GPFS software as
well, as described in the IBM SAP HANA Operations Guide.
If an IBM High IOPS driver or IBM GPFS recommendation to update the software
is given through the individual IBM Support teams (System x, Linux, or GPFS)
that fix known code bugs, it is not recommend to update these drivers without
first asking the IBM System x SAP HANA support team through an SAP OSS
Customer Message.
If the other hardware or software recommendations for IBM components of the
SAP HANA appliance are given through the individual IBM Support teams that fix
known code bugs, it is the clients responsibility to upgrade or downgrade to the
recommended levels as instructed by IBM Support.
2
3
For more information about the IBM Statement of Limited Warranty, see the following website:
http://www.ibm.com/servers/support/machine_warranties
IBM sends a technician after attempting to diagnose and resolve the problem remotely.
251
252
253
254
Appendix A.
Additional topics
This appendix covers the following topics:
GPFS license information
File-based backup with IBM Tivoli Storage Manager for ERP
255
256
Table A-1 GPFS licenses that are included in the custom models for SAP HANA
MTM
PVUs
included
7147-H1x
1400
7147-H2x
1400
7147-H3x
1400
7147-H7x
1400
7147-H8x
1400
7147-H9x
1400
7147-HAx
1400
7147-HBx
1400
7143-H1x
1400
7143-H2x
4000
7143-H3x
5600
7143-H4x
1400
7143-H5x
4000
7143-HAx
4000
7143-HBx
4000
7143-HCx
5600
Licenses for IBM GPFS on x86 Single Server for Integrated Offerings V3
(referred to as Integrated in the table) cannot be ordered independently of the
select hardware for which it is included. This type of license provides file system
capabilities for single-node integrated offerings. Therefore, the model 7143-HAx
includes 4000 PVUs of GPFS on x86 Single Server for Integrated Offerings V3
licenses, so that an upgrade to the 7143-HBx model does not require additional
licenses. The PVU rating for the 7143-HAx model to consider when purchasing
other GPFS license types is 1400 PVUs.
Clients with highly available, multi-node clustered scale-out configurations must
purchase the GPFS on x86 Server and GPFS File Placement Optimizer product.
257
258
node01
node02
node03
SAP HANA DB
DB partition 1
5
backup
2 - GPFS
Shared file system
data02 + log02
3
Backup files
DB partition 3
data01 + log01
First replica
Tivoli
Storage
Manager
Server
DB partition 2
backup
backup
local storage
local storage
data03 + log03
backup
local storage
backup.sh
restore
259
Instead of having the backup files of the individual nodes written to the local
storage of the nodes, an external storage system can be used to provide space
to store the backup files. All nodes must be able to access this storage, for
example, using NFS. Figure A-2 shows this scenario.
node01
node02
node03
SAP HANA DB
DB partition 1
2
First replica
Tivoli
Storage
Manager
Server
5
backup
DB partition 2
DB partition 3
data01 + log01
data02 + log02
data03 + log03
local storage
local storage
local storage
backup.sh
restore
4
Tivoli Storage Manager
Storage
backup
backup
backup
SAP HANA
backup file storage
Figure A-2 Backup process with Data Protection for SAP HANA using external storage for backup files
Running log and data backups requires the Data Protection for SAP HANA
backup.sh command to be run as the SAP HANA administration user
(<sid>adm).
The backup.sh command provides two basic functions:
1. Completes the data-backup (including HANA instance and landscape
configuration files).
2. Completes log-backup and removes successfully saved redo log files from
disk.
260
backup.sh --logs
By using this command, a backup of the SAP HANA database into Tivoli Storage
Manager can be automated fully.
261
To restore data backups, including SAP HANA configuration files and log file
backups, the Tivoli Storage Manager BACKUP-Filemanager is used. Figure A-3
shows a sample panel of the BACKUP-Filemanager.
BACKUP-Filemanager V6.4.0.0, Copyright IBM 2001-2012
.------------------+---------------------------------------------------------------.
| Backup IDs
| Files stored under TSM___A0H7K1C4QI
|
|------------------+---------------------------------------------------------------|
| TSM___A0H7KM0XF4 | */hana/log_backup/log_backup_2_0_1083027170688_1083043933760 |
| TSM___A0H7KLYP3Z | */hana/log_backup/log_backup_2_0_1083043933760_1083060697664 |
| TSM___A0H7KHNLU6 | */hana/log_backup/log_backup_2_0_1083060697664_1083077461376 |
| TSM___A0H7KE6V19 | */hana/log_backup/log_backup_2_0_1083077461376_1083094223936 |
| TSM___A0H7K9KR7F | */hana/log_backup/log_backup_2_0_1083094223936_1083110986880 |
| TSM___A0H7K7L73W | */hana/log_backup/log_backup_2_0_1083110986880_1083127750848 |
| TSM___A0H7K720A4 | */hana/log_backup/log_backup_2_0_1083127750848_1083144513792 |
| TSM___A0H7K4BDXV | */hana/log_backup/log_backup_2_0_1083144513792_1083161277760 |
| TSM___A0H7K472YC | */hana/log_backup/log_backup_2_0_1083161277760_1083178040064 |
| TSM___A0H7K466HK | */hana/log_backup/log_backup_2_0_1083178040064_1083194806336 |
| TSM___A0H7K1C4QI | */hana/log_backup/log_backup_2_0_1083194806336_1083211570688 |
| TSM___A0H7JX1S77 | */hana/log_backup/log_backup_2_0_1083211570688_1083228345728 |
| TSM___A0H7JSRG2B | */hana/log_backup/log_backup_2_0_1083228345728_1083245109824 |
| TSM___A0H7JOH1ZP | */hana/log_backup/log_backup_2_0_1083245109824_1083261872960 |
| TSM___A0H7JK6ONC | */hana/log_backup/log_backup_2_0_1083261872960_1083278636608 |
| TSM___A0H7JJWUI8 | */hana/log_backup/log_backup_2_0_1083278636608_1083295400384 |
| TSM___A0H7JJU5YN | */hana/log_backup/log_backup_2_0_1083295400384_1083312166016 |
| TSM___A0H7JFWAV4 | */hana/log_backup/log_backup_2_0_1083312166016_1083328934016 |
| TSM___A0H7JBG625 | */hana/log_backup/log_backup_2_0_1083328934016_1083345705856 |
| TSM___A0H7JBAASN | */hana/log_backup/log_backup_2_0_1083345705856_1083362476352 |
| TSM___A0H7J7BLDK | */hana/log_backup/log_backup_2_0_1083362476352_1083379244416 |
| TSM___A0H7J5U8S7 | */hana/log_backup/log_backup_2_0_1083379244416_1083396008064 |
| TSM___A0H7J5T92O | */hana/log_backup/log_backup_2_0_1083396008064_1083412772928 |
| TSM___A0H7J4TWPG | */hana/log_backup/log_backup_2_0_1083412772928_1083429538688 |
|
| */hana/log_backup/log_backup_2_0_1083429538688_1083446303424 |
|
| */hana/log_backup/log_backup_2_0_1083446303424_1083463079488 |
|
| */hana/log_backup/log_backup_2_0_1083463079488_1083479846528 V
|------------------+---------------------------------------------------------------|
| 24 BIDs
| 190 File(s) - 190 marked
|
`------------------+---------------------------------------------------------------'
TAB change windows
F2 Restore
F3 Mark all
F4 Unmark allF5 reFresh
F6 fileInfo
F7 redireCt
F8 Delete
F10 eXit
ENTER mark file
Figure A-3 The BACKUP-Filemanager interface
Data and log backups can be selected and then restored to the wanted location.
If no directory is specified for the restore, the BACKUP-Filemanager restores the
backups to the original location from which the backup was done.
262
After the backup files are restored, the recovery process must be started by
using SAP HANA Studio. More information about this process and the various
options for a recovery is contained in the SAP HANA Backup and Recovery
Guide, found at the following website:
http://help.sap.com/hana_platform
After successfully completing the recovery process, if the backup files are no
longer needed, they must be removed manually from the disk.
263
264
Advanced Business
Application Programming
HDD
HPI
ACID
atomicity, consistency,
isolation, durability
I/O
input/output
APO
IBM
International Business
Machines
BI
Business Intelligence
ID
identifier
BICS
BI Consumer Services
IDs
identifiers
BM
bridge module
IMM
integrated management
module
BW
Business Warehouse
IOPS
CD
compact disc
ISICC
CPU
CRC
ITSO
CRM
customer relationship
management
International Technical
Support Organization
JDBC
CRU
customer-replaceable unit
JRE
DB
database
KPIs
DEV
development
LM
landscape management
DIMM
LUW
DSO
DataStore Object
MB
megabyte
DR
disaster recovery
MCA
DXC
MCOD
ECC
ECC
MCOS
ERP
ETL
MDX
Multidimensional Expressions
FTSS
NOS
GB
gigabyte
NSD
GBS
NUMA
ODBC
GPFS
ODBO
GTS
OLAP
HA
high availability
OLTP
265
OS
operating system
SQL
OSS
SSD
solid-state drive
PAM
SSR
PC
personal computer
PCI
Peripheral Component
Interconnect
STG
POC
proof of concept
SUM
PSA
TB
terabyte
PVU
TCO
QA
quality assurance
TCP/IP
QPI
QuickPath Interconnect
Transmission Control
Protocol/Internet Protocol
RAID
Redundant Array of
Independent Disks
TDMS
TREX
RAM
RAS
UEFI
RDS
RHEL
RPM
RPO
RTO
SAN
SAPS
SAS
serial-attached SCSI
SATA
Serial ATA
SCM
SCM
software configuration
management
SD
SDRAM
SLD
SLES
SLO
System Landscape
Optimization
SMI
266
Related publications
The publications that are listed in this section are considered suitable for a more
detailed description of the topics that are covered in this book.
IBM Redbooks
The following IBM Redbooks publications provide additional information about
the topics in this document. Some publications that are referenced in this list
might be available in softcopy only:
The Benefits of Running SAP Solutions on IBM eX5 Systems, REDP-4234
IBM eX5 Portfolio Overview: IBM System x3850 X5, x3950 X5, x3690 X5, and
BladeCenter HX5, REDP-4650
Implementing the IBM General Parallel File System (GPFS) in a Cross
Platform Environment, SG24-7844
You can search for, view, download, or order these documents and other
Redbooks, Redpapers, Web Docs, drafts, and additional materials, at the
following website:
ibm.com/redbooks
Online resources
These websites are also relevant as further information sources:
IBM and SAP: Business Warehouse Accelerator
http://www.ibm-sap.com/bwa
IBM Systems and Services for SAP HANA
http://www.ibm-sap.com/hana
IBM Systems Solution for SAP HANA
http://www.ibm.com/systems/x/solutions/sap/hana
SAP In-Memory Computing - SAP Help Portal
http://help.sap.com/hana
267
268
(0.5 spine)
0.475<->0.875
250 <-> 459 pages
Back cover
In-memory Computing
with SAP HANA on IBM
eX5 and X6 Systems
IBM System x
solution for
SAP HANA
SAP HANA overview
and use cases
Operational aspects
for SAP HANA
appliances
INTERNATIONAL
TECHNICAL
SUPPORT
ORGANIZATION
BUILDING TECHNICAL
INFORMATION BASED ON
PRACTICAL EXPERIENCE
ISBN 0738439908