Performance Report PRIMERGY RX2530 M1

White Paper  Performance Report PRIMERGY RX2530 M1
White Paper
FUJITSU Server PRIMERGY
Performance Report PRIMERGY RX2530 M1
This document contains a summary of the benchmarks executed for the FUJITSU Server
PRIMERGY RX2530 M1.
The PRIMERGY RX2530 M1 performance data are compared with the data of other
PRIMERGY models and discussed. In addition to the benchmark results, an explanation
has been included for each benchmark and for the benchmark environment.
Version
1.1
2015-04-15
http://ts.fujitsu.com/primergy
Page 1 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Contents
Document history ................................................................................................................................................ 2
Technical data .................................................................................................................................................... 3
SPECcpu2006 .................................................................................................................................................... 6
SPECpower_ssj2008 ........................................................................................................................................ 13
Disk I/O: Performance of RAID controllers ....................................................................................................... 18
OLTP-2 ............................................................................................................................................................. 27
vServCon .......................................................................................................................................................... 31
VMmark V2 ....................................................................................................................................................... 38
STREAM ........................................................................................................................................................... 45
LINPACK .......................................................................................................................................................... 47
Literature ........................................................................................................................................................... 51
Contact ............................................................................................................................................................. 52
Document history
Version 1.0
New:






Technical data
SPECpower_ssj2008
Measurement with Xeon E5-2699 v3
Disk I/O: Performance of RAID controllers
Measurements with “LSI SW RAID on Intel C610 (Onboard SATA)”, “PRAID CP400i”, “PRAID
EP400i” and “PRAID EP420i” controllers
OLTP-2
®
®
Results for Intel Xeon Processor E5-2600 v3 Product Family
vServCon
®
®
Results for Intel Xeon Processor E5-2600 v3 Product Family
VMmark V2
Measurement with Xeon E5-2699 v3
“Performance with Server Power” measurement with Xeon E5-2699 v3
“Performance with Server and Storage Power” measurement with Xeon E5-2699 v3
Version 1.1
New:



SPECcpu2006
®
®
Measurements with Intel Xeon Processor E5-2600 v3 Product Family
STREAM
®
®
Measurements with Intel Xeon Processor E5-2600 v3 Product Family
LINPACK
®
®
Measurements with Intel Xeon Processor E5-2600 v3 Product Family
Page 2 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Technical data
PRIMERGY RX2530 M1
PY RX2530 M1 4x 3.5"
PRIMERGY RX2530 M1
PY RX2530 M1 10x 2.5"
Decimal prefixes according to the SI standard are used for measurement units in this white paper (e.g. 1 GB
9
30
= 10 bytes). In contrast, these prefixes should be interpreted as binary prefixes (e.g. 1 GB = 2 bytes) for
the capacities of caches and memory modules. Separate reference will be made to any further exceptions
where applicable.
Model
PRIMERGY RX2530 M1
Model versions
PY RX2530 M1 4x 3.5"
PY RX2530 M1 4x 2.5" expandable
PY RX2530 M1 10x 2.5"
Form factor
Rack server
Chipset
Intel C612
Number of sockets
2
Number of processors orderable
1 or 2
Processor type
Intel Xeon Processor E5-2600 v3 Product Family
Number of memory slots
24 (12 per processor)
Maximum memory configuration
768 GB
Onboard HDD controller
Controller with RAID 0, RAID 1 or RAID 10 for up to 8 SATA HDDs
2 × PCI-Express 3.0 x8
2 × PCI-Express 3.0 x16
PY RX2530 M1 4x 3.5":
4
PY RX2530 M1 4x 2.5" expandable: 8
PY RX2530 M1 10x 2.5":
10
PCI slots
Max. number of internal hard disks
http://ts.fujitsu.com/primergy
®
®
®
Page 3 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Cache
Xeon E5-2623 v3
4
8
10
8.00
3.00
3.50
1866
105
Xeon E5-2637 v3
4
8
15
9.60
3.50
3.70
2133
135
Xeon E5-2603 v3
6
6
15
6.40
1.60
n/a
1600
85
Xeon E5-2609 v3
6
6
15
6.40
1.90
n/a
1600
85
Xeon E5-2620 v3
6
12
15
8.00
2.40
3.20
1866
85
Xeon E5-2643 v3
6
12
20
9.60
3.40
3.70
2133
135
Xeon E5-2630L v3 8
16
20
8.00
1.80
2.90
1866
55
Xeon E5-2630 v3
8
16
20
8.00
2.40
3.20
1866
85
Xeon E5-2640 v3
8
16
20
8.00
2.60
3.40
1866
90
Xeon E5-2667 v3
8
16
20
9.60
3.20
3.60
2133
135
Xeon E5-2650 v3
10
20
25
9.60
2.30
3.00
2133
105
Xeon E5-2660 v3
10
20
25
9.60
2.60
3.3
2133
105
Xeon E5-2650L v3 12
24
30
9.60
1.80
2.50
2133
65
Xeon E5-2670 v3
12
24
30
9.60
2.30
3.10
2133
120
Xeon E5-2680 v3
12
24
30
9.60
2.50
3.30
2133
120
Xeon E5-2690 v3
12
24
30
9.60
2.60
3.50
2133
135
Xeon E5-2683 v3
14
28
35
9.60
2.00
3.00
2133
120
Xeon E5-2695 v3
14
28
35
9.60
2.30
3.30
2133
120
Xeon E5-2697 v3
14
28
35
9.60
2.60
3.60
2133
145
Xeon E5-2698 v3
16
32
40
9.60
2.30
3.60
2133
135
Xeon E5-2699 v3
18
36
45
9.60
2.30
3.60
2133
145
Processor
Cores
Threads
Processors (since system release)
QPI
Speed
Rated
Frequency
[Ghz]
Max.
Turbo
Frequency
[Ghz]
Max.
Memory
Frequency
[MHz]
[MB]
[GT/s]
TDP
[Watt]
All the processors that can be ordered with the PRIMERGY RX2530 M1, apart from Xeon E5-2603 v3 and
®
Xeon E5-2609 v3, support Intel Turbo Boost Technology 2.0. This technology allows you to operate the
processor with higher frequencies than the nominal frequency. Listed in the processor table is "Max. Turbo
Frequency" for the theoretical frequency maximum with only one active core per processor. The maximum
frequency that can actually be achieved depends on the number of active cores, the current consumption,
electrical power consumption and the temperature of the processor.
As a matter of principle Intel does not guarantee that the maximum turbo frequency will be reached. This is
related to manufacturing tolerances, which result in a variance regarding the performance of various
examples of a processor model. The range of the variance covers the entire scope between the nominal
frequency and the maximum turbo frequency.
The turbo functionality can be set via BIOS option. Fujitsu generally recommends leaving the "Turbo Mode"
option set at the standard setting "Enabled", as performance is substantially increased by the higher
frequencies. However, since the higher frequencies depend on general conditions and are not always
guaranteed, it can be advantageous to disable the "Turbo Mode" option for application scenarios with
intensive use of AVX instructions and a high number of instructions per clock unit, as well as for those that
require constant performance or lower electrical power consumption.
Page 4 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
ECC
1
4
2133


Load reduced
8
Low voltage
Registered
Frequency [MHz]
Ranks
8GB (1x8GB) 1Rx4 DDR4-2133 R ECC
Capacity [GB]
Memory module
Bit width of the
memory chips
Memory modules (since system release)
8
2
8
2133


16GB (1x16GB) 2Rx4 DDR4-2133 R ECC
16
2
4
2133


32GB (1x32GB) 2Rx4 DDR4-2133 R ECC
32
2
4
2133


32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
32
4
4
2133


8GB (1x8GB) 2Rx8 DDR4-2133 R ECC
Power supplies (since system release)

Max. number
Modular PSU 450W platinum hp
2
Modular PSU 800W platinum hp
2
Modular PSU 800W titanium hp
2
Some components may not be available in all countries or sales regions.
Detailed technical information is available in the data sheet PRIMERGY RX2530 M1.
http://ts.fujitsu.com/primergy
Page 5 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
SPECcpu2006
Benchmark description
SPECcpu2006 is a benchmark which measures the system efficiency with integer and floating-point
operations. It consists of an integer test suite (SPECint2006) containing 12 applications and a floating-point
test suite (SPECfp2006) containing 17 applications. Both test suites are extremely computing-intensive and
concentrate on the CPU and the memory. Other components, such as Disk I/O and network, are not
measured by this benchmark.
SPECcpu2006 is not tied to a special operating system. The benchmark is available as source code and is
compiled before the actual measurement. The used compiler version and their optimization settings also
affect the measurement result.
SPECcpu2006 contains two different performance measurement methods: the first method (SPECint2006 or
SPECfp2006) determines the time which is required to process single task. The second method
(SPECint_rate2006 or SPECfp_rate2006) determines the throughput, i.e. the number of tasks that can be
handled in parallel. Both methods are also divided into two measurement runs, “base” and “peak” which
differ in the use of compiler optimization. When publishing the results the base values are always used; the
peak values are optional.
Benchmark
Arithmetics
Type
Compiler
optimization
SPECint2006
integer
peak
aggressive
SPECint_base2006
integer
base
conservative
SPECint_rate2006
integer
peak
aggressive
SPECint_rate_base2006
integer
base
conservative
SPECfp2006
floating point
peak
aggressive
SPECfp_base2006
floating point
base
conservative
SPECfp_rate2006
floating point
peak
aggressive
SPECfp_rate_base2006
floating point
base
conservative
Measurement
result
Application
Speed
single-threaded
Throughput
multi-threaded
Speed
single-threaded
Throughput
multi-threaded
The measurement results are the geometric average from normalized ratio values which have been
determined for individual benchmarks. The geometric average - in contrast to the arithmetic average - means
that there is a weighting in favour of the lower individual results. Normalized means that the measurement is
how fast is the test system compared to a reference system. Value “1” was defined for the
SPECint_base2006-, SPECint_rate_base2006, SPECfp_base2006 and SPECfp_rate_base2006 results of
the reference system. For example, a SPECint_base2006 value of 2 means that the measuring system has
handled this benchmark twice as fast as the reference system. A SPECfp_rate_base2006 value of 4 means
that the measuring system has handled this benchmark some 4/[# base copies] times faster than the
reference system. “# base copies” specify how many parallel instances of the benchmark have been
executed.
Not every SPECcpu2006 measurement is submitted by us for publication at SPEC. This is why the SPEC
web pages do not have every result. As we archive the log files for all measurements, we can prove the
correct implementation of the measurements at any time.
Page 6 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
System Under Test (SUT)
Hardware
Model
PRIMERGY RX2530 M1
Processor
Intel Xeon Processor E5-2600 v3 Product Family
Memory
16 × 16GB (1x16GB) 2Rx4 DDR4-2133 R ECC
®
®
Software
Operating system
SPECint_base2006, SPECint2006:
Xeon E5-2630 v3, E5-2650 v3:
Red Hat Enterprise Linux Server release 7.0
All others: Red Hat Enterprise Linux Server release 6.5
SPECint_rate_base2006, SPECint_rate2006:
Xeon E5-2630 v3: Red Hat Enterprise Linux Server release 7.0
All others:
Red Hat Enterprise Linux Server release 6.5
SPECfp_base2006, SPECfp2006, SPECfp_rate_base2006, SPECfp_rate2006:
Red Hat Enterprise Linux Server release 7.0
Operating system
settings
echo always > /sys/kernel/mm/redhat_transparent_hugepage/enabled
Compiler
SPECint_base2006, SPECint2006:
Xeon E5-2630 v3, E5-2650 v3:
C/C++: Version 15.0.0.090 of Intel C++ Studio XE for Linux
All others: C/C++: Version 14.0.0.080 of Intel C++ Studio XE for Linux
SPECint_rate_base2006, SPECint_rate2006:
Xeon E5-2630 v3: C/C++: Version 15.0.0.090 of Intel C++ Studio XE for Linux
All others:
C/C++: Version 14.0.0.080 of Intel C++ Studio XE for Linux
SPECfp_base2006, SPECfp2006, SPECfp_rate_base2006, SPECfp_rate2006:
C/C++: Version 15.0.0.090 of Intel C++ Studio XE for Linux
Fortran: Version 15.0.0.090 of Intel Fortran Studio XE for Linux
Some components may not be available in all countries or sales regions.
http://ts.fujitsu.com/primergy
Page 7 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
Number of processors
SPECint_rate_base2006
SPECint_rate2006
1
209 (est.)
216 (est.)
2
410
424
Xeon E5-2637 v3
2
61.3
64.5
1
233 (est.)
240 (est.)
2
457
471
Xeon E5-2603 v3
2
29.3
30.5
1
135 (est.)
140 (est.)
2
265
275
Xeon E5-2609 v3
2
33.9
35.3
1
156 (est.)
162 (est.)
2
306
317
Xeon E5-2620 v3
2
53.8
57.0
1
261
270
2
508
524
Xeon E5-2643 v3
2
63.6
67.0
1
340 (est.)
352 (est.)
2
667
690
Xeon E5-2630L v3
2
49.9
53.3
1
288 (est.)
298 (est.)
2
564
585
Xeon E5-2630 v3
2
56.1
58.6
1
339 (est.)
353 (est.)
2
664
692
Xeon E5-2640 v3
2
58.4
62.2
1
359 (est.)
371 (est.)
2
703
727
Xeon E5-2667 v3
2
63.3
66.8
1
414 (est.)
429 (est.)
2
812
840
Xeon E5-2650 v3
2
54.6
56.6
1
420 (est.)
435 (est.)
2
823
852
Xeon E5-2660 v3
2
57.9
61.4
1
453 (est.)
468 (est.)
2
888
918
Xeon E5-2650L v3
2
46.3
48.7
1
403 (est.)
416 (est.)
2
790
816
Xeon E5-2670 v3
2
56.4
59.3
1
494 (est.)
510 (est.)
2
968
1000
Xeon E5-2680 v3
2
60.0
63.1
1
531 (est.)
546 (est.)
2
1040
1070
Xeon E5-2690 v3
2
62.0
65.4
1
556 (est.)
571 (est.)
2
1090
1120
Xeon E5-2683 v3
2
53.7
56.8
1
546 (est.)
561 (est.)
2
1070
1100
Xeon E5-2695 v3
2
58.7
61.9
1
577 (est.)
597 (est.)
2
1130
1170
Xeon E5-2697 v3
2
63.2
66.9
1
622 (est.)
643 (est.)
2
1220
1260
Xeon E5-2698 v3
2
62.6
66.3
1
643 (est.)
663 (est.)
2
1260
1300
Xeon E5-2699 v3
2
63.3
66.8
1
704 (est.)
724 (est.)
2
1380
1420
SPECint_rate2006
59.1
Page 8 (52)
SPECint_rate_base2006
56.2
Number of processors
2
SPECint2006
Xeon E5-2623 v3
SPECint_base2006
Processor
Number of processors
In terms of processors the benchmark result depends primarily on the size of the processor cache, the
support for Hyper-Threading, the number of processor cores and on the processor frequency. In the case of
processors with Turbo mode the number of cores, which are loaded by the benchmark, determines the
maximum processor frequency that can be achieved. In the case of single-threaded benchmarks, which
largely load one core only, the maximum processor frequency that can be achieved is higher than with multithreaded benchmarks.
The results marked (est.) are estimates.
http://ts.fujitsu.com/primergy
Number of processors
SPECfp_rate_base2006
SPECfp_rate2006
2
379
389
1
214 (est.)
222 (est.)
2
425
439
57.2
1
142 (est.)
145 (est.)
2
282
287
64.1
1
163 (est.)
167 (est.)
2
324
330
479
Xeon E5-2637 v3
2
Xeon E5-2603 v3
2
55.4
Xeon E5-2609 v3
2
61.9
Xeon E5-2620 v3
2
95.2
Xeon E5-2643 v3
2
Xeon E5-2630L v3
2
Xeon E5-2630 v3
2
102
107
Xeon E5-2640 v3
2
106
Xeon E5-2667 v3
2
Xeon E5-2650 v3
105
SPECfp2006
SPECfp_base2006
2
99.1
108
SPECfp_rate2006
196 (est.)
Xeon E5-2623 v3
Number of processors
190 (est.)
Number of processors
1
Processor
95.6
Version: 1.1  2015-04-15
SPECfp_rate_base2006
White Paper  Performance Report PRIMERGY RX2530 M1
100
1
235 (est.)
242 (est.)
2
468
118
1
289 (est.)
298 (est.)
2
575
591
1
254 (est.)
262 (est.)
2
506
519
1
284 (est.)
293 (est.)
2
566
581
112
1
294 (est.)
303 (est.)
2
586
600
116
121
1
330 (est.)
340 (est.)
2
656
674
2
103
107
1
342 (est.)
354 (est.)
2
681
700
Xeon E5-2660 v3
2
110
114
1
357 (est.)
369 (est.)
2
711
730
Xeon E5-2650L v3
2
1
323 (est.)
3333 (est.)
2
642
659
Xeon E5-2670 v3
2
105
110
1
381 (est.)
394 (est.)
2
759
780
Xeon E5-2680 v3
2
110
115
1
391 (est.)
405 (est.)
2
779
802
Xeon E5-2690 v3
2
114
118
1
403 (est.)
416 (est.)
2
801
824
Xeon E5-2683 v3
2
104
1
405 (est.)
418 (est.)
2
805
827
Xeon E5-2695 v3
2
105
110
1
412 (est.)
426 (est.)
2
820
844
Xeon E5-2697 v3
2
111
117
1
429 (est.)
444 (est.)
2
853
879
Xeon E5-2698 v3
2
107
113
1
437 (est.)
452 (est.)
2
869
895
Xeon E5-2699 v3
2
109
116
1
461 (est.)
478 (est.)
2
918
946
http://ts.fujitsu.com/primergy
113
91.5
88.1
99.4
97.7
92.0
Page 9 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
The following four diagrams illustrate the throughput of the PRIMERGY RX2530 M1 in comparison to its
predecessor PRIMERGY RX200 S8, in their respective most performant configuration.
SPECcpu2006: integer performance
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
67.5
70
62.7
67.0
63.6
60
50
40
30
20
SPECint2006
10
SPECint_base2006
0
PRIMERGY RX200 S8 PRIMERGY RX2530 M1
2 x Xeon E5-2667 v2
2 x Xeon E5-2643 v3
SPECcpu2006: integer performance
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
1420
1500
1380
960
1250
929
1000
750
500
SPECint_rate2006
250
SPECint_rate_base2006
0
PRIMERGY RX200 S8 PRIMERGY RX2530 M1
2 x Xeon E5-2697 v2 2 x Xeon E5-2699 v3
Page 10 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
SPECcpu2006: floating-point performance
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
121
113
125
116
108
100
75
50
SPECfp2006
25
SPECfp_base2006
0
PRIMERGY RX200 S8 PRIMERGY RX2530 M1
2 x Xeon E5-2667 v2
2 x Xeon E5-2667 v3
SPECcpu2006: floating-point performance
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
946
1000
900
918
696
800
678
700
600
500
400
300
SPECfp_rate2006
200
100
SPECfp_rate_base2006
0
PRIMERGY RX200 S8 PRIMERGY RX2530 M1
2 x Xeon E5-2697 v2
2 x Xeon E5-2699 v3
http://ts.fujitsu.com/primergy
Page 11 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
The diagram below reflects how the performance of the PRIMERGY RX2530 M1 scales from one to two
processors when using the Xeon E5-2620 v3.
SPECcpu2006: integer performance
PRIMERGY RX2530 M1 (2 sockets vs. 1 socket)
524
600
508
500
400
270
300
261
200
SPECint_rate2006
100
SPECint_rate_base2006
0
1 x Xeon E5-2620 v3
Page 12 (52)
2 x Xeon E5-2620 v3
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
SPECpower_ssj2008
Benchmark description
SPECpower_ssj2008 is the first industry-standard SPEC benchmark that evaluates the power and
performance characteristics of a server. With SPECpower_ssj2008 SPEC has defined standards for server
power measurements in the same way they have done for performance.
The benchmark workload represents typical server-side Java business applications. The workload is
scalable, multi-threaded, portable across a wide range of platforms and easy to run. The benchmark tests
CPUs, caches, the memory hierarchy and scalability of symmetric multiprocessor systems (SMPs), as well
as the implementation of Java Virtual Machine (JVM), Just In Time (JIT) compilers, garbage collection,
threads and some aspects of the operating system.
SPECpower_ssj2008 reports power consumption for
servers at different performance levels — from 100% to
“active idle” in 10% segments — over a set period of
time. The graduated workload recognizes the fact that
processing loads and power consumption on servers
vary substantially over the course of days or weeks. To
compute a power-performance metric across all levels,
measured transaction throughputs for each segment are
added together and then divided by the sum of the
average power consumed for each segment. The result
is a figure of merit called “overall ssj_ops/watt”. This
ratio provides information about the energy efficiency of
the measured server. The defined measurement
standard enables customers to compare it with other
configurations
and
servers
measured
with
SPECpower_ssj2008. The diagram shows a typical
graph of a SPECpower_ssj2008 result.
The benchmark runs on a wide variety of
operating
systems
and
hardware
architectures and does not require extensive
client or storage infrastructure. The minimum
equipment for SPEC-compliant testing is two
networked computers, plus a power analyzer
and a temperature sensor. One computer is
the System Under Test (SUT) which runs
one of the supported operating systems and
the JVM. The JVM provides the environment
required to run the SPECpower_ssj2008
workload which is implemented in Java. The
other computer is a “Control & Collection
System” (CCS) which controls the operation
of the benchmark and captures the power,
performance and temperature readings for
reporting. The diagram provides an overview
of the basic structure of the benchmark
configuration and the various components.
http://ts.fujitsu.com/primergy
Page 13 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
System Under Test (SUT)
Hardware
Model
PRIMERGY RX2530 M1
Model version
PY RX2530 M1 4x 3.5''
Processor
Xeon E5-2699 v3
Memory
8 × 8GB (1x8GB) 2Rx8 DDR4-2133 R ECC
Network-Interface
1 × PLAN AP 1x1Gbit Cu Intel I210-T1 LP
Disk-Subsystem
Onboard HDD controller
1 × DOM SATA 6G 64GB Main N H-P
Power Supply Unit
1 × Modular PSU 800W titanium hp
Software
BIOS
R1.11.0
BIOS settings
Hardware Prefetcher = Disabled
Adjacent Cache Line Prefetch = Disabled
DCU Streamer Prefetcher = Disabled
Onboard USB Controllers = Disabled
Power Technology = Custom
QPI Link Frequency Select = 6.4 GT/s
Turbo Mode = Disabled
Intel Virtualization Technology = Disabled
ASPM Support = L1 Only
DMI Control = Gen1
COD Enable = Enabled
Early Snoop = Disabled
Firmware
7.69F
Operating system
Microsoft Windows Server 2008 R2 Enterprise SP1
Operating system
settings
Using the local security settings console, “lock pages in memory” was enabled for the user
running the benchmark.
Power Management: Enabled (“Fujitsu Enhanced Power Settings” power plan)
Set “Turn off hard disk after = 1 Minute” in OS.
Benchmark was started via Windows Remote Desktop Connection.
Microsoft Hotfix KB2510206 has been installed due to known problems of the group
assignment algorithm which does not create a balanced group assignment. For more
information see: http://support.microsoft.com/kb/2510206
JVM
IBM J9 VM (build 2.6, JRE 1.7.0 Windows Server 2008 R2 amd64-64 20120322_106209 (JIT
enabled, AOT enabled)
JVM settings
start /NODE [0,1,2,3] /AFFINITY [0x3,0xC,0x30,0xC0,0x300,0xC00,0x3000,0xC000,0x30000]
-Xmn825m -Xms975m -Xmx975m -Xaggressive -Xcompressedrefs -Xgcpolicy:gencon
-XlockReservation -Xnoloa -XtlhPrefetch -Xlp -Xconcurrentlevel0 -Xthr:minimizeusercpu
-Xgcthreads2 (-Xgcthreads1 for JVM5 and JVM23)
Other software
IBM WebSphere Application Server V8.5.0.0, Microsoft Hotfix for Windows (KB2510206)
Some components may not be available in all countries or sales regions.
Page 14 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
The PRIMERGY RX2530 M1 achieved the following result:
SPECpower_ssj2008 = 9,811 overall ssj_ops/watt
The adjoining diagram shows the
result of the configuration described
above. The red horizontal bars show
the performance to power ratio in
ssj_ops/watt (upper x-axis) for each
target load level tagged on the y-axis
of the diagram. The blue line shows
the run of the curve for the average
power consumption (bottom x-axis) at
each target load level marked with a
small rhomb. The black vertical line
shows the benchmark result of 9,811
overall
ssj_ops/watt
for
the
PRIMERGY RX2530 M1. This is the
quotient of the sum of the transaction
throughputs for each load level and
the sum of the average power consumed for each measurement interval.
The following table shows the benchmark results for the throughput in ssj_ops, the power consumption in
watts and the resulting energy efficiency for each load level.
Performance
Power
Energy Efficiency
Target Load
ssj_ops
100%
3,231,698
289
11,191
90%
2,906,711
260
11,173
80%
2,584,271
230
11,236
70%
2,263,620
199
11,377
60%
1,937,836
172
11,283
50%
1,618,047
153
10,544
40%
1,292,232
138
9,367
30%
970,277
123
7,896
20%
646,383
108
5,982
10%
322,218
92.3
3,492
0
47.3
0
Active Idle
Average Power (W)
ssj_ops/watt
∑ssj_ops / ∑power = 9,811
The PRIMERGY RX2530 M1 achieved a new class record with this result, thus surpassing the
best result of the competition by 23% (date: March 03, 2015). Thus, the PRIMERGY RX2530
M1 proves itself to be the most energy-efficient 1U rack server in the world. For the latest
SPECpower_ssj2008 benchmark results, visit:
http://www.spec.org/power_ssj2008/results.
http://ts.fujitsu.com/primergy
Page 15 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
SPECpower_ssj2008: PRIMERGY RX2530 M1 vs. competition
12,000
overall ssj_ops/watt
10,000
9,811
8,000
8,004
Version: 1.1  2015-04-15
The comparison with the competition makes
the advantage of the PRIMERGY RX2530
M1 in the field of energy efficiency evident.
With 23% more energy efficiency than the
best result of the competition in the class of
1U rack servers, the Supermicro SYS1028R-WC1RT server, the PRIMERGY
RX2530 M1 is setting new standards.
6,000
4,000
2,000
0
Supermicro
SYS-1028R-WC1RT
Fujitsu Server
PRIMERGY RX2530 M1
The following diagram shows for each load level the power consumption (on the right y-axis) and the
throughput (on the left y-axis) of the PRIMERGY RX2530 M1 compared to the predecessor PRIMERGY
RX200 S8.
SPECpower_ssj2008: PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
Page 16 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
SPECpower_ssj2008 overall ssj_ops/watt:
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
12,000
2,000
1,800
10,000
1,600
9,811
1,400
8,000
1,200
7,670
6,000
1,000
800
4,000
600
total power [watt]
overall ssj_ops/watt
Thanks to the new Haswell processors the
PRIMERGY RX2530 M1 has in comparison
with the PRIMERGY RX200 S8 a substantially higher throughput. Despite the higher
power consumption this results in an overall
increase in energy efficiency in the
PRIMERGY RX2530 M1 of 28%.
Version: 1.1  2015-04-15
400
2,000
200
0
0
Fujitsu Server
Fujitsu Server
PRIMERGY RX200 S8 PRIMERGY RX2530 M1
http://ts.fujitsu.com/primergy
Page 17 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Disk I/O: Performance of RAID controllers
Benchmark description
Performance measurements of disk subsystems for PRIMERGY servers are used to assess their
performance and enable a comparison of the different storage connections for PRIMERGY servers. As
standard, these performance measurements are carried out with a defined measurement method, which
models the accesses of real application scenarios on the basis of specifications.
The essential specifications are:
 Share of random accesses / sequential accesses
 Share of read / write access types
 Block size (kB)
 Number of parallel accesses (# of outstanding I/Os)
A given value combination of these specifications is known as “load profile”. The following five standard load
profiles can be allocated to typical application scenarios:
Standard load
profile
Access
Type of access
read
write
Block size
[kB]
Application
File copy
random
50%
50%
64
Copying of files
File server
random
67%
33%
64
File server
Database
random
67%
33%
8
Database (data transfer)
Mail server
Streaming
sequential
100%
0%
64
Database (log file),
Data backup;
Video streaming (partial)
Restore
sequential
0%
100%
64
Restoring of files
In order to model applications that access in parallel with a different load intensity, the “# of Outstanding
I/Os” is increased, starting with 1, 3, 8 and going up to 512 (from 8 onwards in increments to the power of
two).
The measurements of this document are based on these standard load profiles.
The main results of a measurement are:



Throughput [MB/s]
Transactions [IO/s]
Latency [ms]
Throughput in megabytes per second
Transaction rate in I/O operations per second
Average response time in ms
The data throughput has established itself as the normal measurement variable for sequential load profiles,
whereas the measurement variable “transaction rate” is mostly used for random load profiles with their small
block sizes. Data throughput and transaction rate are directly proportional to each other and can be
transferred to each other according to the formula
Data throughput [MB/s]
= Transaction rate [IO/s] × Block size [MB]
Transaction rate [IO/s]
= Data throughput [MB/s] / Block size [MB]
12
This section specifies capacities of storage media on a basis of 10 (1 TB = 10 bytes) while all other
20
capacities, file sizes, block sizes and throughputs are specified on a basis of 2 (1 MB/s = 2 bytes/s).
All the details of the measurement method and the basics of disk I/O performance are described in the white
paper “Basics of Disk I/O Performance”.
Page 18 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
All the measurement results discussed in this chapter were determined using the hardware and software
components listed below:
System Under Test (SUT)
Hardware
Controller
1 × “LSI SW RAID on Intel C610 (Onboard SATA)”
1 × “PRAID CP400i”
1 × “PRAID EP400i”
1 × “PRAID EP420i”
Drive
4 × 3.5" SATA HDD Seagate ST3000NM0033
4 × 2.5" SATA SSD Intel SSDSC2BA400G3C
10 × 2.5" SAS SSD Toshiba PX02SMF040
10 × 2.5" SAS HDD HGST HUC156045CSS204
Software
BIOS settings
Intel Virtualization Technology = Disabled
VT-d = Disabled
Energy Performance = Performance
Utilization Profile = Unbalanced
CPU C6 Report = Disabled
Operating system
Microsoft Windows Server 2012 Standard
Operating system
settings
Choose or customize a power plan: High performance
For the processes that create disk I/Os: set the AFFINITY to the CPU node to which the
PCIe slot of the RAID controller is connected
Administration
software
ServerView RAID Manager 5.7.2
Initialization of RAID
arrays
RAID arrays are initialized before the measurement with an elementary block size of 64 kB
(“stripe size”)
File system
NTFS
Measuring tool
Iometer 2006.07.27
Measurement data
Measurement files of 32 GB with 1 – 8 hard disks; 64 GB with 9 – 16 hard disks;
128 GB with 17 or more hard disks
Some components may not be available in all countries / sales regions.
http://ts.fujitsu.com/primergy
Page 19 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
The results presented here are designed to help you choose the right solution from the various configuration
options of the PRIMERGY RX2530 M1 in the light of disk-I/O performance. Various combinations of RAID
controllers and storage media will be analyzed below. Information on the selection of storage media
themselves is to be found in the section “Disk I/O: Performance of storage media”.
Hard disks
The hard disks are the first essential component. If there is a reference below to “hard disks”, this is meant
as the generic term for HDDs (“hard disk drives”, in other words conventional hard disks) and SSDs (“solid
state drives”, i.e. non-volatile electronic storage media).
Mixed drive configurations of SAS and SATA hard disks in one system are permitted, unless they are
excluded in the configurator for special hard disk types.
More hard disks per system are possible as a result of using 2.5" hard disks instead of 3.5" hard disks.
Consequently, the load that each individual hard disk has to overcome decreases and the maximum overall
performance of the system increases.
More detailed performance statements about hard disk types are available in the section “Disk I/O:
Performance of storage media” in this performance report.
Model versions
The maximum number of hard disks in the system depends on the system configuration. The following table
lists the essential cases.
Form
factor
Connection
type
Interface
Number of PCIe Maximum number
controllers
of hard disks
2.5", 3.5"
SATA 6G
direct
0
4
3.5"
SATA 6G, SAS 12G
direct
1
4
2.5"
SATA 6G, SAS 12G
direct
1
8
2.5"
SATA 6G, SAS 12G
Expander
1
10
RAID controller
In addition to the hard disks the RAID controller is the second performance-determining key component. In
the case of these controllers the “modular RAID” concept of the PRIMERGY servers offers a plethora of
options to meet the various requirements of a wide range of different application scenarios.
The following table summarizes the most important features of the available RAID controllers of the
PRIMERGY RX2530 M1. A short alias is specified here for each controller, which is used in the subsequent
list of the performance values.
Controller name
Alias
Cache
Supported
interfaces
In the system
Max. # disks
per controller
BBU/
FBU
RAID levels
LSI SW RAID on Intel
C610 (Onboard SATA)
Onboard C610
-
SATA 6G
-
4 × 2.5"
4 × 3.5"
0, 1, 10
-/-
PRAID CP400i
PRAID CP400i
-
SATA 6G
SAS 12G
PCIe 3.0
x8
8 × 2.5"
4 × 3.5"
0, 1, 1E, 5,
10, 50
-/-
PRAID EP400i
PRAID EP400i
1 GB
SATA 6G
SAS 12G
PCIe 3.0
x8
10 × 2.5"
4 × 3.5"
0, 1, 1E, 5, 6,
10, 50, 60
-/
PRAID EP420i
PRAID EP420i
2 GB
SATA 6G
SAS 12G
PCIe 3.0
x8
10 × 2.5"
4 × 3.5"
0, 1, 1E, 5, 6,
10, 50, 60
-/
The onboard RAID controller is implemented in the chipset Intel C610 on the system board of the server and
uses the CPU of the server for the RAID functionality. This controller is a simple solution that does not
require a PCIe slot.
Page 20 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
System-specific interfaces
The interfaces of a controller in CPU direction (PCIe or in the event of onboard controllers “Direct Media
Interface”, DMI in short) and in the direction of hard disks (SAS or SATA) have in each case specific limits for
data throughput. These limits are listed in the following table. The minimum of these two values is a definite
limit, which cannot be exceeded. This value is highlighted in bold in the following table.
Controller
alias
Effective in the configuration
Onboard C610
4 × SATA 6G
2060 MB/s
4 × DMI 2.0
1716 MB/s
-
PRAID CP400i
8 × SAS 12G
8240 MB/s
8 × PCIe 3.0
6761 MB/s
-/
PRAID EP400i
8 × SAS 12G
8240 MB/s
8 × PCIe 3.0
6761 MB/s
-/
PRAID EP420i
8 × SAS 12G
8240 MB/s
8 × PCIe 3.0
6761 MB/s
-/
# Disk-side
data channels
Limit for
throughput of
disk interface
# CPU-side
data channels
Limit for
throughput of
CPU-side
interface
Connection
via
expander
More details about the RAID controllers of the PRIMERGY systems are available in the white paper “RAID
Controller Performance”.
http://ts.fujitsu.com/primergy
Page 21 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Settings
In most cases, the cache of HDDs has a great influence on disk-I/O performance. It is frequently regarded as
a security problem in case of power failure and is thus switched off. On the other hand, it was integrated by
hard disk manufacturers for the good reason of increasing the write performance. For performance reasons it
is therefore advisable to enable the hard disk cache. To prevent data loss in case of power failure you are
recommended to equip the system with a UPS.
In the case of controllers with a cache there are several parameters that can be set. The optimal settings can
depend on the RAID level, the application scenario and the type of data medium. In the case of RAID levels
5 and 6 in particular (and the more complex RAID level combinations 50 and 60) it is obligatory to enable the
controller cache for application scenarios with write share. If the controller cache is enabled, the data
temporarily stored in the cache should be safeguarded against loss in case of power failure. Suitable
accessories are available for this purpose (e.g. a BBU or FBU).
For the purpose of easy and reliable handling of the settings for RAID controllers and hard disks it is
advisable to use the RAID-Manager software “ServerView RAID” that is supplied for PRIMERGY servers. All
the cache settings for controllers and hard disks can usually be made en bloc – specifically for the
application – by using the pre-defined modi “Performance” or “Data Protection”. The “Performance” mode
ensures the best possible performance settings for the majority of the application scenarios.
More information about the setting options of the controller cache is available in the white paper “RAID
Controller Performance”.
Performance values
In general, disk-I/O performance of a RAID array depends on the type and number of hard disks, on the
RAID level and on the RAID controller. If the limits of the system-specific interfaces are not exceeded, the
statements on disk-I/O performance are therefore valid for all PRIMERGY systems. This is why all the
performance statements of the document “RAID Controller Performance” also apply for the PRIMERGY
RX2530 M1 if the configurations measured there are also supported by this system.
The performance values of the PRIMERGY RX2530 M1 are listed in table form below, specifically for
different RAID levels, access types and block sizes. Substantially different configuration versions are dealt
with separately. The established measurement variables, as already mentioned in the subsection
Benchmark description, are used here. Thus, transaction rate is specified for random accesses and data
throughput for sequential accesses. To avoid any confusion among the measurement units the tables have
been separated for the two access types.
The table cells contain the maximum achievable values. This has three implications: On the one hand hard
disks with optimal performance were used (the components used are described in more detail in the
subsection Benchmark environment). Furthermore, cache settings of controllers and hard disks, which are
optimal for the respective access scenario and the RAID level, are used as a basis. And ultimately each
value is the maximum value for the entire load intensity range (# of outstanding I/Os).
In order to also visualize the numerical values each table cell is highlighted with a horizontal bar, the length
of which is proportional to the numerical value in the table cell. All bars shown in the same scale of length
have the same color. In other words, a visual comparison only makes sense for table cells with the same
colored bars.
Since the horizontal bars in the table cells depict the maximum achievable performance values, they are
shown by the color getting lighter as you move from left to right. The light shade of color at the right end of
the bar tells you that the value is a maximum value and can only be achieved under optimal prerequisites.
The darker the shade becomes as you move to the left, the more frequently it will be possible to achieve the
corresponding value in practice.
Page 22 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
2.5" - Random accesses (maximum performance values in IO/s):
Onboard C610 SSDSC2BA400G3C SATA SSD
HUC156045CSS204 SAS HDD
PRAID CP400i
PX02SMF040 SAS SSD
SSDs random
64 kB blocks
67% read
[IO/s]
SSDs random
8 kB blocks
67% read
[IO/s]
HDDs random
64 kB blocks
67% read
[IO/s]
RAID level
#Disks
Hard disk
type
RAID
Controller
Configuration version
HDDs random
8 kB blocks
67% read
[IO/s]
PRIMERGY RX2530 M1
Model version PY RX2530 M1 4x 2.5' expandable
Model version PY RX2530 M1 10x 2.5'
2
1 N/A
N/A
47337
7870
4
4
0 N/A
10 N/A
N/A
N/A
78887
63426
14951
12256
2
1
1290
1112
75925
12445
8
8
8
10
0
5
3948
5490
2827
2524
3466
1920
100776
137616
29911
55843
77081
19148
2
1
1394
1122
78733
12318
PRAID EP400i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
8
8
8
10
0
5
4113
5558
3292
2610
3504
2176
113462
132049
54614
58778
81445
23046
PRAID EP400i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
10
10
10
10
0
5
4902
6398
3470
3145
4298
2570
112511
133706
54063
47547
94661
22576
2
1
1495
1212
80178
12460
PRAID EP420i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
8
8
8
10
0
5
4212
5452
3194
2804
3506
2010
105915
123219
54214
58569
79893
22894
PRAID EP420i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
10
10
10
10
0
5
4876
6382
3622
3328
4348
2439
112459
133049
54582
48085
94396
22812
http://ts.fujitsu.com/primergy
Page 23 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
2.5" - Sequential accesses (maximum performance values in MB/s):
Onboard C610 SSDSC2BA400G3C SATA SSD
HUC156045CSS204 SAS HDD
PRAID CP400i
PX02SMF040 SAS SSD
SSDs sequential
64 kB blocks
100% write
[MB/s]
SSDs sequential
64 kB blocks
100% read
[MB/s]
HDDs sequential
64 kB blocks
100% write
[MB/s]
RAID level
#Disks
Hard disk
type
RAID
Controller
Configuration version
HDDs sequential
64 kB blocks
100% read
[MB/s]
PRIMERGY RX2530 M1
Model version PY RX2530 M1 4x 2.5' expandable
Model version PY RX2530 M1 10x 2.5'
2
1 N/A
N/A
726
443
4
4
0 N/A
10 N/A
N/A
N/A
1264
1027
1190
605
2
1
394
235
1603
420
8
8
8
10
0
5
1006
1816
1577
913
1820
1583
5918
5838
5844
1652
3295
1868
2
1
411
235
1596
420
PRAID EP400i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
8
8
8
10
0
5
1001
1836
1600
926
1808
1591
5873
5818
5790
1653
3295
2651
PRAID EP400i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
10
10
10
10
0
5
1251
2295
2046
1149
2256
1974
5909
5889
5899
2035
4051
2743
2
1
440
274
1595
421
PRAID EP420i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
8
8
8
10
0
5
1027
1919
1636
958
1851
1605
5888
5848
5847
1650
3281
2611
PRAID EP420i
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
10
10
10
10
0
5
1269
2334
2033
1151
2325
2070
5905
5899
5900
2034
4050
2684
Page 24 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
3.5" - Random accesses (maximum performance values in IO/s):
Onboard C610
PRAID CP400i
PRAID EP400i
PRAID EP420i
ST3000NM0033 SATA HDD
SSDSC2BA400G3C SATA SSD
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
SSDs random
64 kB blocks
67% read
[IO/s]
SSDs random
8 kB blocks
67% read
[IO/s]
HDDs random
64 kB blocks
67% read
[IO/s]
RAID level
#Disks
Hard disk
type
RAID
Controller
Configuration version
HDDs random
8 kB blocks
67% read
[IO/s]
PRIMERGY RX2530 M1
Model version PY RX2530 M1 4x 3.5'
2
1
487
435
47337
7870
4
4
0
10
1081
813
609
464
78887
63426
14951
12256
2
1
1290
1112
75925
12445
4
4
4
10
0
5
2216
2634
1578
1251
1526
953
101722
131832
28395
21815
41626
16470
2
1
1394
1122
78733
12318
HUC156045CSS204 SAS HDD
4
10
2045
1288
112215
22104
PX02SMF040 SAS SSD
4
4
0
5
2681
1516
1770
971
128779
36954
41399
13079
2
1
1495
1212
80178
12460
HUC156045CSS204 SAS HDD
4
10
2133
1397
111969
21776
PX02SMF040 SAS SSD
4
4
0
5
2697
2518
1747
1157
128792
36908
42029
13835
http://ts.fujitsu.com/primergy
Page 25 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
3.5" - Sequential accesses (maximum performance values in MB/s):
Onboard C610
PRAID CP400i
PRAID EP400i
PRAID EP420i
ST3000NM0033 SATA HDD
SSDSC2BA400G3C SATA SSD
HUC156045CSS204 SAS HDD
PX02SMF040 SAS SSD
SSDs sequential
64 kB blocks
100% write
[MB/s]
SSDs sequential
64 kB blocks
100% read
[MB/s]
HDDs sequential
64 kB blocks
100% write
[MB/s]
RAID level
#Disks
Hard disk
type
RAID
Controller
Configuration version
HDDs sequential
64 kB blocks
100% read
[MB/s]
PRIMERGY RX2530 M1
Model version PY RX2530 M1 4x 3.5'
2
1
178
173
726
443
4
4
0
10
667
350
672
336
1264
1027
1190
605
2
1
394
235
1603
420
4
4
4
10
0
5
537
913
682
459
914
686
3239
3224
3206
822
1649
1162
2
1
411
235
1596
420
HUC156045CSS204 SAS HDD
4
10
549
472
3270
831
PX02SMF040 SAS SSD
4
4
0
5
946
698
938
673
3274
3189
1660
1238
421
2
1
440
274
1595
HUC156045CSS204 SAS HDD
4
10
576
480
3249
837
PX02SMF040 SAS SSD
4
4
0
5
963
726
958
677
3233
3197
1670
1251
Conclusion
At full configuration with powerful hard disks the PRIMERGY RX2530 M1 achieves a throughput of up to
5918 MB/s for sequential load profiles and a transaction rate of up to 137616 IO/s for typical, random
application scenarios.
For best possible performance we recommend one of the plug-in PCIe controllers. To operate SSDs within
the maximum performance range the PRAID CP400i is already suited for the simpler RAID levels 0, 1 and
10, and a controller with cache is to be preferred for RAID 5.
In the event of HDDs the controller cache for random load profiles with a significant write share has
performance advantages for all RAID levels.
Page 26 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
OLTP-2
Benchmark description
OLTP stands for Online Transaction Processing. The OLTP-2 benchmark is based on the typical application
scenario of a database solution. In OLTP-2 database access is simulated and the number of transactions
achieved per second (tps) determined as the unit of measurement for the system.
In contrast to benchmarks such as SPECint and TPC-E, which were standardized by independent bodies
and for which adherence to the respective rules and regulations are monitored, OLTP-2 is an internal
benchmark of Fujitsu. OLTP-2 is based on the well-known database benchmark TPC-E. OLTP-2 was
designed in such a way that a wide range of configurations can be measured to present the scaling of a
system with regard to the CPU and memory configuration.
Even if the two benchmarks OLTP-2 and TPC-E simulate similar application scenarios using the same load
profiles, the results cannot be compared or even treated as equal, as the two benchmarks use different
methods to simulate user load. OLTP-2 values are typically similar to TPC-E values. A direct comparison, or
even referring to the OLTP-2 result as TPC-E, is not permitted, especially because there is no priceperformance calculation.
Further information can be found in the document Benchmark Overview OLTP-2.
Benchmark environment
The measurement set-up is symbolically illustrated below:
Tier A
Driver
Network
Network
Application Server
Clients
Tier B
Database Server
Disk
subsystem
System Under Test (SUT)
All results were determined by way of example on a PRIMERGY RX2540 M1.
http://ts.fujitsu.com/primergy
Page 27 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Database Server (Tier B)
Hardware
Model
PRIMERGY RX2540 M1
Processor
Intel Xeon Processor E5-2600 v3 Product Family
Memory
1 processor:
2 processors:
Network interface
2 × onboard LAN 10 Gb/s
Disk subsystem
RX2540 M1:
Onboard RAID controller PRAID EP400i
2 × 300 GB 15k rpm SAS Drive, RAID1 (OS),
4 × 450 GB 15k rpm SAS Drive, RAID10 (LOG)
5 × LSI MegaRAID SAS 9286CV-8e or 5 × PRAID EP420e
(same performance with OLTP-2)
5 × JX40: 13 × 400 GB SSD Drive each, RAID5 (data)
®
®
8 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
16 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
Software
BIOS
Version R1.0.0
Operating system
Microsoft Windows Server 2012 R2 Standard
Database
Microsoft SQL Server 2014 Enterprise
Application Server (Tier A)
Hardware
Model
1 × PRIMERGY RX200 S8
Processor
2 × Xeon E5-2667 v2
Memory
64 GB, 1600 MHz registered ECC DDR3
Network interface
2 × onboard LAN 1 Gb/s
1 × Dual Port LAN 10 Gb/s
Disk subsystem
2 × 250 GB 7.2k rpm SATA Drive
Software
Operating system
Microsoft Windows Server 2012 Standard
Client
Hardware
Model
2 × PRIMERGY RX200 S7
Processor
2 × Xeon E5-2670
Memory
32 GB, 1600 MHz registered ECC DDR3
Network interface
2 × onboard LAN 1 Gb/s
1 × Dual Port LAN 1Gb/s
Disk subsystem
1 × 250 GB 7.2k rpm SATA Drive
Software
Operating system
Microsoft Windows Server 2008 R2 Standard
Benchmark
OLTP-2 Software EGen version 1.13.0
Some components may not be available in all countries / sales regions.
Page 28 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
Database performance greatly depends on the configuration options with CPU, memory and on the
connectivity of an adequate disk subsystem for the database. In the following scaling considerations for the
processors we assume that both the memory and the disk subsystem has been adequately chosen and is
not a bottleneck.
A guideline in the database environment for selecting main memory is that sufficient quantity is more
important than the speed of the memory accesses. This why a configuration with a total memory of 512 GB
was considered for the measurements with two processors and a configuration with a total memory of
256 GB for the measurements with one processor. Both memory configurations have memory access of
2133 MHz. Further information about memory performance can be found in the White Paper Memory
performance of Xeon E5-2600 v3 (Haswell-EP)-based systems.
The following diagram shows the OLTP-2 transaction rates that can be achieved with one and two
®
®
processors of the Intel Xeon Processor E5-2600 v3 Product Family.
OLTP-2 tps
E5-2699 v3 - 18C, HT
3768.70
2136.52
E5-2698 v3 - 16C, HT
3435.90
1915.90
E5-2697 v3 - 14C, HT
3235.78
1842.92
E5-2695 v3 - 14C, HT
2985.40
1689.76
E5-2683 v3 - 14C, HT
1606.03
E5-2690 v3 - 12C, HT
1586.30
E5-2680 v3 - 12C, HT
2889.68
2838.84
2765.09
1545.09
E5-2670 v3 - 12C, HT
2601.38
1454.37
E5-2650L v3 - 12C, HT
2188.13
1219.78
E5-2660 v3 - 10C, HT
2357.78
1310.06
E5-2650 v3 - 10C, HT
1240.39
E5-2667 v3 - 8C, HT
1224.02
E5-2640 v3 - 8C, HT
1068.08
E5-2630L v3 - 8C, HT
874.06
E5-2630 v3 - 8C, HT
1019.98
E5-2643 v3 - 6C, HT
979.99
E5-2620 v3 - 6C, HT
773.16
732.37
409.34
684.82
372.70
E5-2609 v3 - 6C
E5-2603 v3 - 6C
2232.40
2244.71
1996.46
1623.49
1906.56
1823.73
1449.64
2CPUs 512GB RAM
1CPU 256GB RAM
E5-2637 v3 - 4C, HT
667.62
E5-2623 v3 - 4C, HT
608.60
0
HT:
Hyper-Threading
http://ts.fujitsu.com/primergy
500
1247.97
1128.35
1000
1500
2000
2500
3000
3500
4000
tps
bold:
measured
cursive: calculated
Page 29 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
It is evident that a wide performance range is covered by the variety of released processors. If you compare
the OLTP-2 value of the processor with the lowest performance (Xeon E5-2603 v3) with the value of the
processor with the highest performance (Xeon E5-2699 v3), the result is a 5.2-fold increase in performance.
The features of the processors are summarized in the section “Technical data”.
The relatively large performance differences between the processors can be explained by their features. The
values scale on the basis of the number of cores, the size of the L3 cache and the CPU clock frequency and
as a result of the features of Hyper-Threading and turbo mode, which are available in most processor types.
Furthermore, the data transfer rate between processors (“QPI Speed”) also determines performance.
A low performance can be seen in the Xeon E5-2603 v3 and E5-2609 v3 processors, as they have to
manage without Hyper-Threading (HT) and turbo mode (TM).
Within a group of processors with the same number of cores scaling can be seen via the CPU clock
frequency.
If you compare the maximum achievable OLTP-2 values of the current system generation with the values
that were achieved on the predecessor systems, the result is an increase of about 52%.
Maximum OLTP-2 tps
Comparison of system generations
tps
4500
+ ~52%
4000
3500
3000
2500
2000
1500
2 × E5-2697 v2
512 GB
2 × E5-2699 v3
512 GB
1000
SQL 2012
SQL 2014
500
0
Predecessor System
Current System
TX2560 M1 RX2530 M1 RX2540 M1 RX2560 M1
Predecessor System TX300 S8
Page 30 (52)
Current System
RX200 S8
RX300 S8
RX350 S8
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
vServCon
Benchmark description
vServCon is a benchmark used by Fujitsu to compare server configurations with hypervisor with regard to
their suitability for server consolidation. This allows both the comparison of systems, processors and I/O
technologies as well as the comparison of hypervisors, virtualization forms and additional drivers for virtual
machines.
vServCon is not a new benchmark in the true sense of the word. It is more a framework that combines
already established benchmarks (or in modified form) as workloads in order to reproduce the load of a
consolidated and virtualized server environment. Three proven benchmarks are used which cover the
application scenarios database, application server and web server.
Application scenario
Benchmark
No. of logical CPU cores
Memory
Database
Sysbench (adapted)
2
1.5 GB
Java application server
SPECjbb (adapted, with 50% - 60% load)
2
2 GB
Web server
WebBench
1
1.5 GB
Each of the three application scenarios is allocated to a dedicated virtual machine (VM). Add to these a
fourth machine, the so-called idle VM. These four VMs make up a “tile”. Depending on the performance
capability of the underlying server hardware, you may as part of a measurement also have to start several
identical tiles in parallel in order to achieve a maximum performance score.
System Under Test
Database
VM
Java
VM
Web
VM
…
Database
Java
VM
VM
Database
Java
VM
VM
Database
Java
VM
VM
Web
VM
Web
VM
Web
VM
Idle
VM
Tile n
…
Idle
VM
Idle
VM
Idle
VM
Tile 3
Tile 2
Tile 1
Each of the three vServCon application scenarios provides a specific benchmark result in the form of
application-specific transaction rates for the respective VM. In order to derive a normalized score, the
individual benchmark results for one tile are put in relation to the respective results of a reference system.
The resulting relative performance values are then suitably weighted and finally added up for all VMs and
tiles. The outcome is a score for this tile number.
Starting as a rule with one tile, this procedure is performed for an increasing number of tiles until no further
significant increase in this vServCon score occurs. The final vServCon score is then the maximum of the
vServCon scores for all tile numbers. This score thus reflects the maximum total throughput that can be
achieved by running the mix defined in vServCon that consists of numerous VMs up to the possible full
utilization of CPU resources. This is why the measurement environment for vServCon measurements is
designed in such a way that only the CPU is the limiting factor and that no limitations occur as a result of
other resources.
The progression of the vServCon scores for the tile numbers provides useful information about the scaling
behavior of the “System under Test”.
A detailed description of vServCon is in the document: Benchmark Overview vServCon.
http://ts.fujitsu.com/primergy
Page 31 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
The measurement set-up is symbolically illustrated below:
Framework
controller
Server
Disk subsystem
Multiple
1Gb or 10Gb
networks
System Under Test (SUT)
Load generators
All results were determined by way of example on a PRIMERGY RX2540 M1.
System Under Test (SUT)
Hardware
®
®
Processor
Intel Xeon Processor E5-2600 v3 Product Family
Memory
1 processor:
2 processors:
Network interface
1 × dual port 1GbE adapter
1 × dual port 10GbE server adapter
Disk subsystem
1 × dual-channel FC controller Emulex LPe12002
ETERNUS DX80 storage systems:
Each Tile: 50 GB LUN
Each LUN: RAID 0 with 2 × Seagate ST3300657SS-Disks (15 krpm)
8 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
16 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
Software
Operating system
VMware ESXi 5.5.0 U2 Build 2068190
Load generator (incl. Framework controller)
Hardware (Shared)
Enclosure
PRIMERGY BX900
Hardware
Model
18 × PRIMERGY BX920 S1 server blades
Processor
2 × Xeon X5570
Memory
12 GB
Network interface
3 × 1 Gbit/s LAN
Software
Operating system
Page 32 (52)
Microsoft Windows Server 2003 R2 Enterprise with Hyper-V
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Load generator VM (per tile 3 load generator VMs on various server blades)
Hardware
Processor
1 × logical CPU
Memory
512 MB
Network interface
2 × 1 Gbit/s LAN
Software
Operating system
Microsoft Windows Server 2003 R2 Enterprise Edition
Some components may not be available in all countries or sales regions.
Benchmark results
®
The PRIMERGY dual-socket rack and tower systems dealt with here are based on processors of the Intel
®
Xeon Processor E5-2600 v3 Product Family. The features of the processors are summarized in the section
“Technical data”.
The available processors of these systems with their results can be seen in the following table.
Score
#Tiles
4 Cores
Hyper-Threading, Turbo Mode
E5-2623 v3
E5-2637 v3
7.71
8.65
4
4
6 Cores
E5-2603 v3
E5-2609 v3
5.13
5.83
5
5
6 Cores
Hyper-Threading, Turbo Mode
E5-2620 v3
E5-2643 v3
10.1
13.1
6
6
8 Cores
Hyper-Threading, Turbo Mode
E5-2630L v3
E5-2630 v3
E5-2640 v3
E5-2667 v3
11.4
13.6
14.1
15.9
8
8
8
8
10 Cores
Hyper-Threading, Turbo Mode
E5-2650 v3
E5-2660 v3
16.6
17.8
10
10
12 Cores
Hyper-Threading, Turbo Mode
E5-2650L v3
E5-2670 v3
E5-2680 v3
E5-2690 v3
16.2
20.0
21.4
22.4
11
12
12
13
14 Cores
Hyper-Threading, Turbo Mode
E5-2683 v3
E5-2695 v3
E5-2697 v3
21.6
23.5
25.5
14
14
15
16 Cores
Hyper-Threading, Turbo Mode
E5-2698 v3
27.3
16
18 Cores
Hyper-Threading, Turbo Mode
E5-2699 v3
30.3
18
®
®
Intel Xeon Processor
E5 v3 Product Family
Processor
These PRIMERGY dual-socket rack and tower systems are very suitable for application virtualization thanks
to the progress made in processor technology. Compared with a system based on the previous processor
generation an approximate 76% higher virtualization performance can be achieved (measured in vServCon
score in their maximum configuration).
http://ts.fujitsu.com/primergy
Page 33 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
The first diagram compares the virtualization performance values that can be achieved with the processors
reviewed here.
®
®
6
8
8
8
8
10
10
11
12
12
13
E5-2637 v3 - 4 Cores
E5-2603 v3 - 6 Cores
E5-2609 v3 - 6 Cores
E5-2620 v3 - 6 Cores
E5-2643 v3 - 6 Cores
E5-2630L v3 - 8 Cores
E5-2630 v3 - 8 Cores
E5-2640 v3 - 8 Cores
E5-2667 v3 - 8 Cores
E5-2650 v3 - 10 Cores
E5-2660 v3 - 10 Cores
E5-2650L v3 - 12 Cores
E5-2670 v3 - 12 Cores
E5-2680 v3 - 12 Cores
E5-2690 v3 - 12 Cores
#Tiles
14
14
15
16
18
E5-2699 v3 - 18 Cores
6
E5-2698 v3 - 16 Cores
5
E5-2697 v3 - 14 Cores
5
E5-2695 v3 - 14 Cores
4
E5-2683 v3 - 14 Cores
4
E5-2623 v3 - 4 Cores
Intel Xeon Processor E5-2600 v3 Product Family
30
Final vServCon Score
25
20
15
10
5
0
The relatively large performance differences between the processors can be explained by their features. The
values scale on the basis of the number of cores, the size of the L3 cache and the CPU clock frequency and
as a result of the features of Hyper-Threading and turbo mode, which are available in most processor types.
Furthermore, the data transfer rate between processors (“QPI Speed”) also determines performance.
A low performance can be seen in the Xeon E5-2603 v3 and E5-2609 v3 processors, as they have to
manage without Hyper-Threading (HT) and turbo mode (TM). In principle, these weakest processors are only
to a limited extent suitable for the virtualization environment.
Within a group of processors with the same number of cores scaling can be seen via the CPU clock
frequency.
As a matter of principle, the memory access speed also influences performance. A guideline in the
virtualization environment for selecting main memory is that sufficient quantity is more important than the
speed of the memory accesses. The vServCon scaling measurements presented here were all performed
with a memory access speed – depending on the processor type – of at most 2133 MHz. More information
about the topic “Memory Performance” and QPI architecture can be found in the White Paper Memory
performance of Xeon E5-2600 v3 (Haswell-EP)-based systems.
Page 34 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Until now we have looked at the virtualization performance of a
fully configured system. However, with a server with two sockets
the question also arises as to how good performance scaling is
from one to two processors. The better the scaling, the lower the
overhead usually caused by the shared use of resources within a
server. The scaling factor also depends on the application. If the
server is used as a virtualization platform for server consolidation,
the system scales with a factor of 1.97. When operated with two
processors, the system thus achieves a significantly better
performance than with one processor, as is illustrated in the
diagram opposite using the processor version Xeon E5-2699 v3 as
an example.
× 1.97
25
15
10
5
30.3@18 tiles
20
15.4@9 tiles
Final vServCon Score
35
30
Version: 1.1  2015-04-15
1 x E5-2699 v3
2 x E5-2699 v3
0
2.65
5.47
8.20
10.7
13.3
14.8
17.2
18.9
19.8
21.0
22.0
22.7
23.4
23.5
0
2.89
5.70
8.36
10.3
11.6
12.8
13.8
14.1
vServCon Score
The next diagram illustrates the virtualization performance for increasing numbers of VMs based on the
Xeon E5-2640 v3 (8 core) and E5-2695 v2 (14 core) processors.
In addition to the increased
number of physical cores,
E5-2640 v3
E5-2695 v3
Hyper-Threading, which is
25
supported by almost all
®
processors of the Intel
®
Xeon Processor E5-2600 v3
20
Product
Family,
is
an
additional reason for the high
number of VMs that can be
operated. As is known, a
15
physical processor core is
consequently divided into two
logical cores so that the
10
number of cores available for
the hypervisor is doubled.
This standard feature thus
generally
increases
the
5
virtualization performance of
a system.
1 2 3 4 5 6 7 8
1 2 3 4 5 6 7 8 9 10 11 12 13 14
#Tiles
The previous diagram examined the total performance of all application VMs of a host. However, studying
the performance from an individual application VM viewpoint is also interesting. This information is in the
previous diagram. For example, the total optimum is reached in the above Xeon E5-2640 v3 situation with 24
application VMs (eight tiles, not including the idle VMs); the low load case is represented by three application
VMs (one tile, not including the idle VM). Remember: the vServCon score for one tile is an average value
across the three application scenarios in vServCon. This average performance of one tile drops when
changing from the low load case to the total optimum of the vServCon score - from 2.89 to 14.1/8=1.76, i.e.
to 61%. The individual types of application VMs can react very differently in the high load situation. It is thus
clear that in a specific situation the performance requirements of an individual application must be balanced
against the overall requirements regarding the numbers of VMs on a virtualization host.
http://ts.fujitsu.com/primergy
Page 35 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
The virtualization-relevant progress in processor technology since 2008 has an effect on the one hand on an
individual VM and, on the other hand, on the possible maximum number of VMs up to CPU full utilization.
The following comparison shows the proportions for both types of improvements.
Six systems with similar housing
2008
2009
2011
2012
2013
2014/2015
construction are compared: a
RX200 S4
RX200 S5 RX200 S6
RX200 S7 RX200 S8 RX2530 M1
system from 2008, a system from
RX300
S4
RX300
S5
RX300
S6
RX300 S7 RX300 S8 RX2540 M1
2009, a system from 2011, a
TX300 S6
RX350 S7 RX350 S8 RX2560 M1
system from 2012, a system from
TX300 S4
TX300 S5 TX300 S6
TX300 S7 TX300 S8 TX2560 M1
2013 and a current system with
the best processors each (see
table below) for few VMs and for highest maximum performance.
2008
2009
2011
2012
2013
2014
Best
vServCon
Best
vServCon
Performance Score
Maximum
Score
Few VMs
1 Tile Performance
max.
X5460
1.91
X5460
2.94@2 tiles
X5570
2.45
X5570
6.08@ 6 tiles
X5690
2.63
X5690
9.61@ 9 tiles
E5-2643
2.73
E5-2690
13.5@ 8 tiles
E5-2667 v2
2.85
E5-2697 v2 17.1@11 tiles
E5-2643 v3
3.22
E5-2699 v3
30.3@18tiles
The clearest performance improvements arose from 2008 to 2009 with the introduction of the Xeon 5500
1
processor generation (e. g. via the feature “Extended Page Tables” (EPT) ). One sees an increase of the
vServCon score by a factor of 1.28 with a few VMs (one tile).
Virtualization relevant improvements
10
Few VMs (1 Tile)
9
8
vServCon Score
7
6
5
× 1.13
× 1.04
× 1.04
4
× 1.07
× 1.28
3
2
1
1.91
2.45
2.63
2.73
2.85
3.22
2009
X5570
2.93 GHz
4C
2011
X5690
2.93 GHz
6C
2012
E5-2643
3.3 GHz
4C
2013
E5-2667 v2
3.3 GHz
8C
2014
E5-2643 v3
3.4 GHz
6C
0
2008
X5460
3.17 GHz
4C
Year
Y CPU
Freq.
#Cores
With full utilization of the systems with VMs there was an increase by a factor of 2.07. The one reason was
the performance increase that could be achieved for an individual VM (see score for a few VMs). The other
reason was that more VMs were possible with total optimum (via Hyper-Threading). However, it can be seen
that the optimum was “bought” with a triple number of VMs with a reduced performance of the individual VM.
1
EPT accelerates memory virtualization via hardware support for the mapping between host and guest memory
addresses.
Page 36 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Virtualization relevant improvements
35
Score at optimum Tile count
× 1.77
30
25
vServCon Score
× 1.27
20
× 1.40
15
30.3
× 1.58
× 2.07
10
17.1
13.5
5
0
9.61
2.94
2008
X5460
3.17 GHz
4C
6.08
2009
X5570
2.93 GHz
4C
2011
X5690
2.93 GHz
6C
2012
E5-2690
2.9 GHz
8C
2013
E5-2697 v2
2.7 GHz
12C
2014
E5-2699 v3
2.3 GHz
18C
Year
Y CPU
Freq.
#Cores
Where exactly is the technology progress between 2009 and 2014?
The performance for an individual VM in low-load situations has only slightly increased for the processors
compared here with the highest clock frequency per core. We must explicitly point out that the increased
virtualization performance as seen in the score cannot be completely deemed as an improvement for one
individual VM.
The decisive progress is in the higher number of physical cores and – associated with it – in the increased
values of maximum performance (factor 1.58, 1.40, 1.27 and 1.77 in the diagram).
Up to and including 2011 the best processor type of a processor generation had both the highest clock
frequency and the highest number of cores. From 2012 there have been differently optimized processors on
offer: Versions with a high clock frequency per core for few cores and versions with a high number of cores,
but with a lower clock frequency per core. The features of the processors are summarized in the section
“Technical data”.
Performance increases in the virtualization environment since 2009 are mainly achieved by increased VM
numbers due to the increased number of available logical or physical cores. However, since 2012 it has
been possible - depending on the application scenario in the virtualization environment – to also select a
CPU with an optimized clock frequency if a few or individual VMs require maximum computing power.
http://ts.fujitsu.com/primergy
Page 37 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
VMmark V2
Benchmark description
VMmark V2 is a benchmark developed by VMware to compare server configurations with hypervisor
solutions from VMware regarding their suitability for server consolidation. In addition to the software for load
generation, the benchmark consists of a defined load profile and binding regulations. The benchmark results
can be submitted to VMware and are published on their Internet site after a successful review process. After
the discontinuation of the proven benchmark “VMmark V1” in October 2010, it has been succeeded by
“VMmark V2”, which requires a cluster of at least two servers and covers data center functions, like Cloning
and Deployment of virtual machines (VMs), Load Balancing, as well as the moving of VMs with vMotion and
also Storage vMotion.
In addition to the “Performance Only” result, it is also possible from version 2.5 of VMmark to alternatively
measure the electrical power consumption and publish it as a “Performance with Server Power” result (power
consumption of server systems only) and/or “Performance with Server and Storage Power” result (power
consumption of server systems and all storage components).
VMmark V2 is not a new benchmark in the actual sense. Application scenario Load tool
# VMs
It is in fact a framework that consolidates already
LoadGen
1
established benchmarks, as workloads in order to Mail server
Web
2.0
Olio
client
2
simulate the load of a virtualized consolidated server
environment. Three proven benchmarks, which cover E-commerce
DVD Store 2 client
4
the application scenarios mail server, Web 2.0, and
Standby server
(IdleVMTest)
1
e-commerce were integrated in VMmark V2.
Each of the three application scenarios is assigned to a total of seven dedicated virtual machines. Then add
to these an eighth VM called the “standby server”. These eight VMs form a “tile”. Because of the
performance capability of the underlying server hardware, it is usually necessary to have started several
identical tiles in parallel as part of a measurement in order to achieve a maximum overall performance.
A new feature of VMmark V2 is an infrastructure component, which is present once for every two hosts. It
measures the efficiency levels of data center consolidation through VM Cloning and Deployment, vMotion
and Storage vMotion. The Load Balancing capacity of the data center is also used (DRS, Distributed
Resource Scheduler).
The result of VMmark V2 for test type „Performance Only“ is a number, known as a “score”, which provides
information about the performance of the measured virtualization solution. The score reflects the maximum
total consolidation benefit of all VMs for a server configuration with hypervisor and is used as a comparison
criterion of various hardware platforms.
This score is determined from the individual results of the VMs and an infrastructure result. Each of the five
VMmark V2 application or front-end VMs provides a specific benchmark result in the form of applicationspecific transaction rates for each VM. In order to derive a normalized score the individual benchmark results
for one tile are put in relation to the respective results of a reference system. The resulting dimensionless
performance values are then averaged geometrically and finally added up for all VMs. This value is included
in the overall score with a weighting of 80%. The infrastructure workload is only present in the benchmark
once for every two hosts; it determines 20% of the result. The number of transactions per hour and the
average duration in seconds respectively are determined for the score of the infrastructure workload
components.
In addition to the actual score, the number of VMmark V2 tiles is always specified with each VMmark V2
score. The result is thus as follows: “Score@Number of Tiles”, for example “4.20@5 tiles”.
In the case of the two test types “Performance with Server Power” and “Performance with Server and
Storage Power” a so-called “Server PPKW Score” and “Server and Storage PPKW Score” is determined,
which is the performance score divided by the average power consumption in kilowatts (PPKW =
performance per kilowatt (KW)).
The results of the three test types should not be compared with each other.
A detailed description of VMmark V2 is available in the document Benchmark Overview VMmark V2.
Page 38 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
The measurement set-up is symbolically illustrated below:
Clients & Management
Server(s)
Storage System
Multiple
1Gb or 10Gb
networks
Load Generators
incl. Prime Client and
Datacenter Management
Server
vMotion
network
System under Test (SUT)
System Under Test (SUT)
Hardware
Number of servers
2
Model
PRIMERGY RX2540 M1
Processor
Memory
2 × Xeon E5-2699 v3
“Performance Only” measurement result:
Network interface
512 GB: 16 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
“Performance with Server Power” and “Performance with Server and Storage
Power” measurement results:
384 GB: 12 × 32GB (1x32GB) 4Rx4 DDR4-2133 LR ECC
“Performance Only” measurement result:
1 × Emulex OneConnect OCe14000 Dual Port Adapter with 10Gb SFP+ DynamicLoM
interface module
1 × PLAN CP 2x1Gbit Cu Intel I350-T2 LP Adapter
“Performance with Server Power” and “Performance with Server and Storage
Power” measurement results:
1 × Emulex OneConnect OCe14000 Dual Port Adapter with 10Gb SFP+ DynamicLoM
interface module
Disk subsystem
1 × Dual port PFC EP LPe16002
2 × PRIMERGY RX300 S8 configured as Fibre Channel target:
7/6 × SAS-SSD (400 GB)
®
2 × Fusion-io ioDrive 2 PCIe-SSD (1.2 TB)
RAID 0 with several LUNs
Total: 8440 GB
Software
BIOS
Version V5.0.0.9 R1.1.0
BIOS settings
See details
Operating system
VMware ESXi 5.5.0 U2 Build 1964139
Operating system
settings
ESX settings: see details
http://ts.fujitsu.com/primergy
Page 39 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Details
See disclosure
http://www.vmware.com/a/assets/vmmark/pdf/2014-12-09-Fujitsu-RX2530M1.pdf
http://www.vmware.com/a/assets/vmmark/pdf/2014-12-23-Fujitsu-RX2530M1serverPPKW.pdf
http://www.vmware.com/a/assets/vmmark/pdf/2014-12-23-Fujitsu-RX2530M1serverstoragePPKW.pdf
Datacenter Management Server (DMS)
Hardware (Shared)
Enclosure
PRIMERGY BX600
Network Switch
1 × PRIMERGY BX600 GbE Switch Blade 30/12
Hardware
Model
1 × server blade PRIMERGY BX620 S5
Processor
2 × Xeon X5570
Memory
24 GB
Network interface
6 × 1 Gbit/s LAN
Software
Operating system
VMware ESXi 5.1.0 Build 799733
Datacenter Management Server (DMS) VM
Hardware
Processor
4 × logical CPU
Memory
10 GB
Network interface
2 × 1 Gbit/s LAN
Software
Operating system
Microsoft Windows Server 2008 R2 Enterprise x64 Edition
Prime Client
Hardware (Shared)
Enclosure
PRIMERGY BX600
Network Switch
1 × PRIMERGY BX600 GbE Switch Blade 30/12
Hardware
Model
1 × server blade PRIMERGY BX620 S5
Processor
2 × Xeon X5570
Memory
12 GB
Network interface
6 × 1 Gbit/s LAN
Software
Operating system
Page 40 (52)
Microsoft Windows Server 2008 Enterprise x64 Edition SP2
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Load generator
Hardware
Model
2 × PRIMERGY RX600 S6
Processor
4 × Xeon E7-4870
Memory
512 GB
Network interface
5 × 1 Gbit/s LAN
Software
Operating system
VMware ESX 4.1.0 U2 Build 502767
Load generator VM (per tile 1 load generator VM)
Hardware
Processor
4 × logical CPU
Memory
4 GB
Network interface
1 × 1 Gbit/s LAN
Software
Operating system
Microsoft Windows Server 2008 Enterprise x64 Edition SP2
Some components may not be available in all countries or sales regions.
http://ts.fujitsu.com/primergy
Page 41 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
th
“Performance Only” measurement result (December 9 2014)
On December 9, 2014 Fujitsu achieved with a PRIMERGY RX2530 M1 with Xeon E5-2699 v3 processors
and VMware ESXi 5.5.0 U2 a VMmark V2 score of “26.37@22 tiles” in a system configuration with a total of
2 × 36 processor cores and when using two identical servers in the “System under Test” (SUT). With this
result the PRIMERGY RX2530 M1 is in the official VMmark V2 “Performance Only” ranking the second most
powerful 2-socket server in a “matched pair” configuration consisting of two identical hosts (valid as of
benchmark results publication date).
th
All comparisons for the competitor products reflect the status of 9 December 2014. The current VMmark V2
“Performance Only” results as well as the detailed results and configuration data are available at
http://www.vmware.com/a/vmmark/.
The diagram shows the “Performance Only” result of the PRIMERGY RX2530 M1 in comparison with the
®
®
PRIMERGY RX2540 M1 and the best competitor system with 2 × 2 processors of the Intel Xeon Processor
E5-2600 Product Family (v1/v2/v3).
The PRIMERGY RX2530 M1 obtains a performance level almost identical to that of the PRIMERGY
RX2540 M1.
PRIMERGY RX2530 M1 compared to PRIMERGY RX2540 M1
and HP ProLiant DL380p Gen8
30
+59%
20
5
16.54@14 tiles
10
26.37@22 tiles
15
26.48@22 tiles
VMmark V2 Score
25
2 × Fujitsu
PRIMERGY
RX2540 M1
2 × 2 × Xeon
E5-2699 v3
2 × Fujitsu
PRIMERGY
RX2530 M1
2 × 2 × Xeon
E5-2699 v3
2 × HP
ProLiant DL380p
Gen8
2 × 2 × Xeon
E5-2697 v2
0
The processors used, which with a good hypervisor setting could make optimal use of their processor
features, were the essential prerequisites for achieving the PRIMERGY RX2530 M1 result. These features
include Hyper-Threading. All this has a particularly positive effect during virtualization.
All VMs, their application data, the host operating system as well as additionally required data were on a
powerful Fibre Channel disk subsystem. As far as possible, the configuration of the disk subsystem takes the
specific requirements of the benchmark into account. The use of flash technology in the form of SAS SSDs
and PCIe-SSDs in the powerful Fibre Channel disk subsystem resulted in further advantages in response
times of the storage medium used.
The network connection to the load generators was implemented via 10Gb LAN ports. The infrastructureworkload connection between the hosts was by means of 1Gb LAN ports.
All the components used were optimally attuned to each other.
Page 42 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
th
“Performance with Server Power” measurement result (December 23 2014)
On December 23, 2014 Fujitsu achieved with a PRIMERGY RX2530 M1 with Xeon E5-2699 v3 processors
and VMware ESXi 5.5.0 U2 a VMmark V2 “Server PPKW Score” of “25.2305@22 tiles” in a system
configuration with a total of 2 × 36 processor cores and when using two identical servers in the “System
under Test” (SUT). With this result the PRIMERGY RX2530 M1 is in the official VMmark V2 “Performance
with Server Power” ranking the second most energy-efficient virtualization server worldwide (valid as of
benchmark results publication date).
rd
All comparisons for the competitor products reflect the status of 23
December 2014. The current
VMmark V2 “Performance with Server Power” results as well as the detailed results and configuration data
are available at http://www.vmware.com/a/vmmark/2/.
The diagram shows all VMmark V2 “Performance with Server Power“ results.
The PRIMERGY RX2530 M1 as rack server system with only one unit has a very good value for this
compact construction; beaten by only about 6% by the twice as high PRIMERGY RX2540 M1, but clearly
better than the competition system with two units.
Performance with Server Power
30
20
5
17.6899@20 tiles
10
23.6493@22 tiles
15
25.2305@22 tiles
VMmark V2 Server PPKW Score
+34%
25
2 × Fujitsu
PRIMERGY
RX2540 M1
2 × 2 × Xeon
E5-2699 v3
2 × Fujitsu
PRIMERGY
RX2530 M1
2 × 2 × Xeon
E5-2699 v3
2 × HP
ProLiant DL380
Gen9
2 × 2 × Xeon
E5-2699 v3
0
http://ts.fujitsu.com/primergy
Page 43 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
th
“Performance with Server and Storage Power” measurement result (December 23 2014)
On December 23, 2014 Fujitsu achieved with a PRIMERGY RX2530 M1 with Xeon E5-2699 v3 processors
and VMware ESXi 5.5.0 U2 a VMmark V2 “Server and Storage PPKW Score” of “20.8067@22 tiles” in a
system configuration with a total of 2 × 36 processor cores and when using two identical servers in the
“System under Test” (SUT). With this result the PRIMERGY RX2530 M1 is in the official VMmark V2
“Performance with Server and Storage Power” ranking the second most energy-efficient virtualization
platform worldwide (valid as of benchmark results publication date).
rd
All comparisons for the competitor products reflect the status of 23
December 2014. The current
VMmark V2 “Performance with Server and Storage Power” results as well as the detailed results and
configuration data are available at http://www.vmware.com/a/vmmark/3/.
The diagram shows all VMmark V2 “Performance with Server and Storage Power“ results.
The PRIMERGY RX2530 M1 as a rack server system with only one unit – together with the energy-efficient
disk subsystem – has a very good value for this compact construction, beaten by approx. 5% by the twice as
high PRIMERGY RX2540 M1 with the same energy-efficient disk subsystem, but clearly better than the
overall competition configuration which contains servers with two height units.
Performance with Server and Storage Power
VMmark V2 Server and Storage PPKW Score
25
+55%
20
12.7058@20 tiles
5
19.7263@22 tiles
10
20.8067@22 tiles
15
2 × Fujitsu
PRIMERGY
RX2540 M1
2 × 2 × Xeon
E5-2699 v3
2 × Fujitsu
PRIMERGY
RX2530 M1
2 × 2 × Xeon
E5-2699 v3
2 × HP
ProLiant DL380
Gen9
2 × 2 × Xeon
E5-2699 v3
0
Page 44 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
STREAM
Benchmark description
STREAM is a synthetic benchmark that has been used for many years to determine memory throughput and
which was developed by John McCalpin during his professorship at the University of Delaware. Today
STREAM is supported at the University of Virginia, where the source code can be downloaded in either
Fortran or C. STREAM continues to play an important role in the HPC environment in particular. It is for
example an integral part of the HPC Challenge benchmark suite.
The benchmark is designed in such a way that it can be used both on PCs and on server systems. The unit
of measurement of the benchmark is GB/s, i.e. the number of gigabytes that can be read and written per
second.
STREAM measures the memory throughput for sequential accesses. These can generally be performed
more efficiently than accesses that are randomly distributed on the memory, because the processor caches
are used for sequential access.
Before execution the source code is adapted to the environment to be measured. Therefore, the size of the
data area must be at least 12 times larger than the total of all last-level processor caches so that these have
as little influence as possible on the result. The OpenMP program library is used to enable selected parts of
the program to be executed in parallel during the runtime of the benchmark, consequently achieving optimal
load distribution to the available processor cores.
During implementation the defined data area, consisting of 8-byte elements, is successively copied to four
types, and arithmetic calculations are also performed to some extent.
Type
Execution
Bytes per step
Floating-point calculation per step
COPY
a(i) = b(i)
16
0
SCALE
a(i) = q × b(i)
16
1
SUM
a(i) = b(i) + c(i)
24
1
TRIAD
a(i) = b(i) + q × c(i)
24
2
The throughput is output in GB/s for each type of calculation. The differences between the various values are
usually only minor on modern systems. In general, only the determined TRIAD value is used as a
comparison.
The measured results primarily depend on the clock frequency of the memory modules; the processors
influence the arithmetic calculations.
9
This chapter specifies throughputs on a basis of 10 (1 GB/s = 10 Byte/s).
Benchmark environment
System Under Test (SUT)
Hardware
Model
PRIMERGY RX2530 M1
Processor
2 processors of Intel Xeon Processor E5-2600 v3 Product Family
Memory
16 × 16GB (1x16GB) 2Rx4 DDR4-2133 R ECC
®
®
Software
BIOS settings
EnergyPerformance = Performance
Cores per processor < 10: COD Enable = disabled, Early Snoop = enabled
else:
COD Enable = enabled, Early Snoop = disabled
Operating system
Red Hat Enterprise Linux Server release 6.5
Operating system
settings
echo never > /sys/kernel/mm/redhat_transparent_hugepage/enabled
Compiler
Intel C++ Composer XE 2013 SP1 for Linux Update 1
Benchmark
Stream.c Version 5.9
Some components may not be available in all countries or sales regions.
http://ts.fujitsu.com/primergy
Page 45 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark results
Processor
Memory
Frequency
[MHz]
Max. Memory
Bandwidth
[GB/s]
Cores
Processor
Frequency
[GHz]
Number of
Processors
Xeon E5-2603 v3
1600
51
Xeon E5-2609 v3
1600
51
Xeon E5-2623 v3
1866
Xeon E5-2620 v3
Xeon E5-2630L v3
TRIAD
6
1.60
2
47.4
6
1.90
2
58.2
59
4
3.00
2
73.3
1866
59
6
2.40
2
88.9
1866
59
8
1.80
2
86.9
Xeon E5-2630 v3
1866
59
8
2.40
2
89.9
Xeon E5-2640 v3
1866
59
8
2.60
2
90.1
Xeon E5-2637 v3
2133
68
4
3.50
2
89.9
Xeon E5-2643 v3
2133
68
6
3.40
2
90.3
Xeon E5-2667 v3
2133
68
8
3.20
2
Xeon E5-2650 v3
2133
68
10
2.30
2
116
Xeon E5-2660 v3
2133
68
10
2.60
2
115
Xeon E5-2650L v3
2133
68
12
1.80
2
116
Xeon E5-2670 v3
2133
68
12
2.30
2
118
Xeon E5-2680 v3
2133
68
12
2.50
2
118
Xeon E5-2690 v3
2133
68
12
2.60
2
118
Xeon E5-2683 v3
2133
68
14
2.00
2
117
Xeon E5-2695 v3
2133
68
14
2.30
2
118
Xeon E5-2697 v3
2133
68
14
2.60
2
117
Xeon E5-2698 v3
2133
68
16
2.30
2
117
Xeon E5-2699 v3
2133
68
18
2.30
2
116
[GB/s]
90.3
The following diagram illustrates the throughput of the PRIMERGY RX2530 M1 in comparison to its
predecessor, the PRIMERGY RX200 S8, in their most performant configuration.
STREAM TRIAD:
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
GB/s
118
120
101
100
80
60
40
20
0
PRIMERGY RX200 S8
2 × Xeon E5-2697 v2
Page 46 (52)
PRIMERGY RX2530 M1
2 × Xeon E5-2680 v3
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
LINPACK
Benchmark description
LINPACK was developed in the 1970s by Jack Dongarra and some other people to show the performance of
supercomputers. The benchmark consists of a collection of library functions for the analysis and solution of
linear system of equations. A description can be found in the document
http://www.netlib.org/utk/people/JackDongarra/PAPERS/hplpaper.pdf.
LINPACK can be used to measure the speed of computers when solving a linear equation system. For this
purpose, an n × n matrix is set up and filled with random numbers between -2 and +2. The calculation is then
performed via LU decomposition with partial pivoting.
A memory of 8n² bytes is required for the matrix. In case of an n × n matrix the number of arithmetic
2
3
2
operations required for the solution is /3n + 2n . Thus, the choice of n determines the duration of the
measurement: a doubling of n results in an approximately eight-fold increase in the duration of the
measurement. The size of n also has an influence on the measurement result itself: as n increases, the
measured value asymptotically approaches a limit. The size of the matrix is therefore usually adapted to the
amount of memory available. Furthermore, the memory bandwidth of the system only plays a minor role for
the measurement result, but a role that cannot be fully ignored. The processor performance is the decisive
factor for the measurement result. Since the algorithm used permits parallel processing, in particular the
number of processors used and their processor cores are - in addition to the clock rate - of outstanding
significance.
LINPACK is used to measure how many floating point operations were carried out per second. The result is
referred to as Rmax and specified in GFlops (Giga Floating Point Operations per Second).
An upper limit, referred to as Rpeak, for the speed of a computer can be calculated from the maximum
number of floating point operations that its processor cores could theoretically carry out in one clock cycle:
Rpeak = Maximum number of floating point operations per clock cycle
× Number of processor cores of the computer
× Rated processor frequency[GHz]
LINPACK is classed as one of the leading benchmarks in the field of high performance computing (HPC).
LINPACK is one of the seven benchmarks currently included in the HPC Challenge benchmark suite, which
takes other performance aspects in the HPC environment into account.
Manufacturer-independent publication of LINPACK results is possible at http://www.top500.org/. The use of a
LINPACK version based on HPL is prerequisite for this (see: http://www.netlib.org/benchmark/hpl/).
Intel offers a highly optimized LINPACK version (shared memory version) for individual systems with Intel
processors. Parallel processes communicate here via "shared memory", i.e. jointly used memory. Another
version provided by Intel is based on HPL (High Performance Linpack). Intercommunication of the LINPACK
processes here takes place via OpenMP and MPI (Message Passing Interface). This enables communication
between the parallel processes - also from one computer to another. Both versions can be downloaded from
http://software.intel.com/en-us/articles/intel-math-kernel-library-linpack-download/.
Manufacturer-specific LINPACK versions also come into play when graphics cards for General Purpose
Computation on Graphics Processing Unit (GPGPU) are used. These are based on HPL and include
extensions which are needed for communication with the graphics cards.
http://ts.fujitsu.com/primergy
Page 47 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Benchmark environment
System Under Test (SUT)
Hardware
Model
PRIMERGY RX2530 M1
Processor
2 processors of Intel Xeon Processor E5-2600 v3 Product Family
Memory
16 × 16GB (1x16GB) 2Rx4 DDR4-2133 R ECC
®
®
Software
BIOS settings
EnergyPerformance = Performance
COD Enable = disabled
Early Snoop = disabled
All processors apart from Xeon E5-2603 v3 and E5-2609 v3:
Turbo Mode
= Enabled (default)
= Disabled
Hyper Threading = Disabled
Operating system
Red Hat Enterprise Linux Server release 7.0
Benchmark
Shared memory version: Intel Optimized LINPACK Benchmark 11.2 for Linux OS
Some components may not be available in all countries or sales regions.
Page 48 (52)
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Rmax (with Turbo Mode)
[GFlops]
2
384
336
88%
365
95%
Xeon E5-2637 v3
4
3.50
2
448
388
87%
388
87%
Xeon E5-2603 v3
6
1.60
2
307
273
89%
Xeon E5-2609 v3
6
1.90
2
365
321
88%
89%
Efficiency
Rpeak
[GFlops]
3.00
Efficiency
Number of processors
4
Cores
Xeon E5-2623 v3
Processor
Rated Frequency [Ghz]
Rmax (without Turbo Mode)
[GFlops]
Benchmark results
Xeon E5-2620 v3
6
2.40
2
461
411
442
96%
Xeon E5-2643 v3
6
3.40
2
653
565
87%
579
89%
Xeon E5-2630L v3
8
1.80
2
461
432
94%
454
98%
Xeon E5-2630 v3
8
2.40
2
614
575
94%
587
96%
Xeon E5-2640 v3
8
2.60
2
666
597
90%
597
90%
Xeon E5-2667 v3
8
3.20
2
819
734 (est.)
90%
734 (est.)
90%
Xeon E5-2650 v3
10
2.30
2
736
686
93%
702
95%
Xeon E5-2660 v3
10
2.60
2
832
713
86%
712
86%
Xeon E5-2650L v3
12
1.80
2
691
542
78%
541
78%
Xeon E5-2670 v3
12
2.30
2
883
823
93%
829
94%
Xeon E5-2680 v3
12
2.50
2
960
838
87%
838
87%
Xeon E5-2690 v3
12
2.60
2
998
896
90%
896
90%
Xeon E5-2683 v3
14
2.00
2
896
835
93%
874
98%
Xeon E5-2695 v3
14
2.30
2
1030
929
90%
929
90%
Xeon E5-2697 v3
14
2.60
2
1165
983 (est.)
84%
982 (est.)
84%
Xeon E5-2698 v3
16
2.30
2
1178
1084
92%
1086
92%
Xeon E5-2699 v3
18
2.30
2
1325
1185
89%
1186
90%
The results marked (est.) are estimates.
Rmax = Measurement result
Rpeak = Maximum number of floating point operations per clock cycle
× Number of processor cores of the computer
× Rated frequency [GHz]
As explained in the section "Technical Data", Intel does not as a matter of principle guarantee that the
maximum turbo frequency can be reached in the processor models due to manufacturing tolerances. A
further restriction applies for workloads, such as those generated by LINPACK: with intensive use of AVX
instructions and a high number of instructions per clock unit. Here the frequency of a core can also be limited
if the upper limits of the processor for power consumption and temperature are reached before the upper
limit for the current consumption. This can result in the achievement of a lower performance with turbo mode
than without turbo mode. In such cases, you should disable the turbo functionality via BIOS option.
http://ts.fujitsu.com/primergy
Page 49 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
System comparison
The following diagram illustrates the throughput of the PRIMERGY RX2530 M1 in comparison to its
predecessor, the PRIMERGY RX200 S8, in their most performant configuration.
LINPACK:
PRIMERGY RX2530 M1 vs. PRIMERGY RX200 S8
GFlops
1186
1200
1000
800
546
600
400
200
0
PRIMERGY RX200 S8
2 × Xeon E5-2697 v2
Page 50 (52)
PRIMERGY RX2530 M1
2 × Xeon E5-2699 v3
http://ts.fujitsu.com/primergy
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
Literature
PRIMERGY Servers
http://primergy.com/
PRIMERGY RX2530 M1
This White Paper:
http://docs.ts.fujitsu.com/dl.aspx?id=2b2fb785-30f8-4277-9d4a-c3083476c197
http://docs.ts.fujitsu.com/dl.aspx?id=8fcfa982-2a17-4eac-b796-05702e93fba2
http://docs.ts.fujitsu.com/dl.aspx?id=5b7faa9f-d6be-45a3-b14d-938f600ef1c7
Data sheet
http://docs.ts.fujitsu.com/dl.aspx?id=afc62316-7690-4222-814b-ad0203928a07
PRIMERGY Performance
http://www.fujitsu.com/fts/x86-server-benchmarks
Performance of Server Components
http://www.fujitsu.com/fts/products/computing/servers/mission-critical/benchmarks/x86components.html
BIOS optimizations for Xeon E5-2600 v3 based systems
http://docs.ts.fujitsu.com/dl.aspx?id=f154aca6-d799-487c-8411-e5b4e558c88b
Memory performance of Xeon E5-2600 v3 (Haswell-EP)-based systems
http://docs.ts.fujitsu.com/dl.aspx?id=74eb62e6-4487-4d93-be34-5c05c3b528a6
RAID Controller Performance
http://docs.ts.fujitsu.com/dl.aspx?id=e2489893-cab7-44f6-bff2-7aeea97c5aef
Disk I/O: Performance of RAID controllers
Basics of Disk I/O Performance
http://docs.ts.fujitsu.com/dl.aspx?id=65781a00-556f-4a98-90a7-7022feacc602
Information about Iometer
http://www.iometer.org
LINPACK
The LINPACK Benchmark: Past, Present, and Future
http://www.netlib.org/utk/people/JackDongarra/PAPERS/hplpaper.pdf
TOP500
http://www.top500.org/
HPL - A Portable Implementation of the High-Performance Linpack Benchmark for DistributedMemory Computers
http://www.netlib.org/benchmark/hpl/
Intel Math Kernel Library – LINPACK Download
http://software.intel.com/en-us/articles/intel-math-kernel-library-linpack-download/
OLTP-2
Benchmark Overview OLTP-2
http://docs.ts.fujitsu.com/dl.aspx?id=e6f7a4c9-aff6-4598-b199-836053214d3f
SPECcpu2006
http://www.spec.org/osg/cpu2006
Benchmark overview SPECcpu2006
http://docs.ts.fujitsu.com/dl.aspx?id=1a427c16-12bf-41b0-9ca3-4cc360ef14ce
SPECpower_ssj2008
http://www.spec.org/power_ssj2008
Benchmark Overview SPECpower_ssj2008
http://docs.ts.fujitsu.com/dl.aspx?id=166f8497-4bf0-4190-91a1-884b90850ee0
http://ts.fujitsu.com/primergy
Page 51 (52)
White Paper  Performance Report PRIMERGY RX2530 M1
Version: 1.1  2015-04-15
STREAM
http://www.cs.virginia.edu/stream/
VMmark V2
Benchmark Overview VMmark V2
http://docs.ts.fujitsu.com/dl.aspx?id=2b61a08f-52f4-4067-bbbf-dc0b58bee1bd
VMmark V2
http://www.vmmark.com
vServCon
Benchmark Overview vServCon
http://docs.ts.fujitsu.com/dl.aspx?id=b953d1f3-6f98-4b93-95f5-8c8ba3db4e59
Contact
FUJITSU
Website: http://www.fujitsu.com/
PRIMERGY Product Marketing
mailto:[email protected]
PRIMERGY Performance and Benchmarks
mailto:[email protected]
© Copyright 2015 Fujitsu Technology Solutions. Fujitsu and the Fujitsu logo are trademarks or registered trademarks of Fujitsu Limited in Japan and other
countries. Other company, product and service names may be trademarks or registered trademarks of their respective owners. Technical data subject to
modification and delivery subject to availability. Any liability that the data and illustrations are complete, actual or correct is excluded. Designations may be
trademarks and/or copyrights of the respective manufacturer, the use of which by third parties for their own purposes may infringe the rights of such owner.
For further information see http://www.fujitsu.com/fts/resources/navigation/terms-of-use.html
2015-04-15 WW EN
Page 52 (52)
http://ts.fujitsu.com/primergy