Lenovo 4th Gen 4th Generation 8GB RAM Laptop Instruction Manual
- September 6, 2024
- Lenovo
Table of Contents
- Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen
- Memory interleaving
- Balanced memory configurations
- About the tests
- Applying the balanced memory configuration guidelines
- Configuration with 1 DIMM – unbalanced
- Configuration with 2 DIMMs – unbalanced
- Configuration with 4 DIMMs – balanced
- Configuration with 6 DIMMs – unbalanced
- Configuration with 8 DIMMs – balanced
- Configuration with 12 DIMMs – unbalanced
- Configuration with 16 DIMMs – balanced
- Summary of the performance results
- Maximising memory bandwidth
- Summary
- Authors
- Balanced Memory papers
- Related product families
- Notices **
- Trademarks
- References
- Read User Manual Online (PDF format)
- Download This Manual (PDF format)
Lenovo 4th Gen 4th Generation 8GB RAM Laptop Instruction Manual
Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen
Intel Xeon Scalable Processors
Planning / Implementation
The memory subsystem is a key component of 4th and 5th Gen Intel Xeon Scalable server architecture which can greatly affect overall server performance. When properly configured, the memory subsystem can deliver extremely high memory bandwidth and low memory access latency. When the memory subsystem is incorrectly configured, memory bandwidth available to the server can become limited and overall server performance can be reduced.
This paper explains the concept of balanced memory configurations that yield the highest possible memory bandwidth from the 4th and 5th Intel Gen Xeon Scalable architectures. Memory configuration and performance for all supported memory configurations are shown and discussed to illustrate their effect on memory subsystem performance.
This paper specifically covers the 4th Gen and 5th Gen Intel Xeon Scalable processor families (formerly code named “Sapphire Rapids” and “Emerald Rapids”, respectively). For other processor families, see the Balanced Memory papers section.
The Intel processors have four integrated Memory Controllers (iMC), eight
memory channels, and sixteen memory DIMM slots. The following figure
illustrates how the Intel processor’s memory controllers are connected to
memory DIMM slots.
Figure 1. Memory DIMM connections to the Intel processor
Each integrated Memory Controller (iMC) supports two memory channels as below:
- iMC0 supports channels A and B
- iMC1 supports channels C and D
- iMC2 supports channels E and F
- iMC3 supports channels G and H
To illustrate various memory typologies for a processor, different memory
configurations will be designated as H:G:F:E:A:B:C:D where each letter
indicates the number of memory Dim Ms populated on each memory channel. As an
example, a 2:2:2:2:1:1:1:1 memory configuration has 2 memory DIMMs populated
on channels H, G, F, E and 1 memory DIMM populated on channels A, B, C, D.
Click Here To Check For Updates
Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen Intel Xeon Scalable Processors
Memory interleaving
The 4th and 5th Gen Intel Xeon Scalable processor s optimise memory accesses by creating interleave sets across the memory controllers and memory channels. For example, if two memory channels have the same total memory capacity, a 2-channel interleave set is created across the memory channels.
Interleaving enables higher memory bandwidth by spreading contiguous memory accesses across more memory channels rather than sending all memory accesses to one memory channel. In order to form an interleave set between two channels, the two channels are required to have the same channel memory capacity.
If one interleave set cannot be formed for a particular memory configuration, it is possible to have multiple interleave sets. When this happens, performance of a specific memory access depends on which memory region is being accessed and how many memory DIMMs comprise the interleave set. For this reason, memory bandwidth performance on memory configurations with multiple interleave sets can be inconsistent. Contiguous memory accesses to a memory region with fewer channels in the interleave set will have lower performance compared to accesses to a memory region with more channels in the interleave set.
Figure 2 illustrates a 4-channel interleave set which results from populating
identical memory DIMMs on channels A, C, E, G. This 4-channel interleave set
interleaves across memory controllers and between memory channels. Consecutive
addresses alternate between memory controllers with every fourth address going
to each memory channel
Figure 2. 4-channel interleave set across memory controllers and between
memory channels
Balanced memory configurations
Balanced memory configurations enable optimal interleaving which maximises memory bandwidth. Per Intel memory population rules, channels A, E, C, G must be populated with the same total capacity per channel if populated, and channels B, D, F, H must be populated with the same total capacity per channel if populated.
The basic guidelines for a balanced memory subsystem are as follows:
- All populated memory channels should have the same total memory capacity and the same number of ranks per channel.
- All memory controllers on a processor socket should have the same configuration of memory DIMMs.
- All processor sockets on the same physical server should have the same configuration of memory DIMMs.
Tip : We will refer to the above guidelines as Balanced Memory Guidelines 1, 2 and 3 throughout this paper.
About the tests
Intel Memory Latency Checker (Intel MLC) is a tool used to measure memory lateness and bandwidth. The intent of the tool is to measure the highest memory bandwidth available. Intel MLC will be used to measure the sustained memory bandwidth of various memory configurations supported by 4th and 5th Gen Intel Xeon Scalable processors.
All one DIMM per channel and two DIMMs per channel configurations were measured at the following Intel plan of record memory speeds for regular RDIMMs:
- 4th Generation Intel Xeon “Sapphire Rapids”
- One DIMM per channel – 4800MHz
- Two DIMMs per channel – 4400MHz
- 5th Generation Intel Xeon “Emerald Rapids”
- One DIMM per channel – 5600MHz
- Two DIMMs per channel – 4400MHz
For more information about Intel MLC, see the following web page:
https://www.intel.com/content/www/us/en/developer/articles/tool/intelr-
memory-latency-checker.html
Applying the balanced memory configuration guidelines
We will start with the assumption that balanced memory guideline 3 (as listed in Balanced memory configurations ) is followed: all processor sockets on the same physical server have the same configuration of memory DIMMs. Therefore, we only have to look at one processor socket to describe each memory configuration.
For a complete memory rules and population guide, please refer to the Lenovo Press product guide for each specific Think System V3 server that uses the 4th and 5th Gen processors.
In our lab measurements, all memory DIMMs used were 32GB dual-rank (2R) RDIMMs. The examples in this brief follow the recommended memory population sequence as shown in the following table.
Table 1. Memory population
iMC | iMC3 | iMC2 | CPU | iMC0 | iMC1 |
---|---|---|---|---|---|
Channel | H | G | F | E | |
DIMM Slot | 0 | 1 | 0 | 1 | 0 |
Number of DIMMs | 1 | ||||
2 | X | ||||
4 | X | ||||
6 | X | X | |||
8 | X | X | X | ||
12 | X | X | X | X | |
16 | X | X | X | X | X |
Tip : Some Think System V3 servers only implement 1 DIMM per channel.
Take that into consideration
when reviewing the memory recommendations.
Configuration with 1 DIMM – unbalanced
We will start with one memory DIMM which yields the 0:0:0:0:1:0:0:0 memory
configuration shown in the figure below.
Figure 3. 0:0:0:0:1:0:0:0 memory configuration, relative memory bandwidth:
13%
Balanced memory guideline 2 is not followed with only one iMC populated with memory DIMM. This is an unbalanced memory configuration.
A single 1-channel interleave set is formed. Having only one memory channel populated with memory greatly reduces the memory bandwidth of this configuration which was measured at 13% or about one eighth of the full potential memory bandwidth.
The best way to increase the memory bandwidth of this configuration is by using more memory DIMMs. For example, two 16GB RDIMMs populated on two channels A and G would provide the same memory capacity while nearly doubling the memory bandwidth.
Configuration with 2 DIMMs – unbalanced
The recommended memory configuration with 2 memory DIMMs is the
0:1:0:0:1:0:0:0 configuration as shown in the following figure.
Figure 4. 0:1:0:0:1:0:0:0 memory configuration, relative memory bandwidth:
25%
This memory configuration follows guideline 1, but not guideline 2 since not all iMCs were populated with identical memory configuration. This is an unbalanced memory configuration.
A single 2-channel interleave set is formed. Only two memory channels are populated with memory which greatly reduces the memory bandwidth of this memory configuration to about 25% or one fourth of the full potential memory bandwidth.
Configuration with 4 DIMMs – balanced
The recommended configuration with 4 memory DIMMs is 0:1:0:1:1:0:1:0 memory
config as shown in the following figure.
Figure 5. 0:1:0:1:1:0:1:0 memory configuration, relative memory bandwidth:
50%
This memory configuration follows both memory population guidelines 1 and 2. All populated channels have the same channel capacity, and memory configurations are identical with all iMCs. This is a balanced memory configuration.
A single 4-channel interleave set is formed. Although it is a balanced memory configuration, only four of eight memory channels were populated with memory DIMMs. Memory bandwidth is measured at 50% or one half of the full potential memory bandwidth.
Configuration with 6 DIMMs – unbalanced
The recommended configuration with 6 memory DIMMs is 1:1:0:1:1:1:1:0 memory
configuration as shown in the following figure.
Figure 6. 1:1:0:1:1:1:1:0 memory configuration, relative memory bandwidth:
75%
This memory configuration follows memory population guideline 1, but not guideline 2. Memory configurations are not identical between the iMCs. This is an unbalanced memory configuration.
A single 6-channel interleave is formed, and memory bandwidth measured at 75% of the full potential memory bandwidth. Even though a 6 DIMM population is unbalanced, the 4th and 5th Gen Intel Xeon Scalable processor design includes the technology to create a single 6-channel interleave which results in relatively good performance. Without this technology, multiple interleave sets would have been formed and performance would be degraded.
Configuration with 8 DIMMs – balanced
The recommended configuration with 8 DIMMs is 1:1:1:1:1:1:1:1 memory
configuration as shown in the following figure.
Figure 7. 1:1:1:1:1:1:1:1 memory configuration, relative memory bandwidth:
100%
This memory configuration follows both guidelines 1 and 2. All channels were populated with the same memory capacity, and memory configurations were identical between all iMCs. This is a balanced memory configuration.
A single 8-channel interleave is formed, and memory bandwidth measured at 100% of the full potential memory bandwidth.
Configuration with 12 DIMMs – unbalanced
The recommended configuration with 12 DIMMs is 1:2:1:2:2:1:2:1 memory
configuration as shown in the following figure.
Figure 8. 1:2:1:2:2:1:2:1 memory configuration, relative memory bandwidth:
50-85% with SPR and 50-75% with EMR.
This memory configuration follows guideline 2, but not guideline 1. This is an unbalanced memory configuration.
One 8-channel interleave and one 4-channel interleave set is formed, and memory bandwidth measured at 50% – 85% of the full potential memory bandwidth for 4th Gen (Sapphire Rapids) processors, and 50% – 75% of the full potential memory bandwidth for 5th Gen (Emerald Rapids) processors. The lower than potential bandwidth is the result of not utilising all eight memory channels and the Intel supported memory speeds when memory is configured with two DIMMs per channel.
Configuration with 16 DIMMs – balanced
This is a fully populated configuration 2:2:2:2:2:2:2:2 as shown in the
following figure.
Figure 9. 2:2:2:2:2:2:2:2 memory configuration, relative memory bandwidth:
85% with SPR and 76% with EMR
This is a fully populated memory configuration. It follows both guideline 1 and 2. This is a balanced memory configuration.
Each channel is populated with two dual-rank (2R) DIMMs, so the total number
of ranks per channel is 4R.
Memory bandwidth measured at 85% with SPR, and 76% with EMR of the full
potential memory bandwidth due to the memory speed dropping with two DIMMs per
channel as per Intel plan of record.
Summary of the performance results
The following table shows a summary of the relative memory bandwidth of all the memory configurations that were evaluated. It also shows the number of interleave sets formed for each and whether it is a balanced or unbalanced memory configuration.
Table 2. Summary of the performance results
Number of DIMMs populated| Configuration| Number of interleave
sets| Relative performance| Balanced or Unbalanced
---|---|---|---|---
1| 0:0:0:0:1:0:0:0| 1| 13%| Unbalanced
2| 0:1:0:0:1:0:0:0| 1| 25%| Unbalanced
4| 0:1:0:1:1:0:1:0| 1| 50%| Balanced
6| 1:1:0:1:1:1:1:0| 1| 75%| Unbalanced
8| 1:1:1:1:1:1:1:1| 1| 100%| Balanced
12| 1:2:1:2:2:1:2:1| 2| 50% – 85% -SPR50% – 75% – EMR| Unbalanced
16| 2:2:2:2:2:2:2:2| 1| 85% – SPR76% – EMR| Balanced
When using the same memory DIMM, only a memory configuration with 8 DIMMs provides the full potential memory bandwidth. This is the best memory configuration for performance. A balanced memory configuration can also be achieved with four memory DIMMs, but this configuration does not populate all the memory channels which reduces its memory bandwidth and performance. Finally, a balanced memory configuration can also be achieved with 16 memory DIMMs, although this requires two DIMMs per channel which drops the memory speed as per Intel’s plan of record.
Maximising memory bandwidth
To maximize the memory bandwidth of a server, the following rules should be followed:
- Balance the memory across the processor sockets – all processor sockets on the same physical server should have the same configuration of memory DIMMs.
- Balance the memory across the memory controllers – all memory controllers on a processor socket should have the same configuration of memory DIMMs.
- Balance the memory across the populated memory channels – all populated memory channels should have the same total memory capacity and the same total number of ranks.
Peak memory performance is achieved with 8 DIMMs per socket. Given a memory capacity requirement per server, follow these steps to get an optimal memory bandwidth configuration for your requirement:
- Determine your needed memory capacity per socket.
- Divide this memory capacity by eight to determine the minimum memory capacity needed per DDR channel.
- Round this calculated memory capacity per channel up to the closest capacity available with 1 DIMM Per Channel (DPC)
- Populate your server with eight identical DIMM combination per channel derived from step 3.
Examples:
If 512GB of total memory capacity is needed per socket, you can populate
each socket with 8x 64GB DIMMs.
If 768GB of total memory capacity is needed per socket, you can populate each
socket with 8x 96GB DIMMs.
Summary
Overall server performance is affected by the memory subsystem which can provide both high memory bandwidth and low memory access latency when properly configured. Balancing memory across the memory controllers and the memory channels produces memory configurations which can efficiently interleave memory references among its DIMMs producing the highest possible memory bandwidth. An unbalanced memory configuration can reduce the total memory bandwidth to as low as 13% of a balanced memory configuration with 8 identical DIMMs installed per processor.
Implementing all three of the balanced memory guidelines described in this paper results in balanced memory configuration and produces the best possible memory bandwidth and overall performance.
Lenovo recommends installing balanced memory population with 4, 8, or 16 DIMMs per socket. Peak memory performance is achieved with 8 DIMMs per processor.
Authors
This paper was produced by the following team of specialists:
Charles Stephan is a Senior Engineer and Technical Lead for the System
Performance Verification team in the Lenovo Performance Laboratory at the
Lenovo Infrastructure Solutions Group (ISG) campus in Morrisville, NC. His
team is responsible for analyzing the performance of storage adapters, network
adapters, various flash technologies, and complete x86 platforms. Before
transitioning to Lenovo, Charles spent 16 years at IBM as a Performance
Engineer analyzing storage subsystem performance of RAID adapters, Fibre
Channel HBAs, and storage servers for all x86 platforms. He also analyzed
performance of x86 rack systems, blades, and compute nodes. Charles holds a
Master of Science degree in Computer Information Systems from the Florida
Institute of Technology.
Redwan Rahman is a Systems Performance Verification Engineer in the Lenovo Infrastructure Solutions Group Performance Laboratory in Morrisville, NC, USA. His current role includes CPU, Memory, and PCIe subsystem analysis and performance validation against functional specifications and vendor targets. Redwan holds a Bachelor of Science in Computer Engineering from University of Massachusetts Amherst.
Thanks to the following people for their contributions to this project:
- David Watts, Lenovo Press
This paper was based on the paper Balanced Memory Configurations for 2-Socket Servers with 3rd-Gen Intel Xeon Scalable Processors. Thanks to the author of that paper:
- Nathan Pham
Balanced Memory papers
This paper is one of a series of papers on Balanced Memory configurations:
- Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen Intel Xeon Scalable Processors
- Balanced Memory Configurations with 4th Generation AMD EPYC Processors
- Balanced Memory Configurations for 2-Socket Servers with 3rd Gen Intel Xeon Scalable Processors
- Balanced Memory Configurations with 2nd Gen and 3rd Gen AMD EPYC Processors
- Balanced Memory Configurations with 2nd Gen Intel Xeon Scalable Processors
- Balanced Memory Configurations with 1st Generation Intel Xeon Scalable Processors
- Maximizing System x and ThinkServer Performance with a Balanced Memory Configuration
Related product families
Product families related to this document are the following:
- ThinkSystem SD530 V3 Server
- ThinkSystem SD550 V3 Server
- ThinkSystem SR630 V3 Server
- ThinkSystem SR650 V3 Server
- ThinkSystem SR680a V3 Server
- ThinkSystem SR780a V3 Server
- ThinkSystem ST650 V3 Server
Notices **
**
Lenovo may not offer the products, services, or features discussed in this
document in all countries. Consult your
local Lenovo representative for information on the products and services
currently available in your area. Any
reference to a Lenovo product, program, or service is not intended to state or
imply that only that Lenovo product, program, or service may be used. Any
functionally equivalent product, program, or service that does not infringe
any Lenovo intellectual property right may be used instead. However, it is the
user’s responsibility to evaluate and verify the operation of any other
product, program, or service. Lenovo may have patents or pending patent
applications covering subject matter described in this document. The
furnishing of this document does not give you any license to these patents.
You can send license inquiries, in writing, to:
Lenovo (United States), Inc.
8001 Development Drive
Morrisville, NC 27560
U.S.A.
Attention : Lenovo Director of Licensing
LENOVO PROVIDES THIS PUBLICATION ”AS IS” WITHOUT WARRANTY OF ANY KIND, EITHER
EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
NON-INFRINGEMENT,
MERCHANT ABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some jurisdictions do
not allow disclaimer of express or implied warranties in certain transactions,
therefore, this statement may not apply to you.
This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. Lenovo may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice.
The products described in this document are not intended for use in implantation or other life support applications where malfunction may result in injury or death to persons. The information contained in this document does not affect or change Lenovo product specifications or warranties. Nothing in this document shall operate as an express or implied license or indemnity under the intellectual property rights of Lenovo or third parties. All information contained in this document was obtained in specific environments and is presented as an illustration. The result obtained in other operating environments may vary. Lenovo may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you.
Any references in this publication to non-Lenovo Web sites are provided for
convenience only and do not in any
manner serve as an endorsement of those Web sites. The materials at those Web
sites are not part of the materials
for this Lenovo product, and use of those Web sites is at your own risk. Any
performance data contained herein was
determined in a controlled environment. Therefore, the result obtained in
other operating environments may vary
significantly. Some measurements may have been made on development-level
systems and there is no guarantee
that these measurements will be the same on generally available systems.
Furthermore, some measurements may
have been estimated through extrapolation. Actual results may vary. Users of
this document should verify the applicable data for their specific
environment.
© Copyright Lenovo 2024. All rights reserved.
This document, LP2000, was created or updated on August 13, 2024.
Send us your comments in one of the following ways:
Use the online Contact us review form found at:
https://lenovopress.lenovo.com/LP2000
Send your comments in an e-mail to:
comments@lenovopress.com
This document is available online at https://lenovopress.lenovo.com/LP2000.
Trademarks
Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.
The following terms are trademarks of Lenovo in the United States, other
countries, or both:
Lenovo®
System x®
ThinkServer®
ThinkSystem®
The following terms are trademarks of other companies:
AMD and AMD EPYC™ are trademarks of Advanced Micro Devices, Inc.
Intel® and Xeon® are trademarks of Intel Corporation or its subsidiaries.
Other company, product, or service names may be trademarks or service marks of
others.
References
- lenovopress.lenovo.com/assets/images/LP2000/Figure1.png
- lenovopress.lenovo.com/assets/images/LP2000/Figure2.png
- Copyright and Trademark Information | Lenovo US | Lenovo US
Read User Manual Online (PDF format)
Read User Manual Online (PDF format) >>