Lenovo 4th Gen 4th Generation 8GB RAM Laptop Instruction Manual

September 6, 2024
Lenovo

Lenovo 4th Gen 4th Generation 8GB RAM Laptop Instruction Manual

Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen

Intel Xeon Scalable Processors

Planning / Implementation

The memory subsystem is a key component of 4th and 5th Gen Intel Xeon Scalable server architecture which can greatly affect overall server performance. When properly configured, the memory subsystem can deliver extremely high memory bandwidth and low memory access latency. When the memory subsystem is incorrectly configured, memory bandwidth available to the server can become limited and overall server performance can be reduced.

This paper explains the concept of balanced memory configurations that yield the highest possible memory bandwidth from the 4th and 5th Intel Gen Xeon Scalable architectures. Memory configuration and performance for all supported memory configurations are shown and discussed to illustrate their effect on memory subsystem performance.

This paper specifically covers the 4th Gen and 5th Gen Intel Xeon Scalable processor families (formerly code named “Sapphire Rapids” and “Emerald Rapids”, respectively). For other processor families, see the Balanced Memory papers section.

The Intel processors have four integrated Memory Controllers (iMC), eight memory channels, and sixteen memory DIMM slots. The following figure illustrates how the Intel processor’s memory controllers are connected to memory DIMM slots.
Planning Implementation
Figure 1. Memory DIMM connections to the Intel processor

Each integrated Memory Controller (iMC) supports two memory channels as below:

  • iMC0 supports channels A and B
  • iMC1 supports channels C and D
  • iMC2 supports channels E and F
  • iMC3 supports channels G and H

To illustrate various memory typologies for a processor, different memory configurations will be designated as H:G:F:E:A:B:C:D where each letter indicates the number of memory Dim Ms populated on each memory channel. As an example, a 2:2:2:2:1:1:1:1 memory configuration has 2 memory DIMMs populated on channels H, G, F, E and 1 memory DIMM populated on channels A, B, C, D.
Click Here To Check For Updates

Balanced Memory Configurations for 2-Socket Servers with 4th and 5th Gen Intel Xeon Scalable Processors

Memory interleaving

The 4th and 5th Gen Intel Xeon Scalable processor s optimise memory accesses by creating interleave sets across the memory controllers and memory channels. For example, if two memory channels have the same total memory capacity, a 2-channel interleave set is created across the memory channels.

Interleaving enables higher memory bandwidth by spreading contiguous memory accesses across more memory channels rather than sending all memory accesses to one memory channel. In order to form an interleave set between two channels, the two channels are required to have the same channel memory capacity.

If one interleave set cannot be formed for a particular memory configuration, it is possible to have multiple interleave sets. When this happens, performance of a specific memory access depends on which memory region is being accessed and how many memory DIMMs comprise the interleave set. For this reason, memory bandwidth performance on memory configurations with multiple interleave sets can be inconsistent. Contiguous memory accesses to a memory region with fewer channels in the interleave set will have lower performance compared to accesses to a memory region with more channels in the interleave set.

Figure 2 illustrates a 4-channel interleave set which results from populating identical memory DIMMs on channels A, C, E, G. This 4-channel interleave set interleaves across memory controllers and between memory channels. Consecutive addresses alternate between memory controllers with every fourth address going to each memory channel
Memory interleaving
Figure 2. 4-channel interleave set across memory controllers and between memory channels

Balanced memory configurations

Balanced memory configurations enable optimal interleaving which maximises memory bandwidth. Per Intel memory population rules, channels A, E, C, G must be populated with the same total capacity per channel if populated, and channels B, D, F, H must be populated with the same total capacity per channel if populated.

The basic guidelines for a balanced memory subsystem are as follows:

  1. All populated memory channels should have the same total memory capacity and the same number of ranks per channel.
  2. All memory controllers on a processor socket should have the same configuration of memory DIMMs.
  3. All processor sockets on the same physical server should have the same configuration of memory DIMMs.

Tip : We will refer to the above guidelines as Balanced Memory Guidelines 1, 2 and 3 throughout this paper.

About the tests

Intel Memory Latency Checker (Intel MLC) is a tool used to measure memory lateness and bandwidth. The intent of the tool is to measure the highest memory bandwidth available. Intel MLC will be used to measure the sustained memory bandwidth of various memory configurations supported by 4th and 5th Gen Intel Xeon Scalable processors.

All one DIMM per channel and two DIMMs per channel configurations were measured at the following Intel plan of record memory speeds for regular RDIMMs:

  • 4th Generation Intel Xeon “Sapphire Rapids”
    • One DIMM per channel – 4800MHz
    • Two DIMMs per channel – 4400MHz
  • 5th Generation Intel Xeon “Emerald Rapids”
    • One DIMM per channel – 5600MHz
    • Two DIMMs per channel – 4400MHz

For more information about Intel MLC, see the following web page:
https://www.intel.com/content/www/us/en/developer/articles/tool/intelr- memory-latency-checker.html

Applying the balanced memory configuration guidelines

We will start with the assumption that balanced memory guideline 3 (as listed in Balanced memory configurations ) is followed: all processor sockets on the same physical server have the same configuration of memory DIMMs. Therefore, we only have to look at one processor socket to describe each memory configuration.

For a complete memory rules and population guide, please refer to the Lenovo Press product guide for each specific Think System V3 server that uses the 4th and 5th Gen processors.

In our lab measurements, all memory DIMMs used were 32GB dual-rank (2R) RDIMMs. The examples in this brief follow the recommended memory population sequence as shown in the following table.

Table 1. Memory population

iMC iMC3 iMC2 CPU iMC0 iMC1
Channel H G F E
DIMM Slot 0 1 0 1 0
Number of DIMMs 1
2 X
4 X
6 X X
8 X X X
12 X X X X
16 X X X X X

Tip : Some Think System V3 servers only implement 1 DIMM per channel. Take that into consideration
when reviewing the memory recommendations.

Configuration with 1 DIMM – unbalanced

We will start with one memory DIMM which yields the 0:0:0:0:1:0:0:0 memory configuration shown in the figure below.
Configuration with 1 DIMM – unbalanced
Figure 3. 0:0:0:0:1:0:0:0 memory configuration, relative memory bandwidth: 13%

Balanced memory guideline 2 is not followed with only one iMC populated with memory DIMM. This is an unbalanced memory configuration.

A single 1-channel interleave set is formed. Having only one memory channel populated with memory greatly reduces the memory bandwidth of this configuration which was measured at 13% or about one eighth of the full potential memory bandwidth.

The best way to increase the memory bandwidth of this configuration is by using more memory DIMMs. For example, two 16GB RDIMMs populated on two channels A and G would provide the same memory capacity while nearly doubling the memory bandwidth.

Configuration with 2 DIMMs – unbalanced

The recommended memory configuration with 2 memory DIMMs is the 0:1:0:0:1:0:0:0 configuration as shown in the following figure.
Configuration with 2 DIMMs – unbalanced
Figure 4. 0:1:0:0:1:0:0:0 memory configuration, relative memory bandwidth: 25%

This memory configuration follows guideline 1, but not guideline 2 since not all iMCs were populated with identical memory configuration. This is an unbalanced memory configuration.

A single 2-channel interleave set is formed. Only two memory channels are populated with memory which greatly reduces the memory bandwidth of this memory configuration to about 25% or one fourth of the full potential memory bandwidth.

Configuration with 4 DIMMs – balanced

The recommended configuration with 4 memory DIMMs is 0:1:0:1:1:0:1:0 memory config as shown in the following figure.
Configuration with 4 DIMMs – balanced
Figure 5. 0:1:0:1:1:0:1:0 memory configuration, relative memory bandwidth: 50%

This memory configuration follows both memory population guidelines 1 and 2. All populated channels have the same channel capacity, and memory configurations are identical with all iMCs. This is a balanced memory configuration.

A single 4-channel interleave set is formed. Although it is a balanced memory configuration, only four of eight memory channels were populated with memory DIMMs. Memory bandwidth is measured at 50% or one half of the full potential memory bandwidth.

Configuration with 6 DIMMs – unbalanced

The recommended configuration with 6 memory DIMMs is 1:1:0:1:1:1:1:0 memory configuration as shown in the following figure.
Configuration with 6 DIMMs – unbalanced
Figure 6. 1:1:0:1:1:1:1:0 memory configuration, relative memory bandwidth: 75%

This memory configuration follows memory population guideline 1, but not guideline 2. Memory configurations are not identical between the iMCs. This is an unbalanced memory configuration.

A single 6-channel interleave is formed, and memory bandwidth measured at 75% of the full potential memory bandwidth. Even though a 6 DIMM population is unbalanced, the 4th and 5th Gen Intel Xeon Scalable processor design includes the technology to create a single 6-channel interleave which results in relatively good performance. Without this technology, multiple interleave sets would have been formed and performance would be degraded.

Configuration with 8 DIMMs – balanced

The recommended configuration with 8 DIMMs is 1:1:1:1:1:1:1:1 memory configuration as shown in the following figure.
Configuration with 8 DIMMs – balanced
Figure 7. 1:1:1:1:1:1:1:1 memory configuration, relative memory bandwidth: 100%

This memory configuration follows both guidelines 1 and 2. All channels were populated with the same memory capacity, and memory configurations were identical between all iMCs. This is a balanced memory configuration.

A single 8-channel interleave is formed, and memory bandwidth measured at 100% of the full potential memory bandwidth.

Configuration with 12 DIMMs – unbalanced

The recommended configuration with 12 DIMMs is 1:2:1:2:2:1:2:1 memory configuration as shown in the following figure.
Configuration with 12 DIMMs – unbalanced
Figure 8. 1:2:1:2:2:1:2:1 memory configuration, relative memory bandwidth: 50-85% with SPR and 50-75% with EMR.

This memory configuration follows guideline 2, but not guideline 1. This is an unbalanced memory configuration.

One 8-channel interleave and one 4-channel interleave set is formed, and memory bandwidth measured at 50% – 85% of the full potential memory bandwidth for 4th Gen (Sapphire Rapids) processors, and 50% – 75% of the full potential memory bandwidth for 5th Gen (Emerald Rapids) processors. The lower than potential bandwidth is the result of not utilising all eight memory channels and the Intel supported memory speeds when memory is configured with two DIMMs per channel.

Configuration with 16 DIMMs – balanced

This is a fully populated configuration 2:2:2:2:2:2:2:2 as shown in the following figure.
Configuration with 16 DIMMs – balanced
Figure 9. 2:2:2:2:2:2:2:2 memory configuration, relative memory bandwidth: 85% with SPR and 76% with EMR

This is a fully populated memory configuration. It follows both guideline 1 and 2. This is a balanced memory configuration.

Each channel is populated with two dual-rank (2R) DIMMs, so the total number of ranks per channel is 4R.
Memory bandwidth measured at 85% with SPR, and 76% with EMR of the full potential memory bandwidth due to the memory speed dropping with two DIMMs per channel as per Intel plan of record.

Summary of the performance results

The following table shows a summary of the relative memory bandwidth of all the memory configurations that were evaluated. It also shows the number of interleave sets formed for each and whether it is a balanced or unbalanced memory configuration.

Table 2. Summary of the performance results

Number of DIMMs populated| Configuration| Number of interleave sets| Relative performance| Balanced or Unbalanced
---|---|---|---|---
1| 0:0:0:0:1:0:0:0| 1| 13%| Unbalanced
2| 0:1:0:0:1:0:0:0| 1| 25%| Unbalanced
4| 0:1:0:1:1:0:1:0| 1| 50%| Balanced
6| 1:1:0:1:1:1:1:0| 1| 75%| Unbalanced
8| 1:1:1:1:1:1:1:1| 1| 100%| Balanced
12| 1:2:1:2:2:1:2:1| 2| 50% – 85% -SPR50% – 75% – EMR| Unbalanced
16| 2:2:2:2:2:2:2:2| 1| 85% – SPR76% – EMR| Balanced

When using the same memory DIMM, only a memory configuration with 8 DIMMs provides the full potential memory bandwidth. This is the best memory configuration for performance. A balanced memory configuration can also be achieved with four memory DIMMs, but this configuration does not populate all the memory channels which reduces its memory bandwidth and performance. Finally, a balanced memory configuration can also be achieved with 16 memory DIMMs, although this requires two DIMMs per channel which drops the memory speed as per Intel’s plan of record.

Maximising memory bandwidth

To maximize the memory bandwidth of a server, the following rules should be followed:

  1. Balance the memory across the processor sockets – all processor sockets on the same physical server should have the same configuration of memory DIMMs.
  2. Balance the memory across the memory controllers – all memory controllers on a processor socket should have the same configuration of memory DIMMs.
  3. Balance the memory across the populated memory channels – all populated memory channels should have the same total memory capacity and the same total number of ranks.

Peak memory performance is achieved with 8 DIMMs per socket. Given a memory capacity requirement per server, follow these steps to get an optimal memory bandwidth configuration for your requirement:

  1. Determine your needed memory capacity per socket.
  2. Divide this memory capacity by eight to determine the minimum memory capacity needed per DDR channel.
  3. Round this calculated memory capacity per channel up to the closest capacity available with 1 DIMM Per Channel (DPC)
  4. Populate your server with eight identical DIMM combination per channel derived from step 3.

Examples:
If 512GB of total memory capacity is needed per socket, you can populate each socket with 8x 64GB DIMMs.
If 768GB of total memory capacity is needed per socket, you can populate each socket with 8x 96GB DIMMs.

Summary

Overall server performance is affected by the memory subsystem which can provide both high memory bandwidth and low memory access latency when properly configured. Balancing memory across the memory controllers and the memory channels produces memory configurations which can efficiently interleave memory references among its DIMMs producing the highest possible memory bandwidth. An unbalanced memory configuration can reduce the total memory bandwidth to as low as 13% of a balanced memory configuration with 8 identical DIMMs installed per processor.

Implementing all three of the balanced memory guidelines described in this paper results in balanced memory configuration and produces the best possible memory bandwidth and overall performance.

Lenovo recommends installing balanced memory population with 4, 8, or 16 DIMMs per socket. Peak memory performance is achieved with 8 DIMMs per processor.

Authors

This paper was produced by the following team of specialists:
Charles Stephan is a Senior Engineer and Technical Lead for the System Performance Verification team in the Lenovo Performance Laboratory at the Lenovo Infrastructure Solutions Group (ISG) campus in Morrisville, NC. His team is responsible for analyzing the performance of storage adapters, network adapters, various flash technologies, and complete x86 platforms. Before transitioning to Lenovo, Charles spent 16 years at IBM as a Performance Engineer analyzing storage subsystem performance of RAID adapters, Fibre Channel HBAs, and storage servers for all x86 platforms. He also analyzed performance of x86 rack systems, blades, and compute nodes. Charles holds a Master of Science degree in Computer Information Systems from the Florida Institute of Technology.

Redwan Rahman is a Systems Performance Verification Engineer in the Lenovo Infrastructure Solutions Group Performance Laboratory in Morrisville, NC, USA. His current role includes CPU, Memory, and PCIe subsystem analysis and performance validation against functional specifications and vendor targets. Redwan holds a Bachelor of Science in Computer Engineering from University of Massachusetts Amherst.

Thanks to the following people for their contributions to this project:

  • David Watts, Lenovo Press

This paper was based on the paper Balanced Memory Configurations for 2-Socket Servers with 3rd-Gen Intel Xeon Scalable Processors. Thanks to the author of that paper:

  • Nathan Pham

Balanced Memory papers

This paper is one of a series of papers on Balanced Memory configurations:

Related product families

Product families related to this document are the following:

Notices **

**

Lenovo may not offer the products, services, or features discussed in this document in all countries. Consult your
local Lenovo representative for information on the products and services currently available in your area. Any
reference to a Lenovo product, program, or service is not intended to state or imply that only that Lenovo product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any Lenovo intellectual property right may be used instead. However, it is the user’s responsibility to evaluate and verify the operation of any other product, program, or service. Lenovo may have patents or pending patent applications covering subject matter described in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to:

Lenovo (United States), Inc.
8001 Development Drive
Morrisville, NC 27560
U.S.A.
Attention : Lenovo Director of Licensing

LENOVO PROVIDES THIS PUBLICATION ”AS IS” WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESS OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF NON-INFRINGEMENT,
MERCHANT ABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Some jurisdictions do not allow disclaimer of express or implied warranties in certain transactions, therefore, this statement may not apply to you.

This information could include technical inaccuracies or typographical errors. Changes are periodically made to the information herein; these changes will be incorporated in new editions of the publication. Lenovo may make improvements and/or changes in the product(s) and/or the program(s) described in this publication at any time without notice.

The products described in this document are not intended for use in implantation or other life support applications where malfunction may result in injury or death to persons. The information contained in this document does not affect or change Lenovo product specifications or warranties. Nothing in this document shall operate as an express or implied license or indemnity under the intellectual property rights of Lenovo or third parties. All information contained in this document was obtained in specific environments and is presented as an illustration. The result obtained in other operating environments may vary. Lenovo may use or distribute any of the information you supply in any way it believes appropriate without incurring any obligation to you.

Any references in this publication to non-Lenovo Web sites are provided for convenience only and do not in any
manner serve as an endorsement of those Web sites. The materials at those Web sites are not part of the materials
for this Lenovo product, and use of those Web sites is at your own risk. Any performance data contained herein was
determined in a controlled environment. Therefore, the result obtained in other operating environments may vary
significantly. Some measurements may have been made on development-level systems and there is no guarantee
that these measurements will be the same on generally available systems. Furthermore, some measurements may
have been estimated through extrapolation. Actual results may vary. Users of this document should verify the applicable data for their specific environment.

© Copyright Lenovo 2024. All rights reserved.

This document, LP2000, was created or updated on August 13, 2024.

Send us your comments in one of the following ways:

Use the online Contact us review form found at:
https://lenovopress.lenovo.com/LP2000

Send your comments in an e-mail to:
comments@lenovopress.com

This document is available online at https://lenovopress.lenovo.com/LP2000.

Trademarks

Lenovo and the Lenovo logo are trademarks or registered trademarks of Lenovo in the United States, other countries, or both. A current list of Lenovo trademarks is available on the Web at https://www.lenovo.com/us/en/legal/copytrade/.

The following terms are trademarks of Lenovo in the United States, other countries, or both:
Lenovo®
System x®
ThinkServer®
ThinkSystem®

The following terms are trademarks of other companies:

AMD and AMD EPYC™ are trademarks of Advanced Micro Devices, Inc.

Intel® and Xeon® are trademarks of Intel Corporation or its subsidiaries.

Other company, product, or service names may be trademarks or service marks of others.
Lenovo logo

References

Read User Manual Online (PDF format)

Read User Manual Online (PDF format)  >>

Download This Manual (PDF format)

Download this manual  >>

Lenovo User Manuals

Related Manuals