DELL Technologies PowerFlex Appliance Software User Guide
- August 15, 2024
- DELL Technologies
Table of Contents
- Notes, cautions, and warnings
- Introduction
- Revision history
- System overview
- System features
- System components
- Storage schemas
- System architecture
- Deployment options overview
- Base configurations and scaling in a hype converged deployment
- Base configurations and scaling in a two-layer deployment
- Additional references
- Network components
- Storage components
- Virtualization components
- Read User Manual Online (PDF format)
- Download This Manual (PDF format)
DELL Technologies PowerFlex Appliance Software
DELL Technologies PowerFlex Appliance Software
Notes, cautions, and warnings
NOTE: A NOTE indicates important information that helps you make better
use of your product.
CAUTION: A CAUTION indicates either potential damage to hardware or loss
of data and tells you how to avoid the problem.
WARNING : A WARNING indicates a potential for property damage, personal
injury, or death.
Introduction
This document describes the high-level design of the Power Flex appliance. It also describes the hardware and software components in each Power Flex appliance.
The target audience for this document includes customers, sales engineers, field consultants, and advanced services specialists who want to deploy a virtualized infrastructure using Power Flex appliance.
Power Flex appliance architecture is based on Dell PowerEdge R650, R750, R6525, R7525, R640, R740xd, and R840 servers.
Power Flex Manager provides the management and orchestration functionality for a Power Flex appliance.
The dv switch names are for example only and may not match the configured system. Do not change these names or a data unavailable or data lost event may occur.
Dell Power Flex appliance was previously known as Dell Flex appliance. Similarly, Dell Power Flex Manager was previously known as Dell Flex Manager, and Dell Power Flex was previously known as Dell Flex OS. References in the documentation will be updated over time.
For additional Power Flex appliance documentation, go to PowerFlex appliance technical documentation.
Revision history
Date| Document rev i s ion| De s cription of
change s
---|---|---
January 2024| 5.5| Updated D e p loymen t options overview Added
information about S to ra g e schem a s
November 2023| 5.4| Added support for Cisco Nexus 93180YC-FX3
August 2023| 5.3| Added support for Multi -subnet and multi-VLAN
configurations
March 2023| 5.2| Added support for PowerFlexR7525 nodes
December 2021| 5.1| Updated architecture diagrams
November 2021| 5.0| Added content for Power Flex management controller 2.0
June 2021| 4.0| Added content for 3-node Power Flex management controller
November 2020| 3.0| Updated
- System features
- System architecture
June 2020| 2.0| Add content for
- Native asynchronous replication
- Power Flex 3.5
- LACP bonded logical net work
March 2020| 1.1| Removed all software version numbers from the publication
March 2020| 1.0| Initial release
System overview
The Power Flex appliance is a fully integrated, preconfigured, and validated hype converged infrastructure appliance that integrates Power Flex virtualization software with Dell PowerEdge servers.
Power Flex appliance is used in a hype converged or server SAN architecture, heterogeneous virtualized environments, and high performance databases.
Power Flex appliance provides large storage capacity and scalability, enabling you to start small and grow in discrete increments. The flexible architecture enables you to mix and match configuration types within the same cluster. It allows you to grow and manage your compute and storage resources together or independently depending on your business needs. It provides enterprise-grade data protection, multitenant capabilities, and add-on enterprise features such as quality of service (QoS), thin provisioning, and snapshots.
Power Flex appliance uses Power Flex Manager for management and operations. Power Flex Manager allows you to build, automate and simplify implementation, expansion and life cycle management.
System features
-
Power Flex appliance uses Power Flex, which uses the local disks of existing Power Flex nodes to create a storage pool. It is designed to scale to thousands of Power Flex nodes.
The integrated scalable network architecture is available with a choice of the following: -
Hardware consisting of Cisco Nexus switches, Dell Powerwinch switches, or customer switches
-
Bandwidth depending on the customer needs
-
Network configurations
- Trunk
- Channel-group
- Channel-group with LACP
-
Complete flexibility in designing the converged system with the following options:
- Power Flex hype converged nodes
- Power Flex storage-only nodes
- Power Flex compute-only nodes
-
Supports 10 Gbe, 25 Gbe or 100 Gbe ports for backend connectivity
-
Supports native asynchronous replication between sites
-
Power Flex nodes supports SSD and NVMe drive technologies
-
Improved node maintenance using protected maintenance mode (PMM)
-
Supports self-encrypting drives (SEDs) enabled by Dell Cloud Link
-
Supports horizontal scaling by adding Power Flex nodes to extend storage or compute capacity
-
Secure and scalable storage using Power Flex
- Seamlessly increase storage capacity or performance with horizontal node scaling
- Optional data compression feature that improves storage efficiency
- Optional Data at Rest Encryption (D@RE) feature provides data security using Cloud Link as a software encryption layer, or as a key manager for self-encrypting drives
-
End-to-end lifecycle management with Power Flex Manager
-
For the Power Flex management controller 2.0, an optional three-node Power Flex management controller supporting VMware Essie and Power Flex is available for shared storage layer, providing HA for the VM workloads on the Power Flex management controller. The Power Flex management nodes are based on Power Flex R650 nodes and H755 RAID controller.
-
Supports Fibre Channel HBAs on Power Flex nodes to connect to external storage arrays (outside the Power Flex data path) for migrating the data from external storage arrays to Power Flex
-
Power Flex supports multi-subnet or multi-VLAN for all network types other than data, vison, and NSX overlay
-
Power Flex supports multiple Cisco and Dell switch models with different firmware and operating system versions
System components
Power Flex appliance contains compute, network, storage, data encryption, virtualization, and management resources.
The following table shows the supported and optional components of Power Flex appliance:
Resou rc e | Components |
---|---|
Compute | PowerFlex appliance R650, R750, R6525, R7525. R640, R740xd, and R840 |
Network | Customer supplied switches, or the following switches are also |
supported:
- Dell Power Switch S4148F-ON switch
- Dell Power Switch S5224F-ON switch
- Dell Power Switch S5048-ON switch
- Dell Power Switch S5248F-ON switch
- Dell Power Switch S5296F-ON switch
- Cisco Nexus 93180TC-EX switch
- Cisco Nexus 93180YC-FX switch
- Cisco Nexus 93180YC-FX3
- Cisco Nexus 93240YC-FX2 switch
Storage| Power Flex
Virtualization| VMware vSphere Essie
Management| Power Flex Manager, VMware center Server Appliance (vasa)
Power Flex management controller (optional)| Power Flex management controller
2.0 based on Power Flex
Data encryption (optional)| Dell Cloud link or self-encrypting drives (SEDs)
Storage schemas
Protection domains
A protection domain (PD) is a group of nodes or storage data servers that provide data isolation, security, and performance benefits. A node participates in only one protection domain at a time. Only nodes in the same protection domain can affect each other, nodes outside the protection domain are isolated. Secure multi-tenancy can be created with protection domains since data does not mingle across protection domains. You can create different protection domains for different node types with unequal performance profiles. All the hosts in the domain must have the same type and configuration. A Power Flex hype converged node should not be in the same protection domain as a Power Flex storage-only node. The node configuration must match, which includes the drives, CPU, and memory. Any difference in the node configuration leads to an unknown performance impact
Storage pools
Storage pools are a subset of physical storage devices in a protection domain. Each storage device belongs to one (and only one) storage pool. The best practice is to have the same type of storage devices (HDD versus SSD or SSD versus NVMe) within a storage pool to ensure that the volumes are distributed over the same type of storage within the protection domain. Power Flex supports two types of storage pools. You can choose between both layouts. A system can support both fine granularity (FG) and medium granularity (MG) pools on the same storage data server nodes. Volumes can be non-disruptively migrated between the two layouts. Within an fine granularity pool, you can enable or disable compression on a per-volume basis:
- Medium granularity: Volumes are divided into 1MB allocation units, distributed, and replicated across all disks contributing to a pool. MG storage pools support either thick or thin-provisioned volumes, and no attempt is made to reduce the size of user-data written to disk (except with all-zero data). MG storage pools have higher storage access performance than fine granularity storage pools but use more disk space.
- Fine granularity: A space efficient layout, with an allocation unit of just 4 KB and a physical data placement scheme based on log structure array (LSA) architecture. Fine granularity layout requires both flash media (SSD or NVMe) as well as NVDIMM to create an fine granularity storage pool. fine granularity layout is thin-provisioned and zero-padded by nature, and enables Power Flex to support in-line compression, more efficient snapshots, and persistent checksums. FG storage pools use less disk space than MG storage pools but have slightly lower storage access performance.
Fault sets
A fault set is a logical entity that contains a group of storage data servers
within a protection domain that have a higher chance of going down together;
for example, if they are all powered in the same rack. By grouping them into a
fault set, Power Flex mirrors data for a fault set on storage data servers
that are outside the fault set. Thus, availability is assured even if all the
servers within one fault set fail simultaneously.
System architecture
There are several options for a Power Flex appliance deployment.
Power Flex appliance is deployed as one of the following:
- Hype converged
- Two-layer
- Hybrid-hype converged
- Storage-only
The following figure is a high-level view of the physical Power Flex
appliance. This figure is not specific to a particular system but is a generic
representation. The Power Flex management environment depicted below can be
deployed on an optional Power Flex appliance management cluster, or a customer
provided VMware Essie host.
The following figure displays a Power Flex appliance replication environment.
The Power Flex management environment depicted below can be deployed on an
optional Power Flex appliance management cluster, or a customer provided
VMware Essie host.
Power Flex metadata managers (MDM) and storage data replication (SDR) need to
communicate to each other across the WAN.
Components in a dep l oyment | Description |
---|---|
Power Flex appliance management environment | Power Flex Manager requires a |
VMware Essie server hosting applications. The Power Flex management
environment can be deployed on an optional 3-node Power Flex management
cluster , or a customer provided VMware ESX host.
Customer provided resources| Embedded operating system jump server and VMware
center server. It is anticipated that these resources will already be
installed and accessible.
Access A and 8 switches| These switches are partially configured by Power Flex
Manager or they are fully customer configured and connect directly to the
Power Flex appliance nodes.
Hardware management switch| Provides separate network connectivity y of iDRAC
and access switch management interfaces.
Power Flex appliance customer environment|
- Power Flex hype converged nodes
- Power Flex compute-only nodes
- Power Flex storage-only nodes
VLANs| VLANs that are required to be defined on the various switches, interfaces. and network inter face cards (NICs).
Deployment options overview
There are several options for a Power Flex appliance deployment.
Power Flex appliance nodes use Power Flex to operate storage and tie in workloads. Power Flex appliance uses these Power Flex features:
- Storage data client (SDC), which consumes storage from the Power Flex appliance.
- Storage data server (SDS), which contributes node storage to Power Flex appliance.
- Power Flex metadata manager (MDM), which manages the storage blocks and tracks data location across the system.
- Storage data replication (SDR), which enables native asynchronous replication on Power Flex nodes.
Power Flex enables flexible deployment options by allowing the separation of SDC and SDS components. It addresses data center workload requirements through the following Power Flex appliance deployment options:
Depl o yment type | Des c ripti o n |
---|---|
Hype converged | Metadata manager, compute, and storage reside within the same |
server . SOR is supported on PowerFlex hype converged nodes.
Two-layer| Separate es compute resources from storage resources. allowing the
independent expansion of compute or storage resources.
Consists of PowerFlex compute-only nodes (supporting the SOC) and PowerFlex storage- only nodes ( connected to and managed by the SOS).
PowerFlex compute-only nodes host end-user applications. PowerFlex storage- only nodes contribute storage to the system pool.
PowerFlex metadata manager (MOM) runs on PowerFlex storage -only nodes.
Hybrid hype converged| Consists of e PowerFlex hype converged nodes. PowerFlex compute e-only nodes. and PowerFlex storage-only nodes. Some nodes contribute both compute resources and storage resources (PowerFlex hype converged nodes). some contribute only compute resources (PowerFlex compute-only nodes). and some contribute only storage resources (PowerFlex storage-only nodes).
PowerFlex metadata manager (MOM) runs either on PowerFlex hype converged nodes or PowerFlex storage-only nodes.
PowerFlex compute-only nodes| Consists of PowerFlex compute-only nodes (supporting the SOC) on Microsoft Windows, VMware ESXI. CentOS , and Red Hat Enterprise Linux.
PowerFlex compute-only nodes hosts end-user applications.
NOT E : Red Hat Enterprise Linux. CentOS. and Windows compute- only nodes must be
deployed by the customer instead of using PowerFlex Manager.
PowerFlex storage -only nodes| Consists of embedded operating system nodes which contribute storage resources to the virtual environment. The back-end traffic shares the same PowerFlex data networks. No SOC components are installed on this node.
PowerFlex metadata manager (MOM) runs on PowerFlex storage-only nodes. SOR is supported on PowerFlex storage-only nodes with a dual CPU.
PowerFlex storage-only nodes two-layer deployments use the same four PowerFlex data networks for both SDC and SDS communications. Two-layer deployment allows the rebooting of PowerFlex compute-only nodes without PowerFlex ramifications.
When designing initial deployment or specifying later growth, use PowerFlex hype converged nodes. You can add PowerFlex compute-only nodes, or PowerFlex storage-only nodes as needed.
To control the number of processors or cores, consider separating the compute for the application from the nodes that support storage. This deployment is a pure two-layer deployment. Extra workloads are supported or added using:
- PowerFlex hype converged nodes
- PowerFlex compute-only nodes
- PowerFlex storage-only nodes
When hype converged nodes are mixed with PowerFlex compute-only nodes or PowerFlex storage-only nodes, it creates a hybrid deployment.
Base configurations and scaling in a hype converged deployment
PowerFlex appliance has a base configuration that is a minimum set of hype converged components and fixed network resources. Within the base configuration, the following hardware aspects are customizable:
Hardware | Dell Technologies recommends… |
---|---|
Compute and storage | A minimum of four PowerFlex appliance nodes. |
Network | An optional management switch and a minimum of two access switches. |
In a leaf-spine configuration, one pair of leaf switches and one pair of border-leaf switches are recommended.
Management| An optional three -node PowerFlex management controller running VMware Essie to host PowerFlex Manager and other system Software’s. A PowerFlex management controller can also be used for this purpose.
NOTE: The single node controller is configured with the internal drives in a RAID5 configuration in order to provide enhanced local storage resiliency as there is only one node. The three node management environment does not include local RAID storage since there is additional redundancy provided by the fact that there are multiple hosts in the cluster. The single node management cluster can also be expanded into a three-node cluster.
Cabling|
- 25 GB SFP28 direct attached copper – four for each PowerFlex hype converged node.
- 100 GB QSFP28 direct attached copper – two for access switch uplinks and two for access switch VLT/VPC interconnects.
- 1 GB CAT5 or CAT6 – one for each node for iDRAC connectivity and one for each access switch for management connectivity.
Base configurations and scaling in a two-layer deployment
PowerFlex appliance has a base configuration that is a minimum set of PowerFlex compute-only nodes and PowerFlex storage only nodes and fixed network resources. Within the base configuration, the following hardware aspects are customizable:
Hardware | Dell Technologies recommends… |
---|---|
Compute | A minimum of three PowerFlex compute-only nodes. |
Storage | Six Power Flex storage-only nodes (four minimum). |
Network | An optional management switch and a minimum of two access switches. |
Management | An optional three-node PowerFlex management controller running |
VMware Essie to host PowerFlex Manager and other system software. A single
PowerFlex management controller can also be used for this purpose.
| NOTE: The single node controller is configured with the internal drives
in a RAID5 configuration in order to provide enhanced local storage resiliency
as there is only one node. The three node management environment does not
include local RAID storage since there is additional redundancy provided by
the fact that there are multiple hosts in the cluster. The single node
management cluster can also be expanded into a three-node cluster.
Cabling|
- 25 GB SFP28 direct attached copper – four for each storage only and compute only node.
- 100 GB QSFP28 direct attached copper – two for access switch uplinks and two for access switch VLT/VPC interconnects.
- 1 GB CAT5 or CAT6 – one for each node for iDRAC connectivity and one for each access switch for management connectivity.
Additional references
This section provides references to related documentation for virtualization, compute, network, management and storage components
Network components
Network component information and links to documentation are provided.
Pr oduct | Link to documentation |
---|---|
Dell Power Switch S5200 series | [http](http://www.delltechnologies.com/asset |
/en-us/products/)s://www.delltechnologies.com/asset/en-us/products/
networking/technical support/dell_emc_networking-
s5200_on_spec_sheet.pdf
Dell Power Switch S4100 series| hnps://i.dell.com/sites/doccontent/shared-
content/data- sheets/ en/ Documents/ dell-emic-net working-S4100-series- spec-
sheet.pdf
Dell Power Switch S5000 series| https:// www.dell.com/ support/home Ian
-us/product-
support/product/force10-s5000/docs
Dell Power Switch S5296F-ON| https:/ / www.delltechnologies.com/asset/ en-us/
products/ net working/ technical-support/ dell_emc_net working-
s5200_on_spec_sheet,pdf
Cisco Nexus 93180YC-EX| https://www.cisco.com/ cl en/ us/
products/collateral/ switches/nexus-9000-serles-swit chess/ datasheet-
c78-742283.ht ml
Cisco Nexus 93180YC-FX| https:// www.cisco.com/ c/ en/ us/ support/ switches/
nexus-93180yc- fax – switch/
model.html
Cisco Nexus 93180 YC-F X3| https:/ / www.cisco.com/ c/ en/ us/ products/
collateral/ switches/nexus-9000 -series-switches/ datasheet-
c78-744052.html
Cisco Nexus 93240YC-FX2| https://www.cisco.com/ cl en/ us/
products/collateral/ switches/nexus-9000-serles-swit chess/ datasheet-
c78-742282.ht ml
Storage components
Storage component information and links to documentation are provided.
Product | Description | Link to documentation |
---|---|---|
PowerFlex | Converges storage and compute resources into a single-layer |
architect ure, aggregating capacity and performance, simplifying management,
and scaling to thousands of nodes.| dell.com/
support
NVDIMM-N
| Provides compression high-speed DRAM performance that is coupled with flash- backed persistent storage for PowerFlex storage-only nodes.
NVDIMM is used for compression only on PowerFlex storage-only nodes.
| https://www.dell.com/ support/manuals/us/ en/04 I power edge – t640/nvdimmnug _pub/ int reduction?
guide=guid-8884370c-5553-4089- b613-a3c570b56f0e&lang=en- us
Virtualization components
Virtualization component information and links to documentation are provided.
Product | Description | Link to documentation |
---|---|---|
VMware center Server App lance (v CSA) | vasa is a preconfigured Linux virtual |
machine, which is optimized for running VMware center Server and the
associated services on Linux.| https:/ /www.vmware.com/products/vcenter-
server.html
VMware vSphere Esti| Virtualized infrastructure for hype converged systems.
Virtualizes all application servers and provides VMware High Availability y
(HA) and Dynamic Resource Scheduling (DRS).|
https://www.vmware.com/products/vsphere.html
Read User Manual Online (PDF format)
Read User Manual Online (PDF format) >>