The (Almost) Definitive Guide to PowerFlex Sizing & Other Matters Flashcards
What are the restrictions for software only PowerFlex?
only for existing ScaleIO customers (since most original ScaleIO deployments were SW only) or net new customers purchasing over 2 million in capacity licensing over 2 years
RPQ required
How much RAM is required per SDC?
50MB
also consumes very low host CPU
What is the SDSs responsibilities?
consuming local storage devices within server
making sure second copy of data is written to a different host before acknowledgement back to the SDCs
How much RAM does an SDS consume?
500MB-12GB per host (with MG layout)
very little CPU workload
How many vCPUs/cores are allocated for an SDS?
8-12
What protection level do protection domains have?
only one simultaneous node failure can occur within a PD
fault sets allow for multiple host failures but are typically not required
What should be the max number of SDSs to put in a PD?
30 SDSs per protection domain
to maintain 6 9s of availability
What is the max number of devices in a storage pool?
300 devices
How is data layed out in FG and MG pools?
MG - 1MB
FG - 4KB
What are the two performance profiles on PowerFlex?
High
Compact
What is the high performance profile?
all flash nodes
What is the compact performance profile?
HDD nodes
What is the SVM?
Storage VM
a VM for ESXi environments that runs CentOS - runs the SDS and LIA components
typically maps the local storage devices and RAID controller to the host through DirectPath IO
NVMe drives should use RDM
What is the main different between high and compact performance profile?
amount of CPU resources given
compact - SVM given 2 vCPUs/cores
high - SVM given 8 vCPUs/cores
What is the performance profile recommendation for PowerFlex?
configure High for both SSD and HDD - the new default setting
version 3.5 allows now up to 12 vCPUs and 12 SDS threads for FG pools
What is the recommendation specific to performance profile and CloudLink?
when using CloudLink D@RE increasing SVM by another vCPUs is considered best practice (from 8 to 12)
What is the all-flash performance of PowerFlex?
each SSD can provide about 80,000 Read IOPS
What is the max performance of SDS?
250,000 IOPS for reads
100,000 IOPS for writes (due to second copy being required it’s slower)
What happens after you hit the IOPS limit of an SDS?
can add more SSDs for higher large block performance
will not be able to improve small block performance
What is the typical performance of an SDC?
500,000 read or write IOPS if connected to multiple SDSs
What is the bandwidth performance of PowerFlex?
for large block IO - can saturate 200GbE per node to deliver 20GB/s of bandwidth for SDCs per node (2 x 100GbE interfaces)
SDS normally max out around 10GB/s per node when using NVMe and 7-8GB/s when using SSD
What are the latency metrics of PowerFlex?
0.2-1ms w/ All Flash and very high IOPS
100-200 microseconds w/ NVMe and low queue depth IO profiles
How does PowerFlex Manager run?
runs an an ova. VM on dedicated management nodes external to PowerFlex workload nodes
can also be managed through REST API
What is a recommendation for sizing Oracle environments?
often beneficial to have compute-only nodes to extract the most value out of expensive licenses which are by the core
What is the only node type that can support GPUs?
R740xd
Why do you need to configure minimum 10% spare capacity?
to maintain protection of a single node failure
What is the rule when sizing for performance?
the more nodes the better
size with smaller drives to get more nodes
Why is bandwidth and latency better on NVMe devices rather than SSDs?
IO does not need to go through the HBA330 controller and instead goes directly to the CPU
How is PowerFlex licensed?
per raw TiB - perpetual
can carry it over to net new hardware to reduce TCO
licenses per customer not per site or per cluster
How many tiers of licensing are their?
11 Tiers
What is tier 1 licensing?
12-100TBs
What is tier 2 licensing?
101-250TBs
What is tier 3 licensing?
251-500TBs
What is tier 4 licensing?
501-1000TBs
What is tier 5 licensing?
1001-5000TBs
What is tier 6 licensing?
5001-10000TBs
What is tier 7 licensing?
10001-20000TBs
What is tier 8 licensing?
20001-40000TBs
What is tier 9 licensing?
40001-60000TBs
What is tier 10 licensing?
60001-100000TBs
What is tier 11 licensing?
100000+TBs
What is the RPO for asynchronous replication?
30 seconds for version 3.5
How much RAM is required for SDR?
22GB RAM per node
What is the connectivity recommendation for SDR?
25GbE or faster
Where is replication enabled on PowerFlex?
at the protection domain level
SDRs must be present in each PD if any volumes from that PD need to be replicated
What are the replication design choices for PowerFlex?
SDR only enabled on one PD
enable only Partial SDRs per PD
enable max nodes w/ SDR
What is the SDR only enabled on one PD option?
allow only for volumes on that PD to be replicated
if this PD is down for maintenance replication will be unavailable during that time
What is the enable only partial SDRs per PD?
will allow all PDs to be able to participate in replication activities
if the workload is low at least 3 nodes per PD need to be running the SDR
design may require an RPQ
What is the enable max nodes with SDRs?
current supported max SDR per cluster is 64 (version 3.5)
What is the relationship between WAN bandwidth and replication?
replication performance most often limited by the WAN bandwidth
need to properly size WAN bandwidth and journal capacities
What is the relationship between local bandwidth and replication?
bandwidth consumption w/ replication is 3x higher than without which is why additional bandwidth is required
8K workload EX:
normal write = 16K (2 copies of 8K)
replicated write = 48K
What is the networking recommendation for PowerFlex?
use multipathing whenever possible along with LACP
LACP is first layer of protection and then there are multiple VLANs running on top providing IP Multipathing
What is important for OLTP workloads?
require high IOPS and low latency
What is important for OLAP workloads?
high bandwidth
How is SDS installed in a preconfigured ESXi environment?
deploying SVM
How does the MDM interact with the SDS and SDC?
interacts asynchronously using a separate data path and will not impact performance
What do larger storage pools provide?
wider striping by spreading IOs across more devices - benefits apps w/ high IO concurrency and bandwidth requirements
more flexible shared capacity for multiple volumes by providing larger pool for volume allocation
faster rebuild time - more spindles typically means faster rebuilds
What do smaller storage pools provide?
better protection from multiple drive failures
better guaranteed IOPS by provisioning SPs for apps that require guaranteed performance
How is physical and logical data protection done on PowerFlex?
PDs provide physical separation and SPs provide logical separation
SDCs can access volumes regardless of PD or SP
What is important to know before implementing fault sets?
can only create and configure fault sets before adding SDSs to the system
configuring them incorrecty may prevent the creation of volumes since an SDS can only be added to a fault set during the creation of the SDS
What are the rules for fault sets?
data mirroring for all the devices in the fault set must take place in a separate fault set
minimum of 3 fault sets are defined
Where is spare capacity implemented in PowerFlex?
the SP level
What is the relationship between rebuild/rebalance operations and bandwidth?
rebuild/rebalance can affect the available bandwidth for client traffic if the physical ports are shared between SDC and SDS nodes
What is the networking recommendation when it comes to rebuild/rebalance operations?
Have multiple interfaces connected to allow front end client traffic to flow unimpeded by back end operations
back end rebuild/rebalance can compete w/ client bandwidth if only over a single interface
How can interfaces be designated in PowerFlex?
SDC to SDS communication
SDS to SDS traffic
What is the networking recommendation if you only have a single interface connection per node?
limit client bandwidth to keep from overloading system while a rebuild/rebalance occurs
How can MDMs be mounted in PowerFlex?
can be located anywhere in the system and don’t require dedicated nodes
in larger configurations they should be distributed for availability reasons
EX: in multirack config they each (3-5) should be placed in different racks and different fault units
What is ideal network configurations for MDM connectivity?
all on the same subnet for minimal latency
What is the networking recommendation for smaller configurations?
running full mesh architecture w/ SDSs and SDCs sharing a non-routed subnet (access-agg)
What is the networking recommendation for larger configurations?
leaf-spine
Why should you size for multiple storage pools initially?
allows customer growth without having to provision multiple storage pools and keeps them balanced with respect to number of drives in them