EC2 - Placement groups
Sometimes you want control over the EC2 Instance placement strategy.
That strategy can be defined using placement groups.
When you create a placement group, you specify one of the following strategies for the group.
Cluster - clusters instances into a low-latency group in a single Availability Zone
Spread - spreads instances across underlying hardware (max 7 instances per group per AZ)
Partition - spreads instances across many different partitions (which rely on different sets of racks) within an AZ. Scales to 100s of EC2 instances per group (Hadoop, Cassandra, Kafka)
Cluster - Placement Group
Pros:
Great network (10 Gbps bandwidth between instances with Enhanced Networking enabled - recommended)
Cons:
If the rack fails, all instances fail at the same time.
Use case:
Big Data job that needs to complete fast
Application that needs extremely low latency and high network throughput
Spread- Placement Group
Pros:
Can span across Availability Zones (AZ)
Reduced risk is simultaneous failure.
EC2 Instances are on different physical hardware.
Cons:
Limited to 7 instances per AZ per placement group.
Use case:
Application that needs to maximize high availability.
Critical Applications where each instance must be isolated from failure from each other.
Partition - Placement Group
Pros:
Can span across multiple AZs in the same region.
The instances in a partition do not share racks with the instances in the other partitions.
Up to 100s of EC2 instances
A partition failure can affect many EC2 but won’t affect other partitions.
EC2 instances get access to the partition information as metadata.
Cons:
Up to 7 partitions per AZ
Use case:
HDFS, HBase, Cassandra, Kafka.
Critical Applications where at least one instance or more need to be available at any time and in any case.
Last updated