Create efficient two-node edge infrastructure with Red Hat OpenShift and Portworx/Pure Storage

Share

Subscribe to RSS

The demand to extend applications to the edge has never been greater. From retail shops to industrial and manufacturing sites, there's a need to create, consume, and store data at the edge. Deploying applications at the edge comes with a set of physical constraints, but also with the need to deliver a truly cost-efficient and resilient architecture. When building applications at the edge, you must consider the needs of the individual site as well as the cost to deploy, manage, and maintain applications across multiple edge locations.

The good news is that Red Hat OpenShift is evolving to meet this demand head-on. With the introduction of the two-node OpenShift with arbiter topology, Red Hat and partners like Portworx by Pure Storage are delivering a cost-efficient and resilient architecture designed specifically for the edge.

The edge dilemma: High availability vs. cost optimization

The primary motivation behind the two-node OpenShift with arbiter (TNA) initiative is simple: Cost for large-scale edge deployments. For mission-critical edge sites, high availability is non-negotiable. Should a node fail, applications must continue running without disruption, which is why the control plane requires a quorum (typically three or more nodes) to prevent split-brain scenarios and maintain consistency (a principle known as the CAP theorem).

Two-node OpenShift with arbiter, now generally available (GA) with OpenShift 4.20, solves this by dramatically reducing the hardware footprint while preserving three-node consistency and availability.

Two-node OpenShift with arbiter explained

Two-node OpenShift with arbiter architecture is a specialized, cost-sensitive solution that is technically a three-node cluster. Here's how it works:

  • The main nodes: Two full-sized nodes run the OpenShift control plane, all worker workloads, and the bulk of the infrastructure services.
  • The arbiter node: A small, third node dedicated solely to running the third etcd instance necessary to maintain quorum for the control plane. By removing all other responsibilities, the arbiter node can be tiny.
[画像:Diagram shows two main nodes with the workload only on those, and a small third arbiter node where only etcd is running]

This approach achieves single node outage tolerance without impacting your workload, while significantly reducing the bill of materials compared to three full nodes.

The arbiter node requires only minimal system resources (2 vCPU, 8 GB RAM, and a 50 GB SSD, or similar, disk). Note that it's generally recommended to place the arbiter within close proximity to the other nodes, ideally the same location or site, but in a different failure domain. However, it is possible to have the arbiter at a greater distance (for example, in the cloud or in the datacenter).

The term "end-to-end", in this case, means that it includes network roundtrip and disk IO (and really the maximum, not average latency). Note that this requires the etcd slow profile to be configured, and that the cluster will behave slow, because every etcd write transaction must be synced and committed with the arbiter. If additional components are added to the arbiter (for a software defined storage solution, for example), then the requirements for those components must be added. See below for an example.

The architecture supports both x86 and Arm architectures on bare metal hardware certified for Red Hat Enterprise Linux 9 (RHEL) and OpenShift. This means platform=none or platform=baremetal, which allows installation of the arbiter as a virtual machine on any hypervisor supported on RHEL (including kvm, Hyper-V, VMware).

To install a cluster, the usual OpenShift installation methods are supported: UPI, IPI, Agent Based, and Assisted Service.

What's the high-availability stance of TNA?

The high-availability stance of TNA is basically the same as a regular three-node compact cluster, but more sensitive in terms of capacity planning. If no performance degradation is allowed during loss of a node, then each node can be utilised at 50% maximum to have enough capacity to handle the load of the failed node. The load that fails over is 50%, not just 33% like in a three-node compact cluster. Capacity must be planned accordingly.

Is the arbiter node a regular node?

Yes, the arbiter node is a regular node. It's Red Hat CoreOS, Kubelet, Networking, and so on, but the node has a special role ("arbiter"). This prevents regular workload and components from being scheduled to it:

$ oc get nodes
NAME STATUS ROLES
arbiter-0 Ready arbiter           
master-0 Ready control-plane,master,worker 
master-1 Ready control-plane,master,worker

This also means that pods can be explicitly scheduled to it when necessary, by adding a corresponding toleration to the deployment:

tolerations:
 - effect: NoSchedule
  key: node-role.kubernetes.io/master
 - effect: NoSchedule
  key: node-role.kubernetes.io/arbiter

This is used by infrastructure components that also require three nodes to establish quorum to provide consistency. A good example of this is a software-defined hyperconverged storage solution.

Unified data services at the edge

For many edge workloads, especially those leveraging virtualized infrastructure, true high availability requires a robust software-defined storage (SDS) layer that can persist and protect data regardless of which node is running the application. This is where Portworx by Pure Storage comes in, providing essential services like synchronous disaster recovery (DR) and automated data management.

The integration of OpenShift with OpenShift Virtualization and Portworx delivers a powerful, unified platform for virtual machines (VMs) and containers from a single control plane. This hyperconverged Kubernetes-native storage platform brings high availability, disaster recovery, and performance to edge environments, all without dedicated SAN or external storage arrays.

Key benefits of using two-node OpenShift with arbiter and Portworx include:

  • Dual-quorum maintenance: The small-footprint appliance that serves as the arbiter maintains quorum for both the OpenShift and Portworx control planes. This helps ensure platform and storage resilience with minimal hardware. Portworx storage requires only 4 vCPU (x86) and 4GB of RAM of additional resources on each of the nodes, including the arbiter.
  • Data resiliency: Portworx uses a replication factor of 2 across the two main worker nodes. Portworx provides persistent storage for containerized apps and virtual machines managed by OpenShift Virtualization, ensuring data integrity for both types of workloads.
  • Data services: Portworx provides unified VM and container data management benefits to the edge, helping organizations:
    • Architect data resiliency with zero recovery point objective (RPO) disaster recovery
    • Automate development and operations with self-service storage and automated capacity management
    • Protect application data with container-granular and VM file-level backups, and ransomware protection

By handling data for VMs, disks, and containers simultaneously, Portworx helps organizations modernize at the edge while preserving investments in existing virtualized applications. Portworx helps you maintain the Day 0, Day 1, and Day 2 storage and data management capabilities you expect (synchronous DR, automated capacity management, seamless backup and restore).

Portworx 3.5.0 supports two-node OpenShift with arbiter as part of OpenShift 4.20, making this resilient, cost-optimized edge solution available for production deployments.

Cost efficiency and resilience for the open hybrid cloud

The combination of two-node OpenShift with arbiter and Portworx by Pure Storage directly addresses the edge dilemma by reducing node count without sacrificing data integrity or workload availability. It provides a powerful foundation for running stateful, mission-critical applications at scale, ensuring consistent storage management, and achieving near-zero RPO (recovery point objective) for business continuity. The future of edge infrastructure is lean, resilient, and ready for modern containerized and virtualized workloads.

To learn more about how Red Hat OpenShift and Portworx can scale and protect your stateful applications at the edge, explore these OpenShift edge resources. You can contact the Red Hat OpenShift team to inquire about the upcoming availability of two-node OpenShift with arbiter.

Learn more by checking out this demo:

Product trial

Red Hat OpenShift Container Platform | Product Trial

A consistent hybrid cloud foundation for building and scaling containerized applications.

About the authors

Browse by channel

automation icon

Automation

The latest on IT automation for tech, teams, and environments

AI icon

Artificial intelligence

Updates on the platforms that free customers to run AI workloads anywhere

open hybrid cloud icon

Open hybrid cloud

Explore how we build a more flexible future with hybrid cloud

security icon

Security

The latest on how we reduce risks across environments and technologies

edge icon

Edge computing

Updates on the platforms that simplify operations at the edge

Infrastructure icon

Infrastructure

The latest on the world’s leading enterprise Linux platform

application development icon

Applications

Inside our solutions to the toughest application challenges

Virtualization icon

Virtualization

The future of enterprise virtualization for your workloads on-premise or across clouds