RWX Volume Fast Failover (Experimental)

Release 1.7.0 adds a feature that minimizes the downtime for ReadWriteMany volumes when a node fails. When enabled Longhorn uses a lease-based mechanism to monitor the state of the NFS server pod that exports the volume Longhorn reacts quickly to move it to another node if it becomes unresponsive. See RWX Volumes for details on how the NFS server works.

To enable the feature, you set RWX Volume Fast Failover to “true”. Existing RWX volumes will need to be restarted to use the feature after the setting is changed. That is done by scaling the workload down to zero and then back up again. New volumes will pick up the setting at creation and be configured appropriately.

With the feature enabled, when a pod is created or re-created, Longhorn also creates an associated lease object in the longhorn-system namespace, with the same name as the volume. The NFS server pod keeps the lease renewed as proof of life. If the renewal stops happening, Longhorn will take steps to create a new NFS server pod on another node and to re-attach the workload, even before the old node is marked as Not Ready by Kubernetes.

Along with adding the monitoring and fast reaction, the feature also changes the NFS server configuration to use a shortened grace period for client re-connection.

If the setting is changed back to “false”, the lease check is disabled and pod relocation will use regular Kubernetes rules for node failure, even on existing volumes. When the server pod is next restarted, it will revert to the normal grace period configuration.

For more information, see https://github.com/longhorn/longhorn/issues/6205.

Note: In rare circumstances, it is possible for the failover to become deadlocked. This happens if the NFS server pod creation is blocked by a recovery action that is itself blocked by the failover-in-process state. If that is the case, and failover takes more than a minute or two, the workaround is to delete the associated lease object. That clears the state, and a new lease is created along with the replacement server pod. For example, if the stuck volume is named pvc-2ce4e82e-7ccc-46c0-90a8-a141501fbf93 and the feature is enabled, there will be a lease with the same name. To delete the associated lease object:

kubectl -n longhorn-system delete lease pvc-2ce4e82e-7ccc-46c0-90a8-a141501fbf93

See, for example, https://github.com/longhorn/longhorn/issues/9093.

Resource Consumption and System Performance Impact

The Longhorn team has investigated the impact of RWX volumes on resource consumption and system performance. The benchmarking studies, which were completed using 60 RWX volumes, show that enabling the RWX Volume Fast Failover feature results in the following:

  • More requests sent to the Kubernetes API server (kube-apiserver)
  • More remote procedure calls (RPCs) sent from kube-apiserver to etcd
  • Slight increase in CPU and memory usage

Environment:

  • Setup: 1 Control Node + 3 Worker Nodes (v1.27.15+rke2r1)
  • Workload: 60 Deployments with 60 RWX volumes with soft mount

Test Results:

MetricFast Failover DisabledFast Failover EnabledDifference
API request rate (kube-apiserver)37.5 req/s59 req/s+57.3%
RPC rate (kube-apiserver to etcd)37 ops/s57 ops/s+54.1%
Memory usageLower Peaks/MinimaHigher Peaks/MinimaIncreased usage with Fast Failover enabled
Longhorn Manager CPU/RAM usage405 MB / 0.1 CPU417 MB / 0.13 CPU+3% RAM / +30% CPU
Share Manager CPU/RAM usage2.2 GB / 0.235 CPU2.25 GB / 0.26 CPU+2.3% RAM / +10.6% CPU

For detailed screenshots and further context, please refer to the related issue discussion.


© 2019-2024 Longhorn Authors | Documentation Distributed under CC-BY-4.0


© 2024 The Linux Foundation. All rights reserved. The Linux Foundation has registered trademarks and uses trademarks. For a list of trademarks of The Linux Foundation, please see our Trademark Usage page.