Storage Network

Harvester uses Longhorn as its built-in storage system to provide block device volumes for VMs and Pods. If the user wishes to isolate Longhorn replication traffic from the Kubernetes cluster network (i.e. the management network) or other cluster-wide workloads. Users can allocate a dedicated storage network for Longhorn replication traffic to get better network bandwidth and performance.

For more information, please see Longhorn Storage Network

Storage Network - 图1note

Configuring Longhorn settings directly is not recommended, as this can lead to untested situations.

Prerequisites

There are some prerequisites before configuring the Harvester Storage Network setting.

  • Well-configured Cluster Network and VLAN Config.
    • Users have to ensure the Cluster Network is configured and VLAN Config will cover all nodes and ensure the network connectivity is working and expected in all nodes.
  • All VMs should be stopped.
    • We recommend checking the VM status with the following command and should get an empty result.
    • kubectl get -A vmi
  • All pods that are attached to Longhorn Volumes should be stopped.
    • Users could skip this step with the Harvester Storage Network setting. Harvester will stop Longhorn-related pods automatically.
  • All ongoing image uploads or downloads should be either completed or deleted.

Storage Network - 图2caution

If the Harvester cluster was upgraded from v1.0.3, please check if Whereabouts CNI is installed properly before you move on to the next step. We will always recommend following this guide to check. Issue 3168 describes that the Harvester cluster will not always install Whereabouts CNI properly.

  • Verify the ippools.whereabouts.cni.cncf.io CRD exists with the following command.
    • kubectl get crd ippools.whereabouts.cni.cncf.io
  • If the Harvester cluster doesn’t have ippools.whereabouts.cni.cncf.io, please add these two CRDs before configuring storage-network setting.
  1. kubectl apply -f https://raw.githubusercontent.com/harvester/harvester/v1.1.0/deploy/charts/harvester/dependency_charts/whereabouts/crds/whereabouts.cni.cncf.io_ippools.yaml
  2. kubectl apply -f https://raw.githubusercontent.com/harvester/harvester/v1.1.0/deploy/charts/harvester/dependency_charts/whereabouts/crds/whereabouts.cni.cncf.io_overlappingrangeipreservations.yaml

Configuration Example

  • VLAN ID
    • Please check with your network switch setting, and provide a dedicated VLAN ID for Storage Network.
  • Well-configured Cluster Network and VLAN Config
    • Please refer Networking page for more details and configure Cluster Network and VLAN Config but not Networks.
  • IP range for Storage Network
    • IP range should not conflict or overlap with Kubernetes cluster networks(10.42.0.0/16, 10.43.0.0/16, 10.52.0.0/16 and 10.53.0.0/16 are reserved).
    • IP range should be in IPv4 CIDR format and Longhorn pods use Storage Network as follows:
      • instance-manger-e and instance-manager-r pods: These require 2 IPs per node. During an upgrade, two versions of these pods will exist (old and new), and the old version will be deleted once the upgrade is successful.
      • backing-image-ds pods: These are employed to process on-the-fly uploads and downloads of backing image data sources. These pods will be removed once the image uploads or downloads are completed.
      • backing-image-manager pods: 1 IP per disk, similar to the instance manager pods. Two versions of these will coexist during an upgrade, and the old ones will be removed after the upgrade is completed.
      • The required number of IPs is calculated using a simple formula: Required Number of IPs = Number of Nodes * 4 + Number of Disks * 2 + Number of Images to Download/Upload
    • For example, if your cluster has five nodes, each node has two disks, and ten images will be uploaded simultaneously, the IP range should be greater than or equal to /26 (5 * 4 + 5 * 2 * 2 + 10 = 50).

We will take the following configuration as an example to explain the details of the Storage Network

  • VLAN ID for Storage Network: 100
  • Cluster Network: storage
  • IP range: 192.168.0.0/24

Configuration Process

Harvester will create Multus NetworkAttachmentDefinition from the configuration, stop pods related to Longhorn Volume, update Longhorn setting, and restart previous pods.

Before Applying Harvester Storage Network Setting

Here we have two cases.

  • Expect that VM VLAN traffic and Longhorn Storage Network use the same group of physical interfaces.
  • Expect that VM VLAN traffic and Longhorn Storage Network use different physical interfaces.

Longhorn will send replication traffic through the specific interfaces shown as the red line in the figure.

Same Physical Interfaces

Take eth2 and eth3 as an example for VM VLAN traffic and Longhorn Storage Network simultaneously.

Please refer Networking page to configure ClusterNetwork and VLAN Config with eth2 and eth3 and remember the ClusterNetwork name for the further step.

storagenetwork-same.png

Different Physical Interfaces

eth2 and eth3 are for VM VLAN Traffic. eth4 and eth5 are for Longhorn Storage Network.

Please refer Networking page to configure ClusterNetwork and VLAN Config with eth4 and eth5 for Storage Network and remember the ClusterNetwork name for the further step.

storagenetwork-diff.png

Harvester Storage Network Setting

Harvester Storage Network setting will need range, clusterNetwork, vlan field to construct Multus NetworkAttachmentDefinition for Storage Network usage. You could apply this setting via Web UI or CLI.

Web UI

Harvester Storage Network setting could be easily modified on the Settings > storage-network page.

storagenetwork-ui.png

CLI

Users could use this command to edit Harvester Storage Network setting.

  1. kubectl edit settings.harvesterhci.io storage-network

The value format is JSON string or empty string as shown in below.

  1. {
  2. "vlan": 100,
  3. "clusterNetwork": "storage",
  4. "range": "192.168.0.0/24"
  5. }

The full configuration will be like this example.

  1. apiVersion: harvesterhci.io/v1beta1
  2. kind: Setting
  3. metadata:
  4. name: storage-network
  5. value: '{"vlan":100,"clusterNetwork":"storage","range":"192.168.0.0/24"}'

Storage Network - 图6caution

Because of the design, Harvester will treat extra and insignificant characters in JSON string as a different configuration.

After Applying Harvester Storage Network Setting

After applying Harvester’s Storage Network setting, Harvester will stop all pods that are related to Longhorn volumes. Currently, Harvester has some pods listed below that will be stopped during setting.

  • Prometheus
  • Grafana
  • Alertmanager
  • VM Import Controller

Harvester will also create a new NetworkAttachmentDefinition and update the Longhorn Storage Network setting.

Once the Longhorn setting is updated, Longhorn will restart all instance-manager-r, instance-manager-e, and backing-image-manager pods to apply the new network configuration, and Harvester will restart the pods.

Storage Network - 图7note

Harvester will not start VM automatically. Users should check whether the configuration is completed or not in the next section and start VM manually on demand.

Verify Configuration is Completed

Step 1

Check if Harvester Storage Network setting’s status is True and the type is configured.

  1. kubectl get settings.harvesterhci.io storage-network -o yaml

Completed Setting Example:

  1. apiVersion: harvesterhci.io/v1beta1
  2. kind: Setting
  3. metadata:
  4. annotations:
  5. storage-network.settings.harvesterhci.io/hash: da39a3ee5e6b4b0d3255bfef95601890afd80709
  6. storage-network.settings.harvesterhci.io/net-attach-def: ""
  7. storage-network.settings.harvesterhci.io/old-net-attach-def: ""
  8. creationTimestamp: "2022-10-13T06:36:39Z"
  9. generation: 51
  10. name: storage-network
  11. resourceVersion: "154638"
  12. uid: 2233ad63-ee52-45f6-a79c-147e48fc88db
  13. status:
  14. conditions:
  15. - lastUpdateTime: "2022-10-13T13:05:17Z"
  16. reason: Completed
  17. status: "True"
  18. type: configured

Step 2

Verify the readiness of all Longhorn instance-manager-e, instance-manager-r, and backing-image-manager pods, and confirm that their networks are correctly configured.

Execute the following command to inspect a pod’s details:

  1. kubectl -n longhorn-system describe pod <pod-name>

If you encounter an event resembling the following one, the Storage Network might have run out of its available IPs:

  1. Events:
  2. Type Reason Age From Message
  3. ---- ------ ---- ---- -------
  4. ....
  5. Warning FailedCreatePodSandBox 2m58s kubelet Failed to create pod sandbox: rpc error: code = Unknown desc = failed to setup network for
  6. sandbox "04e9bc160c4f1da612e2bb52dadc86702817ac557e641a3b07b7c4a340c9fc48": plugin type="multus" name="multus-cni-network" failed (add): [longhorn-system/ba
  7. cking-image-ds-default-image-lxq7r/7d6995ee-60a6-4f67-b9ea-246a73a4df54:storagenetwork-sdfg8]: error adding container to network "storagenetwork-sdfg8": erro
  8. r at storage engine: Could not allocate IP in range: ip: 172.16.0.1 / - 172.16.0.6 / range: net.IPNet{IP:net.IP{0xac, 0x10, 0x0, 0x0}, Mask:net.IPMask{0xff,
  9. 0xff, 0xff, 0xf8}}
  10. ....

Please reconfigure the Storage Network with a sufficient IP range.

Storage Network - 图8note

If the Storage Network has run out of IPs, you might encounter the same error when you upload/download images. Please delete the related images and reconfigure the Storage Network with a sufficient IP range.

Step 3

Check the k8s.v1.cni.cncf.io/network-status annotations and ensure that an interface named lhnet1 exists, with an IP address within the designated IP range.

Users could use the following command to show all Longhorn Instance Manager to verify.

  1. kubectl get pods -n longhorn-system -l longhorn.io/component=instance-manager -o yaml

Correct Network Example:

  1. apiVersion: v1
  2. kind: Pod
  3. metadata:
  4. annotations:
  5. cni.projectcalico.org/containerID: 2518b0696f6635896645b5546417447843e14208525d3c19d7ec6d7296cc13cd
  6. cni.projectcalico.org/podIP: 10.52.2.122/32
  7. cni.projectcalico.org/podIPs: 10.52.2.122/32
  8. k8s.v1.cni.cncf.io/network-status: |-
  9. [{
  10. "name": "k8s-pod-network",
  11. "ips": [
  12. "10.52.2.122"
  13. ],
  14. "default": true,
  15. "dns": {}
  16. },{
  17. "name": "harvester-system/storagenetwork-95bj4",
  18. "interface": "lhnet1",
  19. "ips": [
  20. "192.168.0.3"
  21. ],
  22. "mac": "2e:51:e6:31:96:40",
  23. "dns": {}
  24. }]
  25. k8s.v1.cni.cncf.io/networks: '[{"namespace": "harvester-system", "name": "storagenetwork-95bj4",
  26. "interface": "lhnet1"}]'
  27. k8s.v1.cni.cncf.io/networks-status: |-
  28. [{
  29. "name": "k8s-pod-network",
  30. "ips": [
  31. "10.52.2.122"
  32. ],
  33. "default": true,
  34. "dns": {}
  35. },{
  36. "name": "harvester-system/storagenetwork-95bj4",
  37. "interface": "lhnet1",
  38. "ips": [
  39. "192.168.0.3"
  40. ],
  41. "mac": "2e:51:e6:31:96:40",
  42. "dns": {}
  43. }]
  44. kubernetes.io/psp: global-unrestricted-psp
  45. longhorn.io/last-applied-tolerations: '[{"key":"kubevirt.io/drain","operator":"Exists","effect":"NoSchedule"}]'
  46. Omitted...

Start VM Manually

After verifying the configuration, users could start VM manually on demand.