You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've read the docummentation and solution provided in 11.13 page, which is "Fail over the workload to another node" doesn't work. Pod is crashing at all nodes. I thought it's related to pvc created using 3.2.1 version, but newly created pvc has this problem too.
Also tried to delete and recreate entire workload, but same problem.
Configuration
Here's my configuration. vaules.yaml:
images:
# Images provided by HuaweihuaweiCSIService: myregistry/huawei-csi:4.2.1storageBackendSidecar: myregistry/storage-backend-sidecar:4.2.1storageBackendController: myregistry/storage-backend-controller:4.2.1# CSI-related sidecar images provided by the Kubernetes community.# These must match the appropriate Kubernetes version.sidecar:
attacher: k8s.gcr.io/sig-storage/csi-attacher:v3.4.0provisioner: k8s.gcr.io/sig-storage/csi-provisioner:v3.0.0resizer: k8s.gcr.io/sig-storage/csi-resizer:v1.4.0registrar: k8s.gcr.io/sig-storage/csi-node-driver-registrar:v2.3.0livenessProbe: k8s.gcr.io/sig-storage/livenessprobe:v2.5.0snapshotter: k8s.gcr.io/sig-storage/csi-snapshotter:v4.2.1snapshotController: k8s.gcr.io/sig-storage/snapshot-controller:v4.2.1# Default image pull policy for sidecar container images, support [IfNotPresent, Always, Never]sidecarImagePullPolicy: "IfNotPresent"# Default image pull policy for Huawei plugin container images, support [IfNotPresent, Always, Never]huaweiImagePullPolicy: "IfNotPresent"# Namespace for installing huawei-csi-nodes and huawei-csi-controllerskubernetes:
# the default value huawei-csi is recommended.namespace: huawei-csi# Specify kubelet config dir path.# kubernetes and openshift is usually /var/lib/kubelet# Tanzu is usually /var/vcap/data/kubelet# CCE is usually /mnt/paas/kubernetes/kubeletkubeletConfigDir: /var/lib/kubeletCSIDriverObject:
# isCreate: create CSIDriver Object# If the Kubernetes version is lower than 1.18, set this parameter to false.# Allowed values:# true: will create CSIDriver object during installation.# false: will not create CSIDriver object during installation.# Default value: falseisCreate: true# If the Kubernetes version is lower than 1.20, set this parameter to null.# fsGroupPolicy: Defines if the underlying volume supports changing ownership and permission of the volume before being mounted.# 'fsGroupPolicy' is only valid when 'isCreate' is true# Allowed values:# ReadWriteOnceWithFSType: supports volume ownership and permissions change only if the fsType is defined# and the volume's accessModes contains ReadWriteOnce.# File: kubernetes may use fsGroup to change permissions and ownership of the volume# to match user requested fsGroup in the pod's security policy regardless of fstype or access mode.# None: volumes will be mounted with no modifications.# Default value: nullfsGroupPolicy: ReadWriteOnceWithFSType# If the Kubernetes version is lower than 1.18, set this parameter to true.# attachRequired: Whether to skip any attach operation altogether.# When 'isCreate' is true and 'attachRequired' is false, csi-attacher sidecar will not be deployed# Allowed values:# true: attach will be called.# false: attach will be skipped.# Default value: trueattachRequired: falsecontroller:
# controllerCount: Define the number of huawei-csi controller# Allowed values: n, where n > 0# Default value: 1# Recommended value: 2controllerCount: 1# volumeNamePrefix: Define a prefix that is prepended to volumes.# THIS MUST BE ALL LOWER CASE.# Default value: pvc# Examples: "volumes", "vol"volumeNamePrefix: pvc# Port used by the webhook service. The default port is 4433.# You can change the port to another port that is not occupied.webhookPort: 4433snapshot:
# enabled: Enable/Disable volume snapshot feature# If the Kubernetes version is lower than 1.17, set this parameter to false.# Allowed values:# true: enable volume snapshot feature(install snapshotter sidecar)# false: disable volume snapshot feature(do not install snapshotter sidecar)# Default value: Noneenabled: trueresizer:
# enabled: Enable/Disable volume expansion feature# Allowed values:# true: enable volume expansion feature(install resizer sidecar)# false: disable volume snapshot feature(do not install resizer sidecar)# Default value: Noneenabled: true# nodeSelector: Define node selection constraints for controller pods.# For the pod to be eligible to run on a node, the node must have each# of the indicated key-value pairs as labels.# Leave as blank to consider all nodes# Allowed values: map of key-value pairs# Default value: NonenodeSelector:
# Uncomment if nodes you wish to use have the node-role.kubernetes.io/master taint# node-role.kubernetes.io/master: ""# Uncomment if nodes you wish to use have the node-role.kubernetes.io/control-plane taint# node-role.kubernetes.io/control-plane: ""# tolerations: Define tolerations that would be applied to controller deployment# Leave as blank to install controller on worker nodes# Allowed values: map of key-value pairs# Default value: Nonetolerations:
# Uncomment if nodes you wish to use have the node-role.kubernetes.io/master taint# - key: "node-role.kubernetes.io/master"# Uncomment if nodes you wish to use have the node-role.kubernetes.io/control-plane taint# - key: "node-role.kubernetes.io/control-plane"# operator: "Exists"# effect: "NoSchedule"node:
# maxVolumesPerNode: Defines the maximum number of volumes that can be used by a node.# Examples: 100# Uncomment if you want to limit the number of volumes that can be used in a Node.# maxVolumesPerNode: 100# nodeSelector: Define node selection constraints for node pods.# For the pod to be eligible to run on a node, the node must have each# of the indicated key-value pairs as labels.# Leave as blank to consider all nodes# Allowed values: map of key-value pairs# Default value: NonenodeSelector:
# Uncomment if nodes you wish to use have the node-role.kubernetes.io/master taint# node-role.kubernetes.io/master: ""# Uncomment if nodes you wish to use have the node-role.kubernetes.io/control-plane taint# node-role.kubernetes.io/control-plane: ""# tolerations: Define tolerations that would be applied to node daemonset# Add/Remove tolerations as per requirement# Leave as blank if you wish to not apply any tolerations# Allowed values: map of key-value pairs# Default value: Nonetolerations:
- key: "node.kubernetes.io/memory-pressure"operator: "Exists"effect: "NoExecute"
- key: "node.kubernetes.io/disk-pressure"operator: "Exists"effect: "NoExecute"
- key: "node.kubernetes.io/network-unavailable"operator: "Exists"effect: "NoExecute"# - key: "node-role.kubernetes.io/control-plane"# operator: "Exists"# effect: "NoSchedule"# - key: "node-role.kubernetes.io/master"# operator: "Exists"# effect: "NoSchedule"# The CSI driver parameter configurationcsiDriver:
# Driver name, it is strongly recommended not to modify this parameter# The CCE platform needs to modify this parameter, e.g. csi.oceanstor.comdriverName: csi.huawei.com# Endpoint, it is strongly recommended not to modify this parameterendpoint: /csi/csi.sock# DR Endpoint, it is strongly recommended not to modify this parameterdrEndpoint: /csi/dr-csi.sock# Maximum number of concurrent disk scans or detaches, support 1~10connectorThreads: 4# Flag to enable or disable volume multipath access, support [true, false]volumeUseMultipath: true# Multipath software used by fc/iscsi. support [DM-multipath, HW-UltraPath, HW-UltraPath-NVMe]scsiMultipathType: DM-multipath# Multipath software used by roce/fc-nvme. only support [HW-UltraPath-NVMe]nvmeMultipathType: HW-UltraPath-NVMe# Timeout interval for waiting for multipath aggregation when DM-multipath is used on the host. support 1~600scanVolumeTimeout: 3# Timeout interval for running command on the host. support 1~600execCommandTimeout: 30# check the number of paths for multipath aggregation# Allowed values:# true: the number of paths aggregated by DM-multipath is equal to the number of online paths# false: the number of paths aggregated by DM-multipath is not checked.# Default value: falseallPathOnline: false# Interval for updating backend capabilities. support 60~600backendUpdateInterval: 60# label enableenableLabel: false# Huawei-csi-controller log configurationcontrollerLogging:
# Log record type, support [file, console]module: file# Log Level, support [debug, info, warning, error, fatal]level: info# Directory for storing logsfileDir: /var/log/huawei# Size of a single log filefileSize: 20M# Maximum number of log files that can be backed up.maxBackups: 9# Huawei-csi-node log configurationnodeLogging:
# Log record type, support [file, console]module: file# Log Level, support [debug, info, warning, error, fatal]level: info# Directory for storing logsfileDir: /var/log/huawei# Size of a single log filefileSize: 20M# Maximum number of log files that can be backed up.maxBackups: 9# leaderElection configurationleaderElection:
leaseDuration: 8srenewDeadline: 6sretryPeriod: 2s
Related problem is in #137, but here is more detailed info.
The text was updated successfully, but these errors were encountered:
My stack
Kubernetes version - 1.25
Huawei CSI driver upgraded from 3.2.1 to 4.2.1
Device Model: 5500 V5
Version: V500R007C10
Problem
PVC provisioning works fine, resizing too, but when trying to mount volume to pod, pod is in pending state with a reason
Solution from docs doesn't work
I've read the docummentation and solution provided in 11.13 page, which is "Fail over the workload to another node" doesn't work. Pod is crashing at all nodes. I thought it's related to pvc created using 3.2.1 version, but newly created pvc has this problem too.
Also tried to delete and recreate entire workload, but same problem.
Configuration
Here's my configuration. vaules.yaml:
Related problem is in #137, but here is more detailed info.
The text was updated successfully, but these errors were encountered: