PowerFlex
Release notes for PowerFlex CSI driver
Release Notes - CSI PowerFlex v2.11.0
New Features/Changes
- #1359 - [FEATURE]: Add Support for OpenShift Container Platform (OCP) 4.16
- #1400 - [FEATURE]: Support for Kubernetes 1.30
- #1358 - [FEATURE]: Support for PowerFlex 4.6
- #1397 - [FEATURE]: Observability upgrade is supported in CSM Operator
Fixed Issues
- #1209 - [BUG]: Doc hyper links in driver Readme is broken
- #1218 - [BUG]: Add the helm-charts-version parameter to the install command for all drivers in csm-docs
- #1237 - [BUG]: Error handling not good in node.go:nodeProbe() and other similar functions
- #1239 - [BUG]: Changes in new release of google.golang.org/protobuf is causing compilation issues
- #1270 - [BUG]: Missing entries for Resiliency in installation wizard template
- #1310 - [BUG]: CSI node pod crash after replacing OCP ingress certificate or restarting kubectl service
- #1350 - [BUG]: Document update : PowerFlex expecting secret CR as
-config in operator - #1355 - [BUG]: Indentation of secret.yaml mentioned on the csm-doc portal for powerflex driver is incorrect.
- #1364 - [BUG]: mkfsFormatOption not working for powerflex
- #1366 - [BUG]: Support Minimum 3GB Volume Size for NFS in CSI-PowerFlex
Known Issues
Issue | Workaround |
---|---|
Delete namespace that has PVCs and pods created with the driver. The External health monitor sidecar crashes as a result of this operation. | Deleting the namespace deletes the PVCs first and then removes the pods in the namespace. This brings a condition where pods exist without their PVCs and causes the external-health-monitor sidecar to crash. This is a known issue and has been reported at https://github.com/kubernetes-csi/external-health-monitor/issues/100 |
When a node goes down, the block volumes attached to the node cannot be attached to another node | This is a known issue and has been reported at https://github.com/kubernetes-csi/external-attacher/issues/215. Workaround: 1. Force delete the pod running on the node that went down 2. Delete the volumeattachment to the node that went down. Now the volume can be attached to the new node. |
sdc:3.6.0.6 is causing issues while installing the csi-powerflex driver on ubuntu,RHEL8.3 | Workaround: Change the powerflexSdc to sdc:3.6 in values.yaml https://github.com/dell/csi-powerflex/blob/72b27acee7553006cc09df97f85405f58478d2e4/helm/csi-vxflexos/values.yaml#L13 |
sdc:3.6.1 is causing issues while installing the csi-powerflex driver on ubuntu. | Workaround: Change the powerflexSdc to sdc:3.6 in values.yaml https://github.com/dell/csi-powerflex/blob/72b27acee7553006cc09df97f85405f58478d2e4/helm/csi-vxflexos/values.yaml#L13 |
A CSI ephemeral pod may not get created in OpenShift 4.13 and fail with the error "error when creating pod: the pod uses an inline volume provided by CSIDriver csi-vxflexos.dellemc.com, and the namespace has a pod security enforcement level that is lower than privileged." |
This issue occurs because OpenShift 4.13 introduced the CSI Volume Admission plugin to restrict the use of a CSI driver capable of provisioning CSI ephemeral volumes during pod admission. Therefore, an additional label security.openshift.io/csi-ephemeral-volume-profile in csidriver.yaml file with the required security profile value should be provided. Follow OpenShift 4.13 documentation for CSI Ephemeral Volumes for more information. |
If the volume limit is exhausted and there are pending pods and PVCs due to exceed max volume count , the pending PVCs will be bound to PVs and the pending pods will be scheduled to nodes when the driver pods are restarted. |
It is advised not to have any pending pods or PVCs once the volume limit per node is exhausted on a CSI Driver. There is an open issue reported with kubenetes at https://github.com/kubernetes/kubernetes/issues/95911 with the same behavior. |
Resource quotas may not work properly with the CSI PowerFlex driver. PowerFlex is only able to assign storage in 8Gi chunks, so if a create volume call is made with a size not divisible by 8Gi, CSI-PowerFlex will round up to the next 8Gi boundary when it provisions storage – however, the resource quota will not record this size but rather the original size in the create request. This means that, for example, if a 10Gi resource quota is set, and a user provisions 10 1Gi PVCs, 80Gi of storage will actually be allocated, which is well over the amount specified in the resource quota. | For now, users should only provision volumes in 8Gi-divisible chunks if they want to use resource quotas. |
Note:
- Support for Kubernetes alpha features like Volume Health Monitoring will not be available in Openshift environment as Openshift doesn’t support enabling of alpha features for Production Grade clusters.
- For fixing #1210 - [BUG]: Helm deployment of PowerFlex driver is failing, a new helm-chart has been released. In order to install this helm chart, we need to pass the flag
--helm-charts-version
during helm installation and flag-v
during offline bundle installation with valuecsi-vxflexos-2.11.0
.
Feedback
Was this page helpful?
Glad to hear it! Please tell us how we can improve.
Sorry to hear that. Please tell us how we can improve.