- KEP-0661: StatefulSet volume resize
- Release Signoff Checklist
- Summary
- Motivation
- Proposal
- Implementation Details
- Production Readiness Review Questionnaire
- Feature Enablement and Rollback - How can this feature be enabled / disabled in a live cluster? - Does enabling the feature change any default behavior? - Can the feature be disabled once it has been enabled (i.e. can we roll back the enablement)? - What happens if we reenable the feature if it was previously rolled back? - Are there any tests for feature enablement/disablement?
- Rollout, Upgrade and Rollback Planning - How can a rollout or rollback fail? Can it impact already running workloads? - What specific metrics should inform a rollback? - Were upgrade and rollback tested? Was the upgrade->downgrade->upgrade path tested? - Is the rollout accompanied by any deprecations and/or removals of features, APIs, fields of API types, flags, etc.?
- Monitoring Requirements - How can an operator determine if the feature is in use by workloads? - How can someone using this feature know that it is working for their instance? - What are the reasonable SLOs (Service Level Objectives) for the enhancement? - What are the SLIs (Service Level Indicators) an operator can use to determine the health of the service? - Are there any missing metrics that would be useful to have to improve observability of this feature?
- Dependencies - Does this feature depend on any specific services running in the cluster?
- Scalability - Will enabling / using this feature result in any new API calls? - Will enabling / using this feature result in introducing new API types? - Will enabling / using this feature result in any new calls to the cloud provider? - Will enabling / using this feature result in increasing size or count of the existing API objects? - Will enabling / using this feature result in increasing time taken by any operations covered by existing SLIs/SLOs? - Will enabling / using this feature result in non-negligible increase of resource usage (CPU, RAM, disk, IO, ...) in any components?
- Troubleshooting - How does this feature react if the API server and/or etcd is unavailable? - What are other known failure modes? - What steps should be taken if SLOs are not being met to determine the problem?
- Implementation History
- Drawbacks
- Alternatives
- Infrastructure Needed (Optional)
Items marked with (R) are required prior to targeting to a milestone / release.
- (R) Enhancement issue in release milestone, which links to KEP dir in kubernetes/enhancements (not the initial KEP PR)
- (R) KEP approvers have approved the KEP status as
implementable
- (R) Design details are appropriately documented
- (R) Test plan is in place, giving consideration to SIG Architecture and SIG Testing input (including test refactors)
- e2e Tests for all Beta API Operations (endpoints)
- (R) Ensure GA e2e tests for meet requirements for Conformance Tests
- (R) Minimum Two Week Window for GA e2e tests to prove flake free
- (R) Graduation criteria is in place
- (R) all GA Endpoints must be hit by Conformance Tests
- (R) Production readiness review completed
- (R) Production readiness review approved
- "Implementation History" section is up-to-date for milestone
- User-facing documentation has been created in kubernetes/website, for publication to kubernetes.io
- Supporting documentation—e.g., additional design documents, links to mailing list discussions/SIG meetings, relevant PRs/issues, release notes
Kubernetes has supported volume expansion as a beta feature since v1.16
. However, this feature support is only
limited to expanding volumes by editing the PVCs as an adhoc operation. Expansion of the volumes associated with a
StatefulSet
created via the volumeClaimTemplate
is not supported directly via the modification of the volumeClaimTemplate
construct of the StatefulSet
. This enhancement proposes to add the ability to resize volumes associated
with a StatefulSet
via modifications to its volumeClaimTemplate
specification.
Stable persistent volumes can be created by specifying .spec.volumeClaimTemplates
for a StatefulSet.
Currently, StatefulSets do not allow the associated volumes to be resized directly by modifying the
.spec.volumeClaimTemplates
. Instead users have to modify each PVC one by one to achieve this. In such cases,
when the StatefulSet scales up, the new PVC(s) will be created with the older size and this again needs
manual intervention. The intent of this proposal is to avoid this operational overhead and simplify the
process of resizing StatefulSet volumes.
Allow users to resize the volumes associated with StatefulSet
by modifying the size of
PVC in the .spec.volumeClaimTemplates
.
- Minimize the amount of unexpected errors/failures that the user might experience during volume resize by having as much validation as we can on the API server.
- In case errors do happen during resize, we shouldn't get in the way of the user trying to remediate those errors
- Provide a feedback mechanism for when a volume finished resizing successfully, so that users can build detection mechanisms around that for their automated CI/CD processes.
This proposal does not try to address the following scenarios but relies on the underlying PVC resize logic to take the appropriate action or indicate error as needed.
- Shrinking of volume is not allowed.
- Few environments require Pod restarts in order to get access to the resized volume (meaning that the underlying CSI driver doesn't support online expansions). This KEP won't address such scenarios.
- Modifying anything but the
resources.request.storage
field of thepersistentVolumeClaim
(eg,storageClassName
,labels
,annotations
). - Resizing a PVC associated with a StatefulSet directly by modifying the PVC spec will not affect the StatefulSet
.spec.volumeClaimTemplates
. - Resizing ephemeral volumes.
- Correcting failures resulting from bad input or environmental issues (such as insufficient capacity/quota). We'll make an attempt to do as much validation as we can in the API server, but cases of errors/failures past admission will have to be handled by the user. This KEP will discuss the different failures that could happen past admission.
Before we proceed to the proposal it is imperative to layout the various components involved
in creating and managing the persistent volumes associated with StatefulSet
.
- StatefulSet controller creates the PVCs from
volumeClaimTemplates
and associates them with the pods spawned from it. Reconciliation logic runs for every change made to the StatefulSetSpec
andStatus
reflects the progress towards making the real world state equal to theSpec
. - The validation logic in the API server is responsible for restricting the fields
which can be edited in a StatefulSet specification.
- Part of the validation that the API server performs for objects (eg
StatefulSet
,PVC
) happens in the object's API server code and part happens in an admission plugin.
- Part of the validation that the API server performs for objects (eg
- The RBAC specification in the controller policy is responsible for adding the required RBAC permission for StatefulSet.
- The StatefulSet controller also supports revision control.
The following changes are proposed to achieve the goal:
- Make the
volumeClaimTemplates
storage size (.spec.resources.request.storage
) editable by modification to the api server validation. We should allow the user to both increase or decrease the storage size in case of expansion failures. Allowing the user to decrease the storage size will give them the ability to address failed expansions, as described in KEP 1790. - Add reconciliation of the associated PVC's size and the individual
volumeClaimTemplates
size into the StatefulSet controller reconciliation. - Add PVC
patch
privilege to the StatefulSet controller'sClusterRole
RBAC. - Modification to the
StatefulSet.Status
to add additional field forvolumeClaimTemplates
to track volume expansion workflow status
Once the above changes are made, modification to any of the volumeClaimTemplates
storage
size (specifically to the .spec.resources.request.storage
field) by user will update (patch) the underlying PVC via the reconciliation performed by (2) above.
The existing PVC resize workflow kicks in at this point and attempts to resize the volume.
Two things to note about the diagram above
- It's an over simplification of the volume resize process or of how StatefulSet reconciliation works - but it should suffice in the context of this discussion.
- The numbers I denote on the arrows do not indicate the order in which operations happen. They are there for easier identification.
There are multiple real or potential places where errors/failures could happen when a user attempts to resize a volume claim template in a StatefulSet.
These are potential failures that could arise during validation or submission that the API server performs when sending the modified StatefulSet object. In the diagram above, it refers to the arrow denoted with (1)
.
We won't expand on any of the errors that could happen during this phase right now, as we'll talk about it during Validation But it's worth noting that this is the ideal place that errors could happen in - as it'll give the user immediate feedback.
These are errors that could happen when the reconciliation logic in the StatefulSet controller attempts to perform a patch
operation on the PVC object. These errors arise from either the PVC validation logic in the API server or its admission plugin. In the diagram above, it refers to the arrow denoted with (6)
.
- The storage class of the volume we're trying to resize doesn't allow for expansions.
- It shouldn't be a problem to have this validation in the StatefulSet validation/admission and we'll talk about it during Validation.
- We are trying to decrease the size of the volume.
- Specifically, we're trying to decrease the size of the volume to a size that's less (or equal) than it's capacity. Decreasing the size of the volume to a size that's greater than its current capacity is allowed, as its part of the recovery from failed expansions mechanism, described in KEP 1790.
- The volume we're trying to resize is not bound.
- This shouldn't be an issue because as we'll discuss later, we'll apply to reconciliation logic to the PVCs only when the pod that they're bound to is healthy and running.
This refers to a class of errors that could happen after the StatefulSet controller has successfully performed a patch
operation on the PVC object. These errors are asynchronous in nature and can happen at any time after the patch
operation. In the diagram above, it refers to the arrow denoted with (10)
. (Though in reality, kubelet
is involved in the resize as well, when an online expansion is performed, but we leave it out of the diagram).
- The CSI driver doesn't support online expansions.
- This is a scenario where the driver cannot do an expansion while the volume is published to a node/the pod that's using it is up. When this happens, the CSI driver will return an error when the CSI resizer calls it (eg
FailedPrecondition
). In this proposal we won't be supporting the scenario of not supporting online expansions. Ideally, we'd want to know ahead of time (during StatefulSet admission/validation) if the CSI driver supports online expansion or not - but it might not be so simple. We'll discuss it in details when we talk about Validation.
- This is a scenario where the driver cannot do an expansion while the volume is published to a node/the pod that's using it is up. When this happens, the CSI driver will return an error when the CSI resizer calls it (eg
- The CSI driver doesn't support offline expansions.
- This is a scenario where the volume must be published to a node/the pod that's using it must be up in order to perform an expansion. This is not exactly an error. The expansion process will be stuck until the volume is published on a node. This could happen in scenarios where a user attempts to resize a PVC that's not currently attached to a pod (eg a PVC is bound to a pod of the StatefulSet and then the user downscales the StatefulSet). We'll have to make sure to not get in the way of the user when try try to recover from this scenario.
- No space left/exceeding quota
- This is a scenario where the driver returns an error to the resizer (eg
ResourceExhaused
/OutOfRange
) because there's no space left for the resize process, or the user exceeds a quota on the provider they are using for storage. There is nothing we can do in this scenario other than not getting in the way of the user when they try to recover from it, and, continue reconciliation once they apply successful recovery. The recovery could involve steps like- Using the mechanisms that are described in KEP 1790
- Requesting more quota from the provider
- It's important to note that, it's possible for partial failures to occur as well - which might be more tricky for the user to handle. For example, a user could do a resize in a StatefulSet that has 3 replicas and the resize could succeed on 2 of them and fail only on the last.
- In such case, the user won't be able to apply the mechanisms in KEP 1790 and their only recourse would be to request more quota from the provider, or downsize their StatefulSet (number of replicas).
- This is a scenario where the driver returns an error to the resizer (eg
We'd ideally want as much validation to happen as early as possible, ie during StatefulSet admission/validation. In reality, perfect validation during StatefulSet admission might be very hard to achieve as it involves communication with parts that the StatefulSet admission plugin won't have direct access to, like the CSI driver. Nonetheless, we'll discuss all the possible validation that we could potentially be performing during admission whether or not we'll actually end up performing them in the final implementation.
We'd want to validate that the storage class of the volume we're trying to resize allows for expansions. This validation can be performed easily by creating a plugin admission for the StatefulSet object.
The plugin will use an informer to get the storage class object that the .spec.storageClassName
field points to, and check its allowVolumeExpansion
field. This should be similar to how the PVC admission plugin does it today
As mentioned earlier, we won't be supporting the scenario of drivers not supporting online expansion in the scope of this proposal. Thus, it would be ideal if we can validate during StatefulSet admission that the CSI driver supports online expansions. However, this could be a lot more tricky than validating whether or not the storage class supports expansions, during admission.
According to the CSI Spec - A driver must publish whether or not it supports online expansion via GetPluginCapabilities
.
However, that API won't be accessible to us from the API server, during StatefulSet admission, because it's part of the CSI driver's deployment/StatefulSet and not necessarily exposed externally via a LoadBalancer or an Ingress.
One alternative would be to have the CSI driver publish its capabilities through the CSIDriver
object. This could be done in two ways
- The CSI driver author would manually put the list of capabilities in the
CSIPlugin
object. - We'd have a plugin in the driver's deployment that would automatically populate the
CSIDriver
object with the capabilities of the driver by callingGetPluginCapabilities
.- We could also modify one of the existing plugins (eg
csi-resizer
,csi-provisioner
) - but none of them looks like a good candidate for this purpose.
- We could also modify one of the existing plugins (eg
The two options have the same drawbacks
- It will require modifying existing CSI drivers to support this feature.
- I noticed that some drivers don't explicitly specify support for online expansion when calling
GetPluginCapabilities
. Here's an example from the EBS CSI Driver
Given the above, this kind of validation won't be handled as part of this KEP.
We don't want to prevent users from decreasing the size of the volume in the StatefulSet PVC templates. Because, we want to support recovery from expansion failures as discussed in KEP 1790.
However, we'd ideally want to prevent users from decreasing the size of the StatefulSet PVC templates below (or equal) to the current capacity of the volume (.status.capacity.storage
).
This is similar to a validation that performed by the PVC API server validation code today.
There are two problems with performing this kind of validation during StatefulSet admission though.
- The validation logic will have to be informed of and access all of the different PVC objects that exist for the StatefulSet. This might involve a lot of logic that is owned by the StatefulSet controller today (eg PVC naming convensions, ordinals, etc)
- There could be a race condition where at the time of admission, we see the requested size as being less than the capacity of a PVC - but by the time that the StatefulSet controller tries to reconcole the PVC's request size to that of the template, the capabity of the PVC has already increased.
Given the above, this kind of validation won't be handled as part of this KEP.
- That being said, At the time of writing this document, I believe that KEP 1790 is still disabled behind a feature flag. In that case, one validation that we can do (and will do if it's still the case during implementation), is validate that as long as the feature flag for KEP 1790 is disabled, the request size of the modified PVC template cannot be less than the request size of the old PVC template.
Errors that happen during validation are trivial so we won't cover them in this section.
The two classes of errors we've like to have error indication than happen
The main mechanism for error reporting will be via events on either the StatefulSet object or the PVC objects. From a UX standpoint, having errors as events on the StatefulSet object, instead of spreading them across the PVC objects, might seem more deseriable - but in some cases, could introduce additional complexity and only marginal benefits.
For errors that happen When Patching The PVC Object, reporting on the StatefulSet object is fairly straightforward, since those errors arise from within the StatefulSet controller.
For example, an error that arises when trying to patch the PVC object to a size that's less or equal to its capacity, can result in the following event on the StatefulSet object.
kubectl describe sts ex1
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
.. .... .. .. .....
Warning FailedToPatchPVC 1s (x11 over 6s) statefulset-controller StatefulSet default/ex1 failed to patch PVC vol1-ex1-2 of Pod 2: PersistentVolumeClaim "vol1-ex1-2" is invalid: spec.resources.requests.storage: Forbidden: field can not be less than previous value
However, for errors that happen asynchronously, reporting on the StatefulSet object is not as straightforward. For example, if the CSI driver fails to resize the volume, the error will be reported as an event on the PVC object.
kubectl describe pvc vol1-ex1-2
Events:
Type Reason Age From Message
---- ------ ---- ---- -------
.. .... .. .. .....
Warning VolumeResizeFailed 4s (x6 over 9s) external-resizer hostpath.csi.k8s.io resize volume "pvc-8f38134a-ee99-4e70-8658-3cb492b8b11e" by resizer "hostpath.csi.k8s.io" failed: rpc error: code = OutOfRange desc = Requested capacity 9223372036854775807 exceeds maximum allowed 1125899906842624
In order to propagate that event to the StatefulSet object, we would have to subscribe to events from all namespaces (since the StatefulSet controller is namespace agnostic) kubeClient.CoreV1().Events("").Watch()
and filter out events that are attached to PVC objects. Then, we'd also have to filter out irrelevant events that weren't generated due to a fault in the resize operation.
This adds extra complication to the StatefulSet controller and could potentially hinder performance. The benefit is marginal improvement in UX (concentration of all error events under the StatefulSet object).
- The improvement is marginal because
- As we'll describe in later section, Events is not a reliable mechanism for getting feedback on the resize operation, and the user will benefit more from relying on the feedback mechanisms described here.
- Even though events might not be realiable for gathering feedback (especially during an automated deployment process, eg a CI/CD pipeline), the user could still rely on them for troubleshooting - but in that case, the user could just get/subscribe (
kubectl get events -w
) to events from the entire namespace.- It should also be possible to create tooling that will allow the user to aggregate events from multiple PVCs that were created by the same StatefulSet/template.
- It will be inconsistent with how events are reported on the StatefulSet pod.
- Error events that are generated by kubelet on a pod (eg
Failed to pull image
) are not propagated to the pod's StatefulSet object.
- Error events that are generated by kubelet on a pod (eg
It is useful to provide feedback to the user about the status of the resize operation. This main benefits are
- For errors that we can't properly report via an event on the StatefulSet object - a delay in feedback could prompt the user to investigate the issue, instead of leaving them oblivious to it.
- Even for scenarios where there are no errors, it could be useful to provide feedback to the user that the resize operation has completed or is still in progress. If the user has multiple StatefulSets and/or cluster they are resizing, this will give them the option to decide if they want to wait for the resize operation to complete before proceeding with other operations.
- Feedback is especially useful if the resize happens as part of an automated deployment mechanism (ie a CI/CD pipeline). In such cases, the user doesn't necessarily keep an eye on the deployment process so proper feedback is really important.
We propose to provide feedback by adding a new field to the StatefulSet status object. The new field will be an array that contains
- The name of the volume claim template
- The current number of replicas for which the size of that PVC template is fully reconciled with the capacity of the PVC object (the
.status.capacity.storage
field) - The last generation (pertains to the
.metadata.generation
field on the StatefulSet object) in which the size of that PVC template was fully reconciled with the capacity of the PVC object for all replicas.
Here's an example of what the StatefulSet status object with the new field will look like
status:
availableReplicas: 3
collisionCount: 0
currentReplicas: 3
currentRevision: ex1-54c5bd476c
observedGeneration: 3
readyReplicas: 3
replicas: 3
updateRevision: ex1-54c5bd476c
updatedReplicas: 3
volumeClaimTemplates:
- finishedReconciliationGeneration: 2
readyReplicas: 1
templateName: vol1
- finishedReconciliationGeneration: 3
readyReplicas: 3
templateName: vol2
Some notes about the new fields
- The
finishedReconciliationGeneration
field is optional and won't appear if the PVC template is newly created and isn't reconciled yet. - The
readyReplicas
will reset to 0 when the size of the PVC template is changed. It will increase only as the new pods become healthy and ready. - The
readyReplicas
field also pertains to updates to the StatefulSet that are not related to the PVC templates. (eg if you modify the pod image,readyReplicas
will reset to 0 and increase only as the new pods become healthy and ready)
In the example above, the size of the vol1
PVC template has been recently resized (while the size of the vol2
PVC template has not). At the time of capturing it, only 1 PVC of that template has been fully reconciled with the capacity of the PVC object.
With the feedback mechanism, the user will be able to wait for partial feedback as well (eg, only 1 of the PVC templates has been fully reconciled, or, only 80% of the replicas for a given PVC template have been fully reconciled).
This new feedback functionality will be able to be integrated into existing deployment tools that already have the ability to poll the status of the StatefulSet object for feedback (eg wait until all the pods in the StatefulSet are fully ready).
For example, Helm
StatefulSet supports revision control. The storage size change will not be added to the fields supporting revision control. Any change in revision - going forward to a new version or going back to an older version would not make any changes to the persistent volume size. Any changes requiring resize of the volume will need editing of the StatefulSet storage size specification.
We are going to change how the StatefulSet reconciliation logic works. Since this is a change to a very central component, to mitigate it, we'd have to introduce a feature flag. The feature flag will be disabled by default. When disabled, the reconciliation logic in the StatefulSet should behave exactly as it behaves today. And, the validation logic that prevents user from modifying the size of a PVC template of a StatefulSet today, should be kept as is as long as the feature flag is disabled.
We are adding a new array field to the StatefulSet status object. I don't believe this change will impact any existing clients/tools - but to be careful, we should not expose this field in the API server as long as the feature flag described above is disabled. (We technically can't put a struct field behind a feature flag - but, by marking it as omitempty
and putting the logic that populates it in the StatefulSet controller, behind a feature flag, we can effectively hide it from the API server)
Apart from mitigating the risk of breaking changes, the new feedback mechanism might not be immediately intuitive to all users and/or immediately be adapted by all deployment tools. We'll have to provide documentation and examples for how to use it and push for adaption by deployment tools. (such as Helm)
We are technically not making any changes to the revision control logic. But, the fact that changes to the PVC template won't be part of revision control, could cause confusion to user. To mitigate this, we should add documentation that explains this behavior.
Unreliability of Events
The error reporting mechanisms that we describe here, suffer from known reliability issues. The event reporting mechanisms works in a fire/forget fashion. The controller tries to push an event into a channel but will abort if the channel is full.
We'll need to document the fact that the user should probably use events as a troubleshooting mechanism, and rely on the feedback mechanism for reliable feedback, especially when deploying from an automated CI/CD pipeline.
This proposal introduces multiple failure scenarios that are going to be handled at the different places, and reported in different manner. Some will be caught during validation, other will be reported as an error event and will require troubleshooting by the user.
We'll have to provide clear documentation on the different failure scenarios, what the user can expect to be caught during valiation (eg storage class not supporting expansion) and what will be caught only during the reconciliation logic (eg user is trying to decrease size below or equal to PVC capacity; the CSI driver doesn't support online expansion).
The only API changes we'll make, is to add the volumeClaimTemplates
array field to the StatefulSet status object as described in the Feedback section.
Modifications would be made to ValidateStatefulSetUpdate
in order to allow changes in the
storage size of the volume (.spec.resources.requests.storage
) in the VolumeClaimTemplates
.
Now that StatefulSet controller needs to patch the PVC, patch
will be added to the statefulset-controller
role in the buildControllerRoles.
In order to be notified of capacity changes in PVC objects (arrow (15)
in the diagram here), which is necessary for the feedback mechanism, we'll perform the following changes
- When creating a new a PVC object during StatefulSet reconciliation, we'll add an annotations to the PVC object, denoting the namespace and name of the StatefulSet object.
- We don't want to use the
ownerReference
in this case, because that field is already utilized by the StatefulSet PVC retention mechanism.
- We don't want to use the
- We'll start listening to PVC update (not add/delete) events in the StatefulSet controller. When an update happens, we'll extract the StatefulSet object that we should do reconciliation for from the annotation we added in the previous step.
- We'll also verify that only the
.status
field has changed in the PVC object. Otherwise, we won't continue reconciliation.
- We'll also verify that only the
We would want to perform PVC reconciliation only for replicas that are in the latest revision of the StatefulSet object, and are running, but not necessarily ready (pass readiness checks).
- The reason for only performing reconciliation on running and up to date (match to the latest revision), is that we don't want to do unnecessary reconciliation works on replicas that are not running and might never run again as far as we're concerned.
- Another reason is that we want for the PVC reconciliation to match the cadence of pod creation/modification in the StatefulSet. If the user uses monotonic (
OrderedReadyPodManagement
) roll out strategy, we'd want the PVC reconciliation process to respect that strategy. This is especially important considering that PVC reconciliation could fail asynchronically (eg, due to out of space/exceeding quota issues) and we want the failure to be predictable. - Another reason is that some CSI drivers don't support offline expansion (meaning, the volume has to be published/attached to a pod)
- This means that if the StatefulSet is scaled to 0, and the user resizes one of the PVC templates, the change won't be reconciled (even if the PVC objects exist), until the StatefulSet is scaled up again.
- Another reason is that we want for the PVC reconciliation to match the cadence of pod creation/modification in the StatefulSet. If the user uses monotonic (
- The reason for only waiting for the replicas/pods to be in a running state, and not necessarily in a ready state, is that the user might have readiness checks that will fail until the PVC is resized. Waiting for the pod to become fully ready, might put us in a deadlock.
- For example, imagine a readiness check that tests that the filesystem has at least 10% space left.
On top of it, when we ecnounter an error during patching the PVC
- If the error is transient (not part of Errors When Patching The PVC Object), record it on the StatefulSet object, and return the error immediately (this will stop the reconciliation process from progressing to other replicas and will trigger a retry).
- If the error is not transient (part of Errors When Patching The PVC Object), record it on the StatefulSet object, and
- If the roll out strategy of the StatefulSet is monotonic (
OrderedReadyPodManagement
), return immediately - but do not return the error, because if the error is not transient, we do not want to retry. - If the roll out strategy is burst (not monotonic), continue.
- If the roll out strategy of the StatefulSet is monotonic (
for pvcTemplate in statefulset {
statefulset.status.pvcTemplateStatus[pvcTemplate.name].readyReplicas = 0
}
for replica in replicas {
// this is part of the existing replicas loop in the StatefulSet reconciliation process (https://github.com/kubernetes/kubernetes/blob/d89d5ab2680bc74fe4487ad71e514f4e0812d9ce/pkg/controller/statefulset/stateful_set_control.go#L397)
// ...
if getPodRevision(replica) == statefulset.revision && replica.Status.Phase == PodRunning && replica.DeletionTimestamp == nil {
for pvcTemplate in statefulset {
pvc, err = getPodPVC(replica, pvcTemplate)
if err != nil {
// this is either a transient error, or an error that shouldn't happen, because the pod is running and the PVC should exist
recordError(statefulset, err)
return err
}
if requestStorage(pvcTemplate) != requestStorage(pvc) {
err = patchPVC(pvc, requestStorage(pvcTemplate))
if err != nil {
recordError(statefulset, err)
if isTransientError(err) {
return err
} else if monotonic {
return nil
}
}
}
if requestStorage(pvcTemplate) == capacityStorage(pvc) {
statefulset.status.pvcTemplateStatus[pvcTemplate.name].readyReplicas++
}
}
}
// ...
}
for pvcTemplate in statefulset {
if statefulset.status.pvcTemplateStatus[pvcTemplate.name].readyReplicas == statefulset.spec.replicas {
statefulset.status.pvcTemplateStatus[pvcTemplate.name].finishedReconciliationGeneration = statefulset.generation
}
}
- E2E tests for volume expansion of PVC associated with statefulsets.
- Ensure statefulset controller revision works properly after resize.
- Test resize failure recovery after KEP 1790 has been implemented.
-
We'll need to make sure that while the feature flag is disabled
- Validation that doesn't allow PVC template to change is still intact
- The new StatefulSet status field is not exposed.
- No change to existing StatefulSet functionality
-
The following scenarios will need to be covered by a combination of unit/integration and manual testing
- A user increases the size of a PVC template with a single replica StatefulSet. The resize goes through. While the resize is in progress, the
finishedReconciliationGeneration
in the StatefulSet status for that PVC template is not increased to the generation of the StatefulSet. (Resize is in progress as long as the.status.capacity.storage
of the PVC is not equal to the.spec.resources.requests.storage
of the PVC) - A user increases the size of a PVC template with multiple replicas. For both monotonic and burst roll out strategies.
- The resize (patching of the PVC object) begins only once the replica is running. (We can simulate a restart to the replica by changing one of its fields in its spec)
- The
readyReplicas
in the StatefulSet status for that PVC template increases as the PVCs are resized.
- A user decreases the size of a PVC template. The user should be able to successfully submit the modified StatefulSet. There should be an error message (event) generated on the StatefulSet object stating that the patching of the PVC object has failed. While in failed state, the
finishedReconciliationGeneration
in the StatefulSet status for that PVC template is not increased to the generation of the StatefulSet, and thereadyReplicas
field stays at 0. When increasing the size of the PVC template back to its original size, the error should be resolved and thefinishedReconciliationGeneration
in the StatefulSet status for that PVC template should be increased to the generation of the StatefulSet and thereadyReplicas
field should be updated to the number of replicas in the StatefulSet.- The same scenario should be tested for having multiple PVC templates. The tester should note that after trying to decrease the size of one of the PVC templates, its
finishedReconciliationGeneration
in the StatefulSet status is not increased to the generation of the StatefulSet, and thereadyReplicas
field stays at 0. But for the other PVC template, itsfinishedReconciliationGeneration
in the StatefulSet status is increased to the generation of the StatefulSet, and thereadyReplicas
field is updated to the number of replicas in the StatefulSet. - The same test should be done for monotonic and burst roll out strategies.
- For monotonic, the StatefulSet reconciliation should stay stuck on the first replica when it encounters the patching error.
- For burst, the StatefulSet reconciliation should continue to the next replica when it encounters the patching error.
- The same scenario should be tested for having multiple PVC templates. The tester should note that after trying to decrease the size of one of the PVC templates, its
- A user increases the size of a PVC template with a single replica StatefulSet. The resize goes through. While the resize is in progress, the
- Alpha in 1.28 behind
StatefulSetVolumeExpansion
feature gate which will be set tofalse
by default. Statefulset controller revision behavior will be explained in the documentation. - Beta in 1.29: We are going to move this to beta with enhanced e2e and more stability improvements.
- GA in 1.30
- [ x ] Feature gate (also fill in values in
kep.yaml
)- Feature gate name: StatefulSetVolumeExpansion
- Components depending on the feature gate:
- kube-controller-manager
- kube-apiserver, to relax the restriction on editing VolumeClaimTemplate size.
Yes, users will now be able to modify the requested size in the VolumeClaimsTemplate section of the statefulset spec.
Yes. If the feature is disabled, the user will not be able to resize PVCs associated with a statefulset anymore. Any existing statefulset reconciliation operations might be affected if feature is disabled while the volume is being resized.
Statefulset reconciliation process for volume size will remove any anomalies which were left behind during disabling the feature. Volumes associated with statefulsets will be allowed to resize.
Unit tests to test behaviour of code when feature gate is enabled and disabled will be written to test both behaviors.
Is the rollout accompanied by any deprecations and/or removals of features, APIs, fields of API types, flags, etc.?
- Events
- Event Reason:
- API .status
- Condition name:
- Other field:
- Other (treat as last resort)
- Details:
What are the SLIs (Service Level Indicators) an operator can use to determine the health of the service?
- Metrics
- Metric name:
- [Optional] Aggregation method:
- Components exposing the metric:
- Other (treat as last resort)
- Details:
Are there any missing metrics that would be useful to have to improve observability of this feature?
Will enabling / using this feature result in increasing time taken by any operations covered by existing SLIs/SLOs?
Will enabling / using this feature result in non-negligible increase of resource usage (CPU, RAM, disk, IO, ...) in any components?
An alternative approach considered with respect to the revision control is to add the volume size as part of the revision control. In such a case, if we did an expand volume, users will not be able to rollback to previous versions since volume shrink is not yet supported. Hence we take a pragmatic path and keep the volume size separate from controller revision changes.