-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Epic: Ensure durability for user workspace files #7901
Comments
|
this is not required for XFS. |
this can be optional for the first iteration |
@sagor999 as a heads up, I added a few observability tasks. One of the first ones we'll need (if it doesn't already exist) is the ability to inspect backups and restores now being done with TAR. For example, this way we can measure duration for both. |
@sagor999 @jenting are there any more integration tests that need to be added for new code we've written? In other words, I see you've fixed existing tests, but wanted to double check for new test needs. For example, one test I can think of, would be a test that kills a pod, relies on a process to backup the orphaned PVC, and then assert that the PVC is gone (because it was snapshotted). |
Question: How would someone who ran out of hours get their data back? (re: #14393) |
Question: Will this change prevent us to download a single file in the workspace? (Will the "Download..." button in the right-click menu of a file still available?) |
No, this is about downloading the workspace content backup. You can still download individual files from your running workspace depending on how you connect to it. E.g. with Vs Code, just drag and drop. |
Maybe this issue should be part of this epic to not lose my workspace's content on a regular basis: #11183 |
Update: After internal discussions, given backup success ratio is high and stable following adjacent improvements, and that the implementation of the new design will be considerably faster to do after #11416, we have decided to pause this effort until then. PS: @6uliver I believe the root cause of that issue is different from the context of this one. I will follow-up on that one there. 🙏 |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions. |
Summary
Better protect user data
Context
Sometimes a workspace, node, or workspace cluster fail and the user data cannot be backed up to cloud storage, resulting in data loss. A related incident for a global outage. A related RFC where we are discussing solutions.
Value
By better handling user data, users will trust that even if the Gitpod service is unavailable, once it is online, they will not lose data.
Acceptance criteria
User data is persisted in such a way that even if there is a workspace, node, or cluster failure, the data is accessible to be backed up at a later time.
Tasks
Ops:
automate deployment of GCP storageClasses as part of cluster creation operation
(specifydiscard
mount option)workspace-clusters
Design:
Product changes:
/workspace
directory is owned bynobody
#14003installer: add a test to ensure that CSI is working as expected prior to installing Gitpod #10614(Moves to Epic: PVC (Persistent Volume Claims) on Self-Hosted #11476)Tests:
Manually test PVC CSI snapshot backup/restore in AWS #10211(Moves to Epic: PVC (Persistent Volume Claims) on Self-Hosted #11476)Manually test PVC CSI snapshot backup/restore in Azure #10212(Moves to Epic: PVC (Persistent Volume Claims) on Self-Hosted #11476)Bug
Should solve:
Day 2:
no backup found
is hidden #14451Front conversations
The text was updated successfully, but these errors were encountered: