-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Velero maintenance job failing with oomkilled and logger name="[index-blob-manager]" sublevel=error #8474
Comments
For the mentioned errors with |
The cause for this problem is the memory usage exceeds the limit assigned to Velero server pod where the maintenance is running. |
Hi @Lyndon-Li, level=warning msg="Found too many index blobs (3519), this may result in degraded performance.\n\nPlease ensure periodic repository maintenance is enabled or run 'kopia maintenance'." logModule=kopia/kopia/format logSource="pkg/kopia/kopia_log.go:101" logger name="[shared-manager]" sublevel=error |
What do you mean by stuck? Is the pod still in Running state? |
maintenance job running for 19hrs. |
Could you observe the cpu and memory usage of the maintenance job pod? Additionally, please share how much data has backed up to the repository. |
backed up is 763 GB |
What is the size of most files in your backup? |
main pvc has around 700 gb of data |
What is the file size in the volume? |
how can i check file size? |
Methods to Check File Sizes Here's a breakdown of the most common and effective methods:
Important Considerations
|
What steps did you take and what happened:
We have installed Velero version 1.14.0, we started facing issues after a few days. The Velero maintenance job pod is failing and Velero pod is restarting
time="2024-11-26T23:51:33Z" level=warning msg="active indexes [xn0_000301012dec8243b5846445a770d7e9-s36cadf8751419dd612e-c1
deletion watermark 0001-01-01 00:00:00 +0000 UTC" logModule=kopia/kopia/format logSource="pkg/kopia/kopia_log.go:101" logger name="[index-blob-manager]" sublevel=error
time="2024-11-26T23:52:00Z" level=warning msg="Found too many index blobs (2438), this may result in degraded performance.\n\nPlease ensure periodic repository maintenance is enabled or run 'kopia maintenance'." logModule=kopia/kopia/format logSource="pkg/kopia/kopia_log.go:101" logger name="[shared-manager]" sublevel=error
time="2024-11-26T23:52:00Z" level=info msg="Start to open repo for maintenance, allow index write on load" logSource="pkg/repository/udmrepo/kopialib/lib_repo.go:165"
time="2024-11-26T23:52:01Z" level=warning msg="active indexes [xn0_000301012dec8243b5846445a770d7e9-s36cadf8751419dd612e-c1 xn0_00280b34b036384ed96d5578acf4c6fb-se642334b6c6a9f9b12e-c1 xn0_0063062125fbddc7796399dc24e67ec9-s0ee97e381a38601b12f-c1 xn0_008deb665d149583be338e10fe647591-s690b6323aa26dd5012e-c1 xn0_0092e2fcacda8223a85b0a7699586ab4-se54981af127c1f4b12f-c1 xn0_00abc8ac6601e30814b8494501424606-se875ff6f29fdd0d412f-c1 xn0_00ade8ccce8aee4c4830601ee2e6ec10-s015f50b8b530ad2d12f-c1
Uploading bundle-2024-12-02-09-55-22.tar.gz…
If you are using velero v1.7.0+:
Please use
velero debug --backup <backupname> --restore <restorename>
to generate the support bundle, and attach to this issue, more options please refer tovelero debug --help
If you are using earlier versions:
Please provide the output of the following commands (Pasting long output into a GitHub gist or other pastebin is fine.)
kubectl logs deployment/velero -n velero
velero backup describe <backupname>
orkubectl get backup/<backupname> -n velero -o yaml
velero backup logs <backupname>
velero restore describe <restorename>
orkubectl get restore/<restorename> -n velero -o yaml
velero restore logs <restorename>
Anything else you would like to add:
Environment:
velero version
): v1.14.0velero client config get features
): v1.12.3kubectl version
): 1.28/etc/os-release
): Ubuntu 22.04Vote on this issue!
This is an invitation to the Velero community to vote on issues, you can see the project's top voted issues listed here.
Use the "reaction smiley face" up to the right of this comment to vote.
The text was updated successfully, but these errors were encountered: