-
Notifications
You must be signed in to change notification settings - Fork 155
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Many cause,not_found errors producing GBs of logs #1001
Comments
@MrApe Thank you for your report. We would like to know your LeoFS' environment and the current state as below:
|
@MrApe Let me ask a few additional questions.
If your LeoFS version isn't the latest one (1.3.8) and you used |
Thanks for having a look. This is the environment:
The only content of the logs of leo_storage are error messages as above. No errors in manager or gateway. I did not use |
@MrApe Thank you for sharing the informative report. We're going to survey this issue from today. |
WIP |
@MrApe I've tried to reproduce this case but still no luck however with 1.3.4, there are many bugs related to handling large objects so I'd recommend you upgrade >= 1.3.8 (1.4.0 would be the best at the moment). Regarding remained queue items, would you like to try this procedure https://gist.github.com/mocchira/1c4852c57c7b328aef46eb234b74093b ? That would free the queue on leo_storage up. I hope you will find it helpful. |
I temporarily solved the problem by resetting the storage completely and importing the objects. It runs fine now. However, I will do an update as soon as possible.
…On 4. Apr 2018, 07:36 +0200, yoshiyuki.kanno ***@***.***>, wrote:
@MrApe I've tried to reproduce this case but still no luck however with 1.3.4, there are many bugs related to handling large objects so I'd recommend you upgrade >= 1.3.8 (1.4.0 would be the best at the moment).
Regarding remained queue items, would you like to try this procedure https://gist.github.com/mocchira/1c4852c57c7b328aef46eb234b74093b ? That would free the queue on leo_storage up. I hope you will find it helpful.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub, or mute the thread.
|
@MrApe Great to hear that. I will close this issue however if you find something wrong after upgrading then feel free to reopen this issue or file another issue :) |
I have a 4 node cluster with N=3, D=R=W=2 setting. I deleted a lot of files using s3cmd. Now each storage node produces GBs of error logs like this:
(as you can see it's a project fifo installation, doing backups to leofs)
The objects all belong to the files deleted. The files I did not delete are fine. I did a
leofs-adm recover-cluster
to make the nodes detect orphan objects and rebuild the ring.leofs-adm mq-stats
shows this asrunning
for about a week(!) now:What is producing these errors and how do I solve this problem?
Thanks in advance.
Best, Jonas
The text was updated successfully, but these errors were encountered: