Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CI: integration timing results are randomly incomplete #8358

Closed
edsantiago opened this issue Nov 16, 2020 · 7 comments · Fixed by #9124
Closed

CI: integration timing results are randomly incomplete #8358

edsantiago opened this issue Nov 16, 2020 · 7 comments · Fixed by #9124
Labels
In Progress This issue is actively being worked by the assignee, please do not work on this at this time. locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments.

Comments

@edsantiago
Copy link
Member

Compare two CI logs, one ubuntu-19 and one 20. Search in-page for test path, or reimport, or check dnsname, or many many others (I'm just offering short easy-to-search-for strings). Note that those appear in the bottom timing results section only on one page, not the other, although the tests ran on both.

Something is gobbling up timing results; this seems like a good place to start looking:

func processTestResult(f GinkgoTestDescription) {
tr := testResult{length: f.Duration.Seconds(), name: f.TestText}
testResults = append(testResults, tr)
}
(maybe this isn't thread-safe?)

@baude weren't you looking into this some months ago?

@github-actions
Copy link

A friendly reminder that this issue had no activity for 30 days.

@rhatdan
Copy link
Member

rhatdan commented Dec 17, 2020

@baude @edsantiago Still an issue?

@edsantiago
Copy link
Member Author

Yes but I seem to be the only one who cares

@vrothberg
Copy link
Member

I just hit performance variations in the remote system tests in #8747.

@edsantiago and I looked into a VM together and noticed performance variations on Ubuntu 20.04. #8747 bumps the time out from 60 to 120 seconds.

@github-actions
Copy link

A friendly reminder that this issue had no activity for 30 days.

@rhatdan
Copy link
Member

rhatdan commented Jan 25, 2021

@edsantiago @vrothberg Since we merged the bump can we close this issue?

@edsantiago
Copy link
Member Author

The issue is not fixed. Example: ubuntu 2010 log for PR 9082. Search in-page for "preserves container". The test is run, but there is no summary line. Compare to ubuntu 2004 log for same PR. Here the test is run and also summarized at bottom.

This is a problem because we still don't understand why ubuntu 2010 runs slower than 2004, and without logged timing results there is no possible way to actually look into the difference.

I really feel like this must be a trivial one-line fix for someone who understands Go arrays and threads.

vrothberg added a commit to vrothberg/libpod that referenced this issue Jan 27, 2021
Use a mutex to synchronize the slice for storing tests results.
Running the e2e tests in parallel is otherwise subject to race
conditions surfacing in lost entries.

Fixes: containers#8358
Signed-off-by: Valentin Rothberg <[email protected]>
@vrothberg vrothberg added In Progress This issue is actively being worked by the assignee, please do not work on this at this time. and removed stale-issue labels Jan 27, 2021
@github-actions github-actions bot added the locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments. label Sep 22, 2023
@github-actions github-actions bot locked as resolved and limited conversation to collaborators Sep 22, 2023
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
In Progress This issue is actively being worked by the assignee, please do not work on this at this time. locked - please file new issue/PR Assist humans wanting to comment on an old issue or PR with locked comments.
Projects
None yet
Development

Successfully merging a pull request may close this issue.

3 participants