Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MultiQC-Additional overview stat table #134
MultiQC-Additional overview stat table #134
Changes from 2 commits
0521047
2d42736
8d7e5c5
eddab30
4d52218
922682c
1d998e0
879e41b
9cf2794
2fb5df0
529182c
1dff437
a9a3d8e
5f00c9b
File filter
Filter by extension
Conversations
Jump to
There are no files selected for viewing
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am not sure I understood all details of the new code, but since it caused quite some memory usage, I have a few questions/suggestions since I am wondering if you really need that many joins:
conditions_peptides
: if the resulting df become too large, you could drop thecondition_name
(anyway the same within this context) andcount
before joining to the predictionspredictions
directly after reading in for entries that are above the threshold, and drop theprediction_score
to reduce the sizeBrings me to the next question: you just provide the count of unique binders, right? Maybe that should be stated somewhere more clearly.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I managed to reduce the used Memory to around 50GB which is only 10GB more compared to the previous version not including peptide predictions!