Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory usage exploding #150

Open
mnsmar opened this issue Feb 2, 2024 · 3 comments
Open

Memory usage exploding #150

mnsmar opened this issue Feb 2, 2024 · 3 comments

Comments

@mnsmar
Copy link

mnsmar commented Feb 2, 2024

Hi,

The memory usage is exploding during m6anet data_prep and the run is killed. Any idea what's causing this? Here is the command I use:

m6anet dataprep --eventalign eventalign.txt --out_dir ouput --readcount_max 1000000 --n_processes 10
image
@yuukiiwa
Copy link
Collaborator

yuukiiwa commented Feb 8, 2024

Hi @mnsmar,

This memory explosion was likely due to --readcount_max 10000000. You can try using the default --readcount_max, which is 1000.

Thanks!

Best wishes,
Yuk Kei

@mnsmar
Copy link
Author

mnsmar commented Feb 8, 2024

Thanks for the reply @yuukiiwa. The reason we use a high --readcount_max is because the majority of our reads are on a very specific locus. As far as I understand, when we use a lower --readcount_max these loci are skipped. Is there a way to avoid skipping these regions but randomly use up to --readcount_max reads in a locus?

@yuukiiwa
Copy link
Collaborator

yuukiiwa commented Feb 9, 2024

Hi @mnsmar,

Thanks for the explanation of your data!

Given that most reads are of the same sites, you can consider splitting them into smaller eventalign.txt files and run m6anet dataprep separately on them.

Thanks!

Best wishes,
Yuk Kei

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants