Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use disk option(s) #75

Open
mirh opened this issue Jul 30, 2020 · 1 comment
Open

Use disk option(s) #75

mirh opened this issue Jul 30, 2020 · 1 comment

Comments

@mirh
Copy link

mirh commented Jul 30, 2020

I get why everything happens in memory, but sometimes you just cannot fit your whole test file into it.
Or maybe you can, but it's the combined size with the RAM requirements of the most "hungry compressors" to make it a bad day for you.
Or even, you could, but the "clue" you are looking for happens below the bandwidth threshold of your non volatile memory.

So, putting aside that it I/O bottlenecks could be detected and flagged (I suppose it may also be true for RAM in some crazy situation?), could we get some kind of knob to control the "disk type"?
I see three possible levels for this:

  • compression on disk (since all algorithms should be asymmetrically more tasking here), everything else as usual, meaning the ending archive gets copied back in memory before decompression
  • an intermediate mode where once a single file/element/part is decompressed, it may as well be deleted on the spot (making best case RAM requirement be: algorithm footprint + archive size + sector size)
  • everything happens on disk
@inikep
Copy link
Owner

inikep commented Sep 1, 2020

Please try -m option which splits large files into parts.
You can also compress a large file in blocks/chunks with -b option.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants