I'm trying to generate multiple transcriptomic datasets in parallel with different parameters (error model, coverage/depth, and sequence identity), however, Badread consumes a lot of memory while doing this and occasionally will cause my machine to freeze while running the simulation.
Is there a way to control the memory usage, or better optimize the process of simulating multiple datasets in parallel?
I'm trying to generate multiple transcriptomic datasets in parallel with different parameters (error model, coverage/depth, and sequence identity), however, Badread consumes a lot of memory while doing this and occasionally will cause my machine to freeze while running the simulation.
Is there a way to control the memory usage, or better optimize the process of simulating multiple datasets in parallel?