I am using multiprocessing.imap_unordered
to perform a computation on a list of values:
def process_parallel(fnc, some_list):
pool = multiprocessing.Pool()
for result in pool.imap_unordered(fnc, some_list):
for x in result:
yield x
pool.terminate()
Each call to fnc
returns a HUGE object as a result, by design. I can store N instances of such object in RAM, where N ~ cpu_count, but not much more (not hundreds).
Now, using this function takes up too much memory. The memory is entirely spent in the main process, not in the workers.
How does imap_unordered
store the finished results? I mean the results that were already returned by workers but not yet passed on to user. I thought it was smart and only computed them "lazily" as needed, but apparently not.
It looks like since I cannot consume the results of process_parallel
fast enough, the pool keeps queueing these huge objects from fnc
somewhere, internally, and then blows up. Is there a way to avoid this? Limit its internal queue somehow?
I'm using Python2.7. Cheers.
Python 3.8 introduced a new module multiprocessing. shared_memory that provides shared memory for direct access across processes.
multiprocessing is a package that supports spawning processes using an API similar to the threading module. The multiprocessing package offers both local and remote concurrency, effectively side-stepping the Global Interpreter Lock by using subprocesses instead of threads.
Yes, with Jython you've real multi-threading.
Python optimizes memory utilization by allocating the same object reference to a new variable if the object already exists with the same value. That is why python is called more memory efficient.
As you can see by looking into the corresponding source file (python2.7/multiprocessing/pool.py
), the IMapUnorderedIterator uses a collections.deque
instance for storing the results. If a new item comes in, it is added and removed in the iteration.
As you suggested, if another huge object comes in while the main thread is still processing the object, those will be stored in memory too.
What you might try is something like this:
it = pool.imap_unordered(fnc, some_list)
for result in it:
it._cond.acquire()
for x in result:
yield x
it._cond.release()
This should cause the task-result-receiver-thread to get blocked while you process an item if it is trying to put the next object into the deque. Thus there should not be more than two of the huge objects in memory. If that works for your case, I don't know ;)
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With