I've got a problem that I want to split across multiple CUDA devices, but I suspect my current system architecture is holding me back;
What I've set up is a GPU class, with functions that perform operations on the GPU (strange that). These operations are of the style
for iteration in range(maxval):
result[iteration]=gpuinstance.gpufunction(arguments,iteration)
I'd imagined that there would be N gpuinstances for N devices, but I don't know enough about multiprocessing to see the simplest way of applying this so that each device is asynchronously assigned, and strangely few of the examples that I came across gave concrete demonstrations of collating results after processing.
Can anyone give me any pointers in this area?
UPDATE Thank you Kaloyan for your guidance in terms of the multiprocessing area; if CUDA wasn't specifically the sticking point I'd be marking you as answered. Sorry.
Perviously to playing with this implementation, the gpuinstance class initiated the CUDA device with import pycuda.autoinit
But that didn't appear to work, throwing invalid context
errors as soon as each (correctly scoped) thread met a cuda command. I then tried manual initialisation in the __init__
constructor of the class with...
pycuda.driver.init()
self.mydev=pycuda.driver.Device(devid) #this is passed at instantiation of class
self.ctx=self.mydev.make_context()
self.ctx.push()
My assumption here is that the context is preserved between the list of gpuinstances is created and when the threads use them, so each device is sitting pretty in its own context.
(I also implemented a destructor to take care of pop/detach
cleanup)
Problem is, invalid context
exceptions are still appearing as soon as the thread tries to touch CUDA.
Any ideas folks? And Thanks to getting this far. Automatic upvotes for people working 'banana' into their answer! :P
You need to get all your bananas lined up on the CUDA side of things first, then think about the best way to get this done in Python [shameless rep whoring, I know].
The CUDA multi-GPU model is pretty straightforward pre 4.0 - each GPU has its own context, and each context must be established by a different host thread. So the idea in pseudocode is:
In Python, this might look something like this:
import threading
from pycuda import driver
class gpuThread(threading.Thread):
def __init__(self, gpuid):
threading.Thread.__init__(self)
self.ctx = driver.Device(gpuid).make_context()
self.device = self.ctx.get_device()
def run(self):
print "%s has device %s, api version %s" \
% (self.getName(), self.device.name(), self.ctx.get_api_version())
# Profit!
def join(self):
self.ctx.detach()
threading.Thread.join(self)
driver.init()
ngpus = driver.Device.count()
for i in range(ngpus):
t = gpuThread(i)
t.start()
t.join()
This assumes it is safe to just establish a context without any checking of the device beforehand. Ideally you would check the compute mode to make sure it is safe to try, then use an exception handler in case a device is busy. But hopefully this gives the basic idea.
What you need is a multi-threaded implementation of the map
built-in function. Here is one implementation. That, with a little modification to suit your particular needs, you get:
import threading
def cuda_map(args_list, gpu_instances):
result = [None] * len(args_list)
def task_wrapper(gpu_instance, task_indices):
for i in task_indices:
result[i] = gpu_instance.gpufunction(args_list[i])
threads = [threading.Thread(
target=task_wrapper,
args=(gpu_i, list(xrange(len(args_list)))[i::len(gpu_instances)])
) for i, gpu_i in enumerate(gpu_instances)]
for t in threads:
t.start()
for t in threads:
t.join()
return result
It is more or less the same as what you have above, with the big difference being that you don't spend time waiting for each single completion of the gpufunction
.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With