I have gone through Cuda programming guide but still not clear where does cuda kernel reside on GPU? In other words, in which memory segment does it reside?
Also, How do I know what is the max kernel size supported by my device? Whether max kernel size depend on number of simultaneous kernels loaded on device?
The instructions are stored in global memory that is inaccessible to the user but are prefetched into an instruction cache during execution.
The maximum kernel size is stated in the Programming Guide in section G.1: 2 million instructions.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With