A site for solving at least some of your technical problems...
A site for solving at least some of your technical problems...
As I'm slowly learning CUDA, I got some of the tests running. Those under 1_Utilities actually define the parameters (stats) of the card. There are the parameters for an NVidia Quadro 600:
Device 0: "Quadro 600" CUDA Driver Version / Runtime Version 5.0 / 5.0 CUDA Capability Major/Minor version number: 2.1 Total amount of global memory: 1024 MBytes (1073283072 bytes) ( 2) Multiprocessors x ( 48) CUDA Cores/MP: 96 CUDA Cores GPU Clock rate: 1280 MHz (1.28 GHz) Memory Clock rate: 800 Mhz Memory Bus Width: 128-bit L2 Cache Size: 131072 bytes Max Texture Dimension Size (x,y,z) 1D=(65536), 2D=(65536,65535), 3D=(2048,2048,2048) Max Layered Texture Size (dim) x layers 1D=(16384) x 2048, 2D=(16384,16384) x 2048 Total amount of constant memory: 65536 bytes Total amount of shared memory per block: 49152 bytes Total number of registers available per block: 32768 Warp size: 32 Maximum number of threads per multiprocessor: 1536 Maximum number of threads per block: 1024 Maximum sizes of each dimension of a block: 1024 x 1024 x 64 Maximum sizes of each dimension of a grid: 65535 x 65535 x 65535 Maximum memory pitch: 2147483647 bytes Texture alignment: 512 bytes Concurrent copy and kernel execution: Yes with 1 copy engine(s) Run time limit on kernels: Yes Integrated GPU sharing Host Memory: No Support host page-locked memory mapping: Yes Alignment requirement for Surfaces: Yes Device has ECC support: Disabled Device supports Unified Addressing (UVA): Yes Device PCI Bus ID / PCI location ID: 1 / 0 Compute Mode: < Default (multiple host threads can use ::cudaSetDevice() with device simultaneously) >
So as I was reading in a document (there are quite a few!) I have the maximum number of threads available: 1536. It's probably not that important for my test, but I was really wondering because the warp size is 32 and the number of cores is 96. All of that is just not well explained in the documentation (at least for what I read so far.) I actually see that I have 2 processors, 48 cores per processor.
So... if I understand properly I could run my alpha computation in 150 "cycles" on such a CUDA card (640x480 / 1024 = 300 which we can share on both processors.) Need to test that theory. That would probably be quite a bit faster than the SIMD implementation (which is already dead fast: about 3750 full frames per second.)
More soon. 8-)