Difference between revisions of "CUDA examples"

From ICPWiki
Jump to: navigation, search
m (Scalar product)
(Scalar product)
Line 12: Line 12:
  
 
{{Download|ScalarGPU.tar.gz|Code example scalar product|tgz}}
 
{{Download|ScalarGPU.tar.gz|Code example scalar product|tgz}}
 +
 +
[http://mathema.tician.de/software/pycuda PyCUDA] is an extremely powerful Python extension that does not only allow to use CUDA code from Python, but can do just-in-time kernel compilation for you, and allows to write code similiar to numpy, just that it will be executed on a GPU - and much faster therefore. This is an example code calculating again the scalar product, just in Python. Thanks to PyCUDA it is as fast as the plain CUDA code (or even faster, using GPUArray...).
 +
 +
{{Download|PyScalar.tar.gz|Code example scalar product - in Python|tgz}}

Revision as of 19:20, 4 February 2011

Gram-Schmidt method

This example implements the Gram-Schmidt orthogonalization method for fully occupied matrices. While for sparse matrices, the Householder method is typically more efficient, the Gram-Schmidt method is still widely used for fully occupied matrices due to its robustness. And due to its structure, the method can be easily implemented on GPUs using CUDA.

tgz.pngCode example Gram-Schmidt method (23 KB)Info circle.png

Scalar product

Due to its very small computational cost, the scalar product is certainly 'not' suited for porting to a GPU. However, if a complex algorithm requires a scalar product and the algorithm is ported to a GPU, it is necessary to also port the scalar product to the GPU. Otherwise, one would need to transfer data back to the CPU, which is even more time consuming.

This example demonstrates several possible strategies. It also serves as a good example for the use of atomic operations, and how they can be avoided. e.g. on older hardware. Some of the examples deliberatedly don't work to show the pitfalls of massively parallel computation, and to demonstrate that atomic operations are really necessary. Note that the use of atomic operations is rather slow, and therefore in this example no speed up is gained by using them. However, the use of atomic operations makes the code considerably shorter and easier to read, which is important in a scientific environment, where code is continously developed further.

tgz.pngCode example scalar product (17 KB)Info circle.png

PyCUDA is an extremely powerful Python extension that does not only allow to use CUDA code from Python, but can do just-in-time kernel compilation for you, and allows to write code similiar to numpy, just that it will be executed on a GPU - and much faster therefore. This is an example code calculating again the scalar product, just in Python. Thanks to PyCUDA it is as fast as the plain CUDA code (or even faster, using GPUArray...).

tgz.pngCode example scalar product - in Python (15 KB)Info circle.png