CUDA编程札记

来源:互联网 发布:腾讯算法工程师待遇 编辑:程序博客网 时间:2024/06/05 01:54






const int N = 33 * 1024;const int threadsPerBlock = 256;const int blocksPerGrid =            imin( 32, (N+threadsPerBlock-1) / threadsPerBlock );__global__ void dot( float *a, float *b, float *c ) {    __shared__ float cache[threadsPerBlock];    int tid = threadIdx.x + blockIdx.x * blockDim.x;    int cacheIndex = threadIdx.x;    float   temp = 0;    while (tid < N) {        temp += a[tid] * b[tid];        tid += blockDim.x * gridDim.x;    }        // set the cache values    cache[cacheIndex] = temp;        // synchronize threads in this block    __syncthreads();    // for reductions, threadsPerBlock must be a power of 2    // because of the following code    int i = blockDim.x/2;    while (i != 0) {        if (cacheIndex < i)            cache[cacheIndex] += cache[cacheIndex + i];        __syncthreads();        i /= 2;    }    if (cacheIndex == 0)        c[blockIdx.x] = cache[0];}int main( void ) {    float   *a, *b, c, *partial_c;    float   *dev_a, *dev_b, *dev_partial_c;    // allocate memory on the cpu side    a = (float*)malloc( N*sizeof(float) );    b = (float*)malloc( N*sizeof(float) );    partial_c = (float*)malloc( blocksPerGrid*sizeof(float) );    // allocate the memory on the GPU    HANDLE_ERROR( cudaMalloc( (void**)&dev_a,                              N*sizeof(float) ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_b,                              N*sizeof(float) ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_partial_c,                              blocksPerGrid*sizeof(float) ) );    // fill in the host memory with data    for (int i=0; i<N; i++) {        a[i] = i;        b[i] = i*2;    }    // copy the arrays 'a' and 'b' to the GPU    HANDLE_ERROR( cudaMemcpy( dev_a, a, N*sizeof(float),                              cudaMemcpyHostToDevice ) );    HANDLE_ERROR( cudaMemcpy( dev_b, b, N*sizeof(float),                              cudaMemcpyHostToDevice ) );     dot<<<blocksPerGrid,threadsPerBlock>>>( dev_a, dev_b,                                            dev_partial_c );    // copy the array 'c' back from the GPU to the CPU    HANDLE_ERROR( cudaMemcpy( partial_c, dev_partial_c,                              blocksPerGrid*sizeof(float),                              cudaMemcpyDeviceToHost ) );    // finish up on the CPU side    c = 0;    for (int i=0; i<blocksPerGrid; i++) {        c += partial_c[i];    }    #define sum_squares(x)  (x*(x+1)*(2*x+1)/6)    printf( "Does GPU value %.6g = %.6g?\n", c,             2 * sum_squares( (float)(N - 1) ) );    // free memory on the gpu side    HANDLE_ERROR( cudaFree( dev_a ) );    HANDLE_ERROR( cudaFree( dev_b ) );    HANDLE_ERROR( cudaFree( dev_partial_c ) );    // free memory on the cpu side    free( a );    free( b );    free( partial_c );}


struct Lock {    int *mutex;    Lock( void ) {        HANDLE_ERROR( cudaMalloc( (void**)&mutex,sizeof(int) ) );        HANDLE_ERROR( cudaMemset( mutex, 0, sizeof(int) ) );    }    ~Lock( void ) {        cudaFree( mutex );    }    __device__ void lock( void ) {        while( atomicCAS( mutex, 0, 1 ) != 0 );    }    __device__ void unlock( void ) {        atomicExch( mutex, 0 );    }};

#define imin(a,b) (a<b?a:b)const int N = 33 * 1024 * 1024;const int threadsPerBlock = 256;const int blocksPerGrid =            imin( 32, (N+threadsPerBlock-1) / threadsPerBlock );__global__ void dot( Lock lock, float *a,                     float *b, float *c ) {    __shared__ float cache[threadsPerBlock];    int tid = threadIdx.x + blockIdx.x * blockDim.x;    int cacheIndex = threadIdx.x;    float   temp = 0;    while (tid < N) {        temp += a[tid] * b[tid];        tid += blockDim.x * gridDim.x;    }        // set the cache values    cache[cacheIndex] = temp;        // synchronize threads in this block    __syncthreads();    // for reductions, threadsPerBlock must be a power of 2    // because of the following code    int i = blockDim.x/2;    while (i != 0) {        if (cacheIndex < i)            cache[cacheIndex] += cache[cacheIndex + i];        __syncthreads();        i /= 2;    }    if (cacheIndex == 0) {        // wait until we get the lock        lock.lock();       // we have the lock at this point, update and release        *c += cache[0];        lock.unlock();    }}int main( void ) {    float   *a, *b, c = 0;    float   *dev_a, *dev_b, *dev_c;    // allocate memory on the cpu side    a = (float*)malloc( N*sizeof(float) );    b = (float*)malloc( N*sizeof(float) );    // allocate the memory on the GPU    HANDLE_ERROR( cudaMalloc( (void**)&dev_a,                              N*sizeof(float) ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_b,                              N*sizeof(float) ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_c,                              sizeof(float) ) );    // fill in the host memory with data    for (int i=0; i<N; i++) {        a[i] = i;        b[i] = i*2;    }    // copy the arrays 'a' and 'b' to the GPU    HANDLE_ERROR( cudaMemcpy( dev_a, a, N*sizeof(float),                              cudaMemcpyHostToDevice ) );    HANDLE_ERROR( cudaMemcpy( dev_b, b, N*sizeof(float),                              cudaMemcpyHostToDevice ) );     HANDLE_ERROR( cudaMemcpy( dev_c, &c, sizeof(float),                              cudaMemcpyHostToDevice ) );     Lock    lock;    dot<<<blocksPerGrid,threadsPerBlock>>>( lock, dev_a,                                            dev_b, dev_c );    // copy c back from the GPU to the CPU    HANDLE_ERROR( cudaMemcpy( &c, dev_c,                              sizeof(float),                              cudaMemcpyDeviceToHost ) );    #define sum_squares(x)  (x*(x+1)*(2*x+1)/6)    printf( "Does GPU value %.6g = %.6g?\n", c,             2 * sum_squares( (float)(N - 1) ) );    // free memory on the gpu side    HANDLE_ERROR( cudaFree( dev_a ) );    HANDLE_ERROR( cudaFree( dev_b ) );    HANDLE_ERROR( cudaFree( dev_c ) );    // free memory on the cpu side    free( a );    free( b );}

__global__ void histo_kernel( unsigned char *buffer,                              long size,                              unsigned int *histo ) {    // calculate the starting index and the offset to the next    // block that each thread will be processing    int i = threadIdx.x + blockIdx.x * blockDim.x;    int stride = blockDim.x * gridDim.x;    while (i < size) {        atomicAdd( &histo[buffer[i]], 1 );        i += stride;    }}int main( void ) {    unsigned char *buffer =                     (unsigned char*)big_random_block( SIZE );    // capture the start time    // starting the timer here so that we include the cost of    // all of the operations on the GPU.    cudaEvent_t     start, stop;    HANDLE_ERROR( cudaEventCreate( &start ) );    HANDLE_ERROR( cudaEventCreate( &stop ) );    HANDLE_ERROR( cudaEventRecord( start, 0 ) );    // allocate memory on the GPU for the file's data    unsigned char *dev_buffer;    unsigned int *dev_histo;    HANDLE_ERROR( cudaMalloc( (void**)&dev_buffer, SIZE ) );    HANDLE_ERROR( cudaMemcpy( dev_buffer, buffer, SIZE,                              cudaMemcpyHostToDevice ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_histo,                              256 * sizeof( int ) ) );    HANDLE_ERROR( cudaMemset( dev_histo, 0,                              256 * sizeof( int ) ) );    // kernel launch - 2x the number of mps gave best timing    cudaDeviceProp  prop;    HANDLE_ERROR( cudaGetDeviceProperties( &prop, 0 ) );    int blocks = prop.multiProcessorCount;    histo_kernel<<<blocks*2,256>>>( dev_buffer, SIZE, dev_histo );        unsigned int    histo[256];    HANDLE_ERROR( cudaMemcpy( histo, dev_histo,                              256 * sizeof( int ),                              cudaMemcpyDeviceToHost ) );    // get stop time, and display the timing results    HANDLE_ERROR( cudaEventRecord( stop, 0 ) );    HANDLE_ERROR( cudaEventSynchronize( stop ) );    float   elapsedTime;    HANDLE_ERROR( cudaEventElapsedTime( &elapsedTime,                                        start, stop ) );    printf( "Time to generate:  %3.1f ms\n", elapsedTime );    long histoCount = 0;    for (int i=0; i<256; i++) {        histoCount += histo[i];    }    printf( "Histogram Sum:  %ld\n", histoCount );    // verify that we have the same counts via CPU    for (int i=0; i<SIZE; i++)        histo[buffer[i]]--;    for (int i=0; i<256; i++) {        if (histo[i] != 0)            printf( "Failure at %d!  Off by %d\n", i, histo[i] );    }    HANDLE_ERROR( cudaEventDestroy( start ) );    HANDLE_ERROR( cudaEventDestroy( stop ) );    cudaFree( dev_histo );    cudaFree( dev_buffer );    free( buffer );    return 0;}


__global__ void histo_kernel( unsigned char *buffer,                              long size,                              unsigned int *histo ) {    // clear out the accumulation buffer called temp    // since we are launched with 256 threads, it is easy    // to clear that memory with one write per thread    __shared__  unsigned int temp[256];    temp[threadIdx.x] = 0;    __syncthreads();    // calculate the starting index and the offset to the next    // block that each thread will be processing    int i = threadIdx.x + blockIdx.x * blockDim.x;    int stride = blockDim.x * gridDim.x;    while (i < size) {        atomicAdd( &temp[buffer[i]], 1 );        i += stride;    }    // sync the data from the above writes to shared memory    // then add the shared memory values to the values from    // the other thread blocks using global memory    // atomic adds    // same as before, since we have 256 threads, updating the    // global histogram is just one write per thread!    __syncthreads();    atomicAdd( &(histo[threadIdx.x]), temp[threadIdx.x] );}int main( void ) {    unsigned char *buffer =                     (unsigned char*)big_random_block( SIZE );    // capture the start time    // starting the timer here so that we include the cost of    // all of the operations on the GPU.  if the data were    // already on the GPU and we just timed the kernel    // the timing would drop from 74 ms to 15 ms.  Very fast.    cudaEvent_t     start, stop;    HANDLE_ERROR( cudaEventCreate( &start ) );    HANDLE_ERROR( cudaEventCreate( &stop ) );    HANDLE_ERROR( cudaEventRecord( start, 0 ) );    // allocate memory on the GPU for the file's data    unsigned char *dev_buffer;    unsigned int *dev_histo;    HANDLE_ERROR( cudaMalloc( (void**)&dev_buffer, SIZE ) );    HANDLE_ERROR( cudaMemcpy( dev_buffer, buffer, SIZE,                              cudaMemcpyHostToDevice ) );    HANDLE_ERROR( cudaMalloc( (void**)&dev_histo,                              256 * sizeof( int ) ) );    HANDLE_ERROR( cudaMemset( dev_histo, 0,                              256 * sizeof( int ) ) );    // kernel launch - 2x the number of mps gave best timing    cudaDeviceProp  prop;    HANDLE_ERROR( cudaGetDeviceProperties( &prop, 0 ) );    int blocks = prop.multiProcessorCount;    histo_kernel<<<blocks*2,256>>>( dev_buffer,                                    SIZE, dev_histo );        unsigned int    histo[256];    HANDLE_ERROR( cudaMemcpy( histo, dev_histo,                              256 * sizeof( int ),                              cudaMemcpyDeviceToHost ) );    // get stop time, and display the timing results    HANDLE_ERROR( cudaEventRecord( stop, 0 ) );    HANDLE_ERROR( cudaEventSynchronize( stop ) );    float   elapsedTime;    HANDLE_ERROR( cudaEventElapsedTime( &elapsedTime,                                        start, stop ) );    printf( "Time to generate:  %3.1f ms\n", elapsedTime );    long histoCount = 0;    for (int i=0; i<256; i++) {        histoCount += histo[i];    }    printf( "Histogram Sum:  %ld\n", histoCount );    // verify that we have the same counts via CPU    for (int i=0; i<SIZE; i++)        histo[buffer[i]]--;    for (int i=0; i<256; i++) {        if (histo[i] != 0)            printf( "Failure at %d!\n", i );    }    HANDLE_ERROR( cudaEventDestroy( start ) );    HANDLE_ERROR( cudaEventDestroy( stop ) );    cudaFree( dev_histo );    cudaFree( dev_buffer );    free( buffer );    return 0;}


注:本文是作者对GPU高性能编程CUDA实战的学习总结。此书的代码可以在下面的链接下载,无需积分哦!

http://download.csdn.net/detail/celerychen2009/6360573


原创粉丝点击