Related
Matthew Sainsbury If for some reason a Linux system fails to boot, the following message is usually seen: kernel panic - not syncing: [Error Message]
But what does "out of sync" mean? What exactly is out of sync? The only place I've seen synchronization befor
Matthew Sainsbury If for some reason a Linux system fails to boot, the following message is usually seen: kernel panic - not syncing: [Error Message]
But what does "out of sync" mean? What exactly is out of sync? The only place I've seen synchronization befor
Abx I have to add two square matrices N x Nusing cuda program . This book requires that configuration parameters for the kernel be written for the following situations: (a) Each thread can only process 1matrix elements (b) Each thread produces one output matri
username I don't know what's wrong, I want to add two vectors together but I get an error. I looked at some tutorials but didn't see the problem, here is my error: Argument of type 'int**' is incompatible with parameter of type 'int*' #include "cuda_runtime.h"
Abx I have to add two square matrices N x Nusing cuda program . This book requires that configuration parameters for the kernel be written for the following situations: (a) Each thread can only process 1matrix elements (b) Each thread produces one output matri
username I don't know what's wrong, I want to add two vectors together but I get an error. I looked at some tutorials but didn't see the problem, here is my error: Argument of type 'int**' is incompatible with parameter of type 'int*' #include "cuda_runtime.h"
Decker I'm trying to analyze some code I've found online and keep thinking I'm stuck. I am looking at a histogram kernel launched with the following parameters histogram<<<2500, numBins, numBins * sizeof(unsigned int)>>>(...);
I know the parameters are grid,
David CUDA kernels are launched using this syntax (at least in the runtime API) mykernel<<<blocks, threads, shared_mem, stream>>>(args);
Is this implemented as a macro or is there a special syntax nvcc removes before handing the host code to gcc? Robert Clovi
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Dirk I'm trying to analyze some code I've found online and keep thinking I'm stuck. I am looking at a histogram kernel launched with the following parameters histogram<<<2500, numBins, numBins * sizeof(unsigned int)>>>(...);
I know the parameters are grid, b
Decker I'm trying to analyze some code I've found online and keep thinking I'm stuck. I am looking at a histogram kernel launched with the following parameters histogram<<<2500, numBins, numBins * sizeof(unsigned int)>>>(...);
I know the parameters are grid,
David CUDA kernels are launched using this syntax (at least in the runtime API) mykernel<<<blocks, threads, shared_mem, stream>>>(args);
Is this implemented as a macro or is there a special syntax nvcc removes before handing the host code to gcc? Robert Clovi
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Decker I'm trying to analyze some code I've found online and keep thinking I'm stuck. I am looking at a histogram kernel launched with the following parameters histogram<<<2500, numBins, numBins * sizeof(unsigned int)>>>(...);
I know the parameters are grid,
David CUDA kernels are launched using this syntax (at least in the runtime API) mykernel<<<blocks, threads, shared_mem, stream>>>(args);
Is this implemented as a macro or is there a special syntax nvcc removes before handing the host code to gcc? Robert Clovi
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
Tyson Hilmer I use the checkCudaErrors helper function from the CUDA toolkit examples. See "helper_cuda.h". I'm confused as to why the startup errors in this example are not caught by checkCudaErrors. The error is too many threads started (2048). From Debug (L
elastic band I am new to Cuda programming and am implementing the classic Floyd APSP algorithm. The algorithm consists of 3 nested loops, and all code in the two inner loops can be executed in parallel. As the main part of my code, this is the kernel code: __g
elastic band I am new to Cuda programming and am implementing the classic Floyd APSP algorithm. The algorithm consists of 3 nested loops, and all code in the two inner loops can be executed in parallel. As the main part of my code, this is the kernel code: __g
elastic band I am new to Cuda programming and am implementing the classic Floyd APSP algorithm. The algorithm consists of 3 nested loops, and all code in the two inner loops can be executed in parallel. As the main part of my code, this is the kernel code: __g
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa
Novitor Can you help me understand how to write a CUDA kernel in Python? AFAIK, numba.vectorize can be executed on cuda, cpu, parallel (multi-cpus) based on target . But target='cuda' requires the CUDA kernel to be set. The main problem is that many of the exa