1.使用thrust:
http://stackoverflow.com/questions/13185221/cuda-host-object-to-device
http://blog.csdn.net/shenlan282/article/details/8237576
http://blog.csdn.net/shenlan282/article/details/8237586
2.使用条件编译
http://stackoverflow.com/questions/6978643/cuda-and-classes
Define the class in a header that you #include, just like in C++.
Any method that must be called from device code should be defined with both __device__
and__host__
declspecs, including the constructor and destructor if you plan to use new
/delete
on the device (note new
/delete
require CUDA 4.0 and a compute capability 2.0 or higher GPU).
You probably want to define a macro like
#ifdef __CUDACC__ #define CUDA_CALLABLE_MEMBER __host__ __device__ #else #define CUDA_CALLABLE_MEMBER #endif
Then use this macro on your member functions
class Foo { public: CUDA_CALLABLE_MEMBER Foo() {} CUDA_CALLABLE_MEMBER ~Foo() {} CUDA_CALLABLE_MEMBER void aMethod() {} };
The reason for this is that only the CUDA compiler knows __device__
and __host__
-- your host C++ compiler will raise an error.