mxlib
c++ tools for analyzing astronomical data and other tasks by Jared R. Males. [git repo]
Loading...
Searching...
No Matches
cudaPtr.hpp
Go to the documentation of this file.
1/** \file cudaPtr.hpp
2 * \author Jared R. Males
3 * \brief A wrapper for cuda device pointers
4 * \ingroup cuda_files
5 *
6 */
7
8//***********************************************************************//
9// Copyright 2019,2020 Jared R. Males (jaredmales@gmail.com)
10//
11// This file is part of mxlib.
12//
13// mxlib is free software: you can redistribute it and/or modify
14// it under the terms of the GNU General Public License as published by
15// the Free Software Foundation, either version 3 of the License, or
16// (at your option) any later version.
17//
18// mxlib is distributed in the hope that it will be useful,
19// but WITHOUT ANY WARRANTY; without even the implied warranty of
20// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
21// GNU General Public License for more details.
22//
23// You should have received a copy of the GNU General Public License
24// along with mxlib. If not, see <http://www.gnu.org/licenses/>.
25//***********************************************************************//
26
27#ifndef math_cudaPtr_hpp
28#define math_cudaPtr_hpp
29
30#include <iostream>
31
32#include <cuda_runtime.h>
33
34#include "templateCuda.hpp"
35
36namespace mx
37{
38namespace cuda
39{
40
41/// A smart-pointer wrapper for cuda device pointers.
42/**
43 * \ingroup cuda
44 */
45template <typename T>
46struct cudaPtr
47{
48 /// The host data type.
49 // typedef typename cudaType<T>::hostType hostPtrT;
50 typedef T hostPtrT;
51
52 /// The device data type
53 // typedef typename cudaType<T>::deviceType devicePtrT;
54 typedef T devicePtrT;
55
56 /// The device pointer
58
59 /// The allocated size
60 size_t m_size{ 0 };
61
62 /// Destructor, frees memory if allocated.
64
65 size_t size()
66 {
67 return m_size;
68 }
69
70 /// Resize the memory allocation, in 1D
71 /** If no size change, this is a no-op.
72 *
73 * \returns 0 on success.
74 * \returns a cuda error code otherwise.
75 *
76 * \test Scenario: scaling a vector with cublas \ref test_math_templateCublas_scal "[test doc]"
77 * \test Scenario: scaling and accumulating a vector with cublas \ref test_math_templateCublas_axpy "[test doc]"
78 * \test Scenario: multiplying two vectors element by element \ref test_math_templateCublas_elementwiseXxY "[test
79 * doc]"
80 */
81 int resize( size_t sz /**< [in] the new size */ );
82
83 /// Resize the memory allocation, in 2D
84 /** If no size change, this is a no-op.
85 *
86 * \returns 0 on success.
87 * \returns a cuda error code otherwise.
88 *
89 */
90 int resize( size_t x_sz, ///< [in] the new x size,
91 size_t y_sz ///< [in] the new y size
92 );
93
94 /// Resize the memory allocation, in 3D
95 /** If no size change, this is a no-op.
96 *
97 * \returns 0 on success.
98 * \returns a cuda error code otherwise.
99 *
100 */
101 int resize( size_t x_sz, ///< [in] the new x size,
102 size_t y_sz, ///< [in] the new y size,
103 size_t z_sz ///< [in] the new z size
104 );
105
106 /// Initialize the array bytes to 0.
107 /** Just a wrapper to cudaMemset.
108 *
109 */
110 cudaError_t initialize();
111
112 /// Free the memory allocation
113 /**
114 * \returns 0 on success.
115 * \returns a cuda error code otherwise.
116 *
117 */
118 int free();
119
120 /// Copy from the host to the device, after allocation.
121 /**
122 * The device pointer must be allocated.
123 *
124 * \returns 0 on success.
125 * \returns a cuda error code otherwise.
126 *
127 * \test Scenario: multiplying two vectors element by element \ref test_math_templateCublas_elementwiseXxY "[test
128 * doc]"
129 */
130 int upload( const hostPtrT *src /**< [in] The host location */ );
131
132 /// Copy from the host to the device with allocation.
133 /**
134 * The device pointer will be re-allocated as needed.
135 *
136 * \returns 0 on success.
137 * \returns a cuda error code otherwise.
138 *
139 * \test Scenario: scaling a vector with cublas \ref test_math_templateCublas_scal "[test doc]"
140 * \test Scenario: scaling and accumulating a vector with cublas \ref test_math_templateCublas_axpy "[test doc]"
141 * \test Scenario: multiplying two vectors element by element \ref test_math_templateCublas_elementwiseXxY "[test
142 * doc]"
143 */
144 int upload( const hostPtrT *src, ///< [in] The host location
145 size_t sz ///< [in] The size of the array
146 );
147
148 /// Copy from the device to the host.
149 /**
150 *
151 * \test Scenario: scaling a vector with cublas \ref test_math_templateCublas_scal "[test doc]"
152 * \test Scenario: scaling and accumulating a vector with cublas \ref test_math_templateCublas_axpy "[test doc]"
153 * \test Scenario: multiplying two vectors element by element \ref test_math_templateCublas_elementwiseXxY "[test
154 * doc]"
155 */
156 int download( hostPtrT *dest /**< [in] The host location, allocated.*/ );
157
158 /// Conversion operator, accesses the device pointer for use in Cuda functions.
159 /**
160 * \test Scenario: scaling and accumulating a vector with cublas \ref test_math_templateCublas_axpy "[test doc]"
161 * \test Scenario: multiplying two vectors element by element \ref test_math_templateCublas_elementwiseXxY "[test
162 * doc]"
163 */
164 typename cpp2cudaType<devicePtrT>::cudaType *operator()()
165 {
166 return (typename cpp2cudaType<devicePtrT>::cudaType *)m_devicePtr;
167 }
168
169 /// Conversion operator, accesses the device pointer for use in Cuda functions.
170 /**
171 *
172 * \test Scenario: scaling a vector with cublas \ref test_math_templateCublas_scal "[test doc]"
173 */
174 const typename cpp2cudaType<devicePtrT>::cudaType *operator()() const
175 {
176 return (typename cpp2cudaType<devicePtrT>::cudaType *)m_devicePtr;
177 }
178};
179
180template <typename T>
182{
183 free();
184}
185
186template <typename T>
187int cudaPtr<T>::resize( size_t sz )
188{
189 if( m_size == sz )
190 return 0;
191
192 m_size = sz;
193
194 cudaError_t rv = cudaMalloc( (void **)&m_devicePtr, sz * sizeof( devicePtrT ) );
195
196 if( rv != cudaSuccess )
197 {
198 std::cerr << "Error from cudaMalloc: ";
199 printf( "[%s] %s\n", cudaGetErrorName( rv ), cudaGetErrorString( rv ) );
200 }
201
202 return 0;
203}
204
205template <typename T>
206int cudaPtr<T>::resize( size_t x_sz, size_t y_sz )
207{
208 return resize( x_sz * y_sz );
209}
210
211template <typename T>
212int cudaPtr<T>::resize( size_t x_sz, size_t y_sz, size_t z_sz )
213{
214 return resize( x_sz * y_sz * z_sz );
215}
216
217template <typename T>
219{
220 return ::cudaMemset( m_devicePtr, 0, m_size * sizeof( devicePtrT ) );
221}
222
223template <typename T>
225{
226 if( m_devicePtr )
227 {
228 int rv = cudaFree( m_devicePtr );
229
230 if( rv != cudaSuccess )
231 {
232 std::cerr << "Cuda Free Error \n";
233 return rv;
234 }
235 }
236
237 m_devicePtr = 0;
238 m_size = 0;
239
240 return 0;
241}
242
243template <typename T>
245{
246 // Copy host memory to device
247 int rv = cudaMemcpy( m_devicePtr, src, m_size * sizeof( devicePtrT ), cudaMemcpyHostToDevice );
248
249 if( rv != cudaSuccess )
250 {
251 std::cerr << "Cuda Memcpy error \n";
252 return rv;
253 }
254
255 return 0;
256}
257
258template <typename T>
259int cudaPtr<T>::upload( const hostPtrT *src, size_t sz )
260{
261 int rv;
262
263 rv = resize( sz );
264
265 if( rv )
266 return rv;
267
268 return upload( src );
269}
270
271template <typename T>
273{
274 // Copy device memory to host
275 int rv = cudaMemcpy( dest, m_devicePtr, m_size * sizeof( devicePtrT ), cudaMemcpyDeviceToHost );
276
277 if( rv != cudaSuccess )
278 {
279 std::cerr << "Cuda Memcpy error \n";
280 return rv;
281 }
282
283 return 0;
284}
285
286} // namespace cuda
287} // namespace mx
288#endif // math_cudaPtr_hpp
The mxlib c++ namespace.
Definition mxError.hpp:106
A smart-pointer wrapper for cuda device pointers.
Definition cudaPtr.hpp:47
const cpp2cudaType< devicePtrT >::cudaType * operator()() const
Conversion operator, accesses the device pointer for use in Cuda functions.
Definition cudaPtr.hpp:174
int resize(size_t x_sz, size_t y_sz)
Resize the memory allocation, in 2D.
Definition cudaPtr.hpp:206
cudaError_t initialize()
Initialize the array bytes to 0.
Definition cudaPtr.hpp:218
T devicePtrT
The device data type.
Definition cudaPtr.hpp:54
size_t m_size
The allocated size.
Definition cudaPtr.hpp:60
int download(hostPtrT *dest)
Copy from the device to the host.
Definition cudaPtr.hpp:272
int upload(const hostPtrT *src)
Copy from the host to the device, after allocation.
Definition cudaPtr.hpp:244
devicePtrT * m_devicePtr
The device pointer.
Definition cudaPtr.hpp:57
int free()
Free the memory allocation.
Definition cudaPtr.hpp:224
T hostPtrT
The host data type.
Definition cudaPtr.hpp:50
int resize(size_t x_sz, size_t y_sz, size_t z_sz)
Resize the memory allocation, in 3D.
Definition cudaPtr.hpp:212
int upload(const hostPtrT *src, size_t sz)
Copy from the host to the device with allocation.
Definition cudaPtr.hpp:259
cpp2cudaType< devicePtrT >::cudaType * operator()()
Conversion operator, accesses the device pointer for use in Cuda functions.
Definition cudaPtr.hpp:164
int resize(size_t sz)
Resize the memory allocation, in 1D.
Definition cudaPtr.hpp:187
~cudaPtr()
Destructor, frees memory if allocated.
Definition cudaPtr.hpp:181
Utilities for a template interface to cuda.