Basics of CuPy¶
In this section, you will learn about the following things:
The concept of current device
host-device and device-device array transfer
Basics of cupy.ndarray¶
CuPy is a GPU array backend that implements a subset of NumPy interface.
In the following code,
cp is an abbreviation of
cupy, following the convention of abbreviating
>>> import numpy as np >>> import cupy as cp
>>> x_gpu = cp.array([1, 2, 3])
x_gpu in the above example is an instance of
You can see its creation of identical to
NumPy’s one, except that
numpy is replaced with
The main difference of
numpy.ndarray is that the content is allocated on the device memory.
Its data is allocated on the current device, which will be explained later.
Most of the array manipulations are also done in the way similar to NumPy.
Take the Euclidean norm (a.k.a L2 norm) for example.
numpy.linalg.norm() to calculate it on CPU.
>>> x_cpu = np.array([1, 2, 3]) >>> l2_cpu = np.linalg.norm(x_cpu)
We can calculate it on GPU with CuPy in a similar way:
>>> x_gpu = cp.array([1, 2, 3]) >>> l2_gpu = cp.linalg.norm(x_gpu)
CuPy implements many functions on
See the reference for the supported subset of NumPy API.
Understanding NumPy might help utilizing most features of CuPy.
So, we recommend you to read the NumPy documentation.
CuPy has a concept of current devices, which is the default device on which the allocation, manipulation, calculation, etc., of arrays are taken place. Suppose the ID of current device is 0. The following code allocates array contents on GPU 0.
>>> x_on_gpu0 = cp.array([1, 2, 3, 4, 5])
The current device can be changed by
cupy.cuda.Device.use() as follows:
>>> x_on_gpu0 = cp.array([1, 2, 3, 4, 5]) >>> cp.cuda.Device(1).use() >>> x_on_gpu1 = cp.array([1, 2, 3, 4, 5])
If you switch the current GPU temporarily, with statement comes in handy.
>>> with cp.cuda.Device(1): ... x_on_gpu1 = cp.array([1, 2, 3, 4, 5]) >>> x_on_gpu0 = cp.array([1, 2, 3, 4, 5])
Most operations of CuPy are done on the current device. Be careful that if processing of an array on a non-current device will cause an error:
>>> with cp.cuda.Device(0): ... x_on_gpu0 = cp.array([1, 2, 3, 4, 5]) >>> with cp.cuda.Device(1): ... x_on_gpu0 * 2 # raises error Traceback (most recent call last): ... ValueError: Array device must be same as the current device: array device = 0 while current = 1
cupy.ndarray.device attribute indicates the device on which the array is allocated.
>>> with cp.cuda.Device(1): ... x = cp.array([1, 2, 3, 4, 5]) >>> x.device <CUDA Device 1>
If the environment has only one device, such explicit device switching is not needed.
Associated with the concept of current devices are current streams, which help avoid explicitly passing streams in every single operation so as to keep the APIs pythonic and user-friendly. In CuPy, any CUDA operations such as data transfer (see the next section) and kernel launches are enqueued onto the current stream, and the queued tasks on the same stream will be executed in serial (but asynchronously with respect to the host).
The default current stream in CuPy is CUDA’s null stream (i.e., stream 0). It is also known as the legacy
default stream, which is unique per device. However, it is possible to change the current stream using the
cupy.cuda.Stream API, please see Accessing CUDA Functionalities for example. The current stream in CuPy can be
It is worth noting that CuPy’s current stream is managed on a per thread basis, meaning that on different Python threads the current stream (if not the null stream) can be different.
Move arrays to a device¶
>>> x_cpu = np.array([1, 2, 3]) >>> x_gpu = cp.asarray(x_cpu) # move the data to the current device.
>>> with cp.cuda.Device(0): ... x_gpu_0 = cp.ndarray([1, 2, 3]) # create an array in GPU 0 >>> with cp.cuda.Device(1): ... x_gpu_1 = cp.asarray(x_gpu_0) # move the array to GPU 1
cupy.asarray() does not copy the input array if possible.
So, if you put an array of the current device, it returns the input object itself.
Move array from a device to the host¶
Moving a device array to the host can be done by
cupy.asnumpy() as follows:
>>> x_gpu = cp.array([1, 2, 3]) # create an array in the current device >>> x_cpu = cp.asnumpy(x_gpu) # move the array to the host.
We can also use
>>> x_cpu = x_gpu.get()
Check Memory Management for a detailed description of how is memory managed in CuPy using memory pools.
How to write CPU/GPU agnostic code¶
The compatibility of CuPy with NumPy enables us to write CPU/GPU generic code.
It can be made easy by the
This function returns the
cupy module based on arguments.
A CPU/GPU generic function is defined using it like follows:
>>> # Stable implementation of log(1 + exp(x)) >>> def softplus(x): ... xp = cp.get_array_module(x) ... return xp.maximum(0, x) + xp.log1p(xp.exp(-abs(x)))
Sometimes, an explicit conversion to a host or device array may be required.
cupy.asnumpy() can be used in agnostic implementations
to get host or device arrays from either CuPy or NumPy arrays.
>>> y_cpu = np.array([4, 5, 6]) >>> x_cpu + y_cpu array([5, 7, 9]) >>> x_gpu + y_cpu Traceback (most recent call last): ... TypeError: Unsupported type <class 'numpy.ndarray'> >>> cp.asnumpy(x_gpu) + y_cpu array([5, 7, 9]) >>> cp.asnumpy(x_gpu) + cp.asnumpy(y_cpu) array([5, 7, 9]) >>> x_gpu + cp.asarray(y_cpu) array([5, 7, 9]) >>> cp.asarray(x_gpu) + cp.asarray(y_cpu) array([5, 7, 9])