Parallel Algorithms¶
Element-wise expression evalution (“map”)¶
Evaluating involved expressions on pyopencl.array.Array
instances by
using overloaded operators can be somewhat inefficient, because a new temporary
is created for each intermediate result. The functionality in the module
pyopencl.elementwise
contains tools to help generate kernels that
evaluate multi-stage expressions on one or several operands in a single pass.
-
class
pyopencl.elementwise.
ElementwiseKernel
(context, arguments, operation, name="kernel", preamble="", options=[])¶ A kernel that takes a number of scalar or vector arguments and performs an operation specified as a snippet of C on these arguments.
Parameters: - arguments – a string formatted as a C argument list.
- operation – a snippet of C that carries out the desired ‘map’
operation. The current index is available as the variable i.
operation may contain the statement
PYOPENCL_ELWISE_CONTINUE
, which will terminate processing for the current element. - name – the function name as which the kernel is compiled
- options – passed unmodified to
pyopencl.Program.build()
. - preamble – a piece of C source code that gets inserted outside of the function context in the elementwise operation’s kernel source code.
Warning
Using a return statement in operation will lead to incorrect results, as some elements may never get processed. Use
PYOPENCL_ELWISE_CONTINUE
instead.Changed in version 2013.1: Added
PYOPENCL_ELWISE_CONTINUE
.-
__call__
(*args, wait_for=None)¶ Invoke the generated scalar kernel. The arguments may either be scalars or
GPUArray
instances.Returns a new
pyopencl.Event
. wait_for may either be None or a list ofpyopencl.Event
instances for whose completion this command waits before starting exeuction.
Here’s a usage example:
.. literalinclude:: ../examples/demo_elementwise.py
(You can find this example as
examples/demo_elementwise.py
in the PyOpenCL distribution.)
Sums and counts (“reduce”)¶
-
class
pyopencl.reduction.
ReductionKernel
(ctx, dtype_out, neutral, reduce_expr, map_expr=None, arguments=None, name="reduce_kernel", options=[], preamble="")¶ Generate a kernel that takes a number of scalar or vector arguments (at least one vector argument), performs the map_expr on each entry of the vector argument and then the reduce_expr on the outcome of that. neutral serves as an initial value. preamble offers the possibility to add preprocessor directives and other code (such as helper functions) to be added before the actual reduction kernel code.
Vectors in map_expr should be indexed by the variable i. reduce_expr uses the formal values “a” and “b” to indicate two operands of a binary reduction operation. If you do not specify a map_expr,
in[i]
is automatically assumed and treated as the only one input argument.dtype_out specifies the
numpy.dtype
in which the reduction is performed and in which the result is returned. neutral is specified as float or integer formatted as string. reduce_expr and map_expr are specified as string formatted operations and arguments is specified as a string formatted as a C argument list. name specifies the name as which the kernel is compiled. options are passed unmodified topyopencl.Program.build()
. preamble specifies a string of code that is inserted before the actual kernels.-
__call__
(*args, queue=None, wait_for=None, return_event=False, out=None)¶ wait_for may either be None or a list of
pyopencl.Event
instances for whose completion this command waits before starting exeuction.With out the resulting single-entry
pyopencl.array.Array
can be specified. Because offsets are supported one can store results anywhere (e.g.out=a[3]
).Returns: the resulting scalar as a single-entry pyopencl.array.Array
if return_event is False, otherwise a tuple(scalar_array, event)
.Note
The returned
pyopencl.Event
corresponds only to part of the execution of the reduction. It is not suitable for profiling.
New in version 2011.1.
Changed in version 2014.2: Added out parameter.
-
Here’s a usage example:
a = pyopencl.array.arange(queue, 400, dtype=numpy.float32)
b = pyopencl.array.arange(queue, 400, dtype=numpy.float32)
krnl = ReductionKernel(ctx, numpy.float32, neutral="0",
reduce_expr="a+b", map_expr="x[i]*y[i]",
arguments="__global float *x, __global float *y")
my_dot_prod = krnl(a, b).get()
Prefix Sums (“scan”)¶
A prefix sum is a running sum of an array, as provided by
e.g. numpy.cumsum
:
>>> import numpy as np
>>> a = [1,1,1,1,1,2,2,2,2,2]
>>> np.cumsum(a)
array([ 1, 2, 3, 4, 5, 7, 9, 11, 13, 15])
This is a very simple example of what a scan can do. It turns out that scans are significantly more versatile. They are a basic building block of many non-trivial parallel algorithms. Many of the operations enabled by scans seem difficult to parallelize because of loop-carried dependencies.
See also
- Prefix sums and their applications, by Guy Blelloch.
- This article gives an overview of some surprising applications of scans.
- Simple / Legacy Interface
- These operations built into PyOpenCL are realized using
GenericScanKernel
.
Usage Example¶
This example illustrates the implementation of a simplified version of
pyopencl.algorithm.copy_if()
,
which copies integers from an array into the (variable-size) output if they are
greater than 300:
knl = GenericScanKernel(
ctx, np.int32,
arguments="__global int *ary, __global int *out",
input_expr="(ary[i] > 300) ? 1 : 0",
scan_expr="a+b", neutral="0",
output_statement="""
if (prev_item != item) out[item-1] = ary[i];
""")
out = a.copy()
knl(a, out)
a_host = a.get()
out_host = a_host[a_host > 300]
assert (out_host == out.get()[:len(out_host)]).all()
The value being scanned over is a number of flags indicating whether each array element is greater than 300. These flags are computed by input_expr. The prefix sum over this array gives a running count of array items greater than 300. The output_statement the compares prev_item (the previous item’s scan result, i.e. index) to item (the current item’s scan result, i.e. index). If they differ, i.e. if the predicate was satisfied at this position, then the item is stored in the output at the computed index.
This example does not make use of the following advanced features also available in PyOpenCL:
- Segmented scans
- Access to the previous item in input_expr (e.g. for comparisons)
See the implementation of
unique()
for an example.
Making Custom Scan Kernels¶
New in version 2013.1.
Debugging aids¶
-
class
pyopencl.scan.
GenericDebugScanKernel
¶ Performs the same function and has the same interface as
GenericScanKernel
, but uses a dead-simple, sequential scan. Works best on CPU platforms, and helps isolate bugs in scans by removing the potential for issues originating in parallel execution.
Simple / Legacy Interface¶
-
class
pyopencl.scan.
ExclusiveScanKernel
(ctx, dtype, scan_expr, neutral, name_prefix="scan", options=[], preamble="", devices=None)¶ Generates a kernel that can compute a prefix sum using any associative operation given as scan_expr. scan_expr uses the formal values “a” and “b” to indicate two operands of an associative binary operation. neutral is the neutral element of scan_expr, obeying scan_expr(a, neutral) == a.
dtype specifies the type of the arrays being operated on. name_prefix is used for kernel names to ensure recognizability in profiles and logs. options is a list of compiler options to use when building. preamble specifies a string of code that is inserted before the actual kernels. devices may be used to restrict the set of devices on which the kernel is meant to run. (defaults to all devices in the context ctx.
-
__call__
(self, input_ary, output_ary=None, allocator=None, queue=None)¶
-
-
class
pyopencl.scan.
InclusiveScanKernel
(ctx, dtype, scan_expr, neutral=None, name_prefix="scan", options=[], preamble="", devices=None)¶ Works like
ExclusiveScanKernel
.Changed in version 2013.1: neutral is now always required.
For the array [1,2,3], inclusive scan results in [1,3,6], and exclusive scan results in [0,1,3].
Here’s a usage example:
knl = InclusiveScanKernel(context, np.int32, "a+b")
n = 2**20-2**18+5
host_data = np.random.randint(0, 10, n).astype(np.int32)
dev_data = cl_array.to_device(queue, host_data)
knl(dev_data)
assert (dev_data.get() == np.cumsum(host_data, axis=0)).all()
Predicated copies (“partition”, “unique”, …)¶
Sorting (radix sort)¶
Building many variable-size lists¶
Bitonic Sort¶
-
class
pyopencl.bitonic_sort.
BitonicSort
(context)¶ Sort an array (or one axis of one) using a sorting network.
Will only work if the axis of the array to be sorted has a length that is a power of 2.
New in version 2015.2.
See also
pyopencl.algorithm.RadixSort
-
__call__
(arr, idx=None, queue=None, wait_for=None, axis=0)¶ Parameters: - arr – the array to be sorted. Will be overwritten with the sorted array.
- idx – an array of indices to be tracked along with the sorting of arr
- queue – a
pyopencl.CommandQueue
, defaults to the array’s queue if None - wait_for – a list of
pyopencl.Event
instances or None - axis – the axis of the array by which to sort
Returns: a tuple (sorted_array, event)
-