|
@@ -0,0 +1,10220 @@
|
|
|
+/*******************************************************************************
|
|
|
+ * Copyright (c) 2008-2016 The Khronos Group Inc.
|
|
|
+ *
|
|
|
+ * Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
+ * copy of this software and/or associated documentation files (the
|
|
|
+ * "Materials"), to deal in the Materials without restriction, including
|
|
|
+ * without limitation the rights to use, copy, modify, merge, publish,
|
|
|
+ * distribute, sublicense, and/or sell copies of the Materials, and to
|
|
|
+ * permit persons to whom the Materials are furnished to do so, subject to
|
|
|
+ * the following conditions:
|
|
|
+ *
|
|
|
+ * The above copyright notice and this permission notice shall be included
|
|
|
+ * in all copies or substantial portions of the Materials.
|
|
|
+ *
|
|
|
+ * MODIFICATIONS TO THIS FILE MAY MEAN IT NO LONGER ACCURATELY REFLECTS
|
|
|
+ * KHRONOS STANDARDS. THE UNMODIFIED, NORMATIVE VERSIONS OF KHRONOS
|
|
|
+ * SPECIFICATIONS AND HEADER INFORMATION ARE LOCATED AT
|
|
|
+ * https://www.khronos.org/registry/
|
|
|
+ *
|
|
|
+ * THE MATERIALS ARE PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
|
+ * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
|
+ * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
|
|
|
+ * IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
|
|
|
+ * CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
|
|
|
+ * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
|
|
|
+ * MATERIALS OR THE USE OR OTHER DEALINGS IN THE MATERIALS.
|
|
|
+ ******************************************************************************/
|
|
|
+
|
|
|
+/*! \file
|
|
|
+ *
|
|
|
+ * \brief C++ bindings for OpenCL 1.0 (rev 48), OpenCL 1.1 (rev 33),
|
|
|
+ * OpenCL 1.2 (rev 15), OpenCL 2.0 (rev 29), OpenCL 2.1 (rev 17),
|
|
|
+ * and OpenCL 2.2 (V2.2-11).
|
|
|
+ * \author Lee Howes and Bruce Merry
|
|
|
+ *
|
|
|
+ * Derived from the OpenCL 1.x C++ bindings written by
|
|
|
+ * Benedict R. Gaster, Laurent Morichetti and Lee Howes
|
|
|
+ * With additions and fixes from:
|
|
|
+ * Brian Cole, March 3rd 2010 and April 2012
|
|
|
+ * Matt Gruenke, April 2012.
|
|
|
+ * Bruce Merry, February 2013.
|
|
|
+ * Tom Deakin and Simon McIntosh-Smith, July 2013
|
|
|
+ * James Price, 2015-
|
|
|
+ * \version 2.2.0
|
|
|
+ * \date 2019-09-18
|
|
|
+ *
|
|
|
+ * Optional extension support
|
|
|
+ *
|
|
|
+ * cl_ext_device_fission
|
|
|
+ * #define CL_HPP_USE_CL_DEVICE_FISSION
|
|
|
+ * cl_khr_d3d10_sharing
|
|
|
+ * #define CL_HPP_USE_DX_INTEROP
|
|
|
+ * cl_khr_sub_groups
|
|
|
+ * #define CL_HPP_USE_CL_SUB_GROUPS_KHR
|
|
|
+ * cl_khr_image2d_from_buffer
|
|
|
+ * #define CL_HPP_USE_CL_IMAGE2D_FROM_BUFFER_KHR
|
|
|
+ *
|
|
|
+ * Doxygen documentation for this header is available here:
|
|
|
+ *
|
|
|
+ * http://khronosgroup.github.io/OpenCL-CLHPP/
|
|
|
+ *
|
|
|
+ * The latest version of this header can be found on the GitHub releases page:
|
|
|
+ *
|
|
|
+ * https://github.com/KhronosGroup/OpenCL-CLHPP/releases
|
|
|
+ *
|
|
|
+ * Bugs and patches can be submitted to the GitHub repository:
|
|
|
+ *
|
|
|
+ * https://github.com/KhronosGroup/OpenCL-CLHPP
|
|
|
+ */
|
|
|
+
|
|
|
+/*! \mainpage
|
|
|
+ * \section intro Introduction
|
|
|
+ * For many large applications C++ is the language of choice and so it seems
|
|
|
+ * reasonable to define C++ bindings for OpenCL.
|
|
|
+ *
|
|
|
+ * The interface is contained with a single C++ header file \em cl2.hpp and all
|
|
|
+ * definitions are contained within the namespace \em cl. There is no additional
|
|
|
+ * requirement to include \em cl.h and to use either the C++ or original C
|
|
|
+ * bindings; it is enough to simply include \em cl2.hpp.
|
|
|
+ *
|
|
|
+ * The bindings themselves are lightweight and correspond closely to the
|
|
|
+ * underlying C API. Using the C++ bindings introduces no additional execution
|
|
|
+ * overhead.
|
|
|
+ *
|
|
|
+ * There are numerous compatibility, portability and memory management
|
|
|
+ * fixes in the new header as well as additional OpenCL 2.0 features.
|
|
|
+ * As a result the header is not directly backward compatible and for this
|
|
|
+ * reason we release it as cl2.hpp rather than a new version of cl.hpp.
|
|
|
+ *
|
|
|
+ *
|
|
|
+ * \section compatibility Compatibility
|
|
|
+ * Due to the evolution of the underlying OpenCL API the 2.0 C++ bindings
|
|
|
+ * include an updated approach to defining supported feature versions
|
|
|
+ * and the range of valid underlying OpenCL runtime versions supported.
|
|
|
+ *
|
|
|
+ * The combination of preprocessor macros CL_HPP_TARGET_OPENCL_VERSION and
|
|
|
+ * CL_HPP_MINIMUM_OPENCL_VERSION control this range. These are three digit
|
|
|
+ * decimal values representing OpenCL runime versions. The default for
|
|
|
+ * the target is 200, representing OpenCL 2.0 and the minimum is also
|
|
|
+ * defined as 200. These settings would use 2.0 API calls only.
|
|
|
+ * If backward compatibility with a 1.2 runtime is required, the minimum
|
|
|
+ * version may be set to 120.
|
|
|
+ *
|
|
|
+ * Note that this is a compile-time setting, and so affects linking against
|
|
|
+ * a particular SDK version rather than the versioning of the loaded runtime.
|
|
|
+ *
|
|
|
+ * The earlier versions of the header included basic vector and string
|
|
|
+ * classes based loosely on STL versions. These were difficult to
|
|
|
+ * maintain and very rarely used. For the 2.0 header we now assume
|
|
|
+ * the presence of the standard library unless requested otherwise.
|
|
|
+ * We use std::array, std::vector, std::shared_ptr and std::string
|
|
|
+ * throughout to safely manage memory and reduce the chance of a
|
|
|
+ * recurrance of earlier memory management bugs.
|
|
|
+ *
|
|
|
+ * These classes are used through typedefs in the cl namespace:
|
|
|
+ * cl::array, cl::vector, cl::pointer and cl::string.
|
|
|
+ * In addition cl::allocate_pointer forwards to std::allocate_shared
|
|
|
+ * by default.
|
|
|
+ * In all cases these standard library classes can be replaced with
|
|
|
+ * custom interface-compatible versions using the CL_HPP_NO_STD_ARRAY,
|
|
|
+ * CL_HPP_NO_STD_VECTOR, CL_HPP_NO_STD_UNIQUE_PTR and
|
|
|
+ * CL_HPP_NO_STD_STRING macros.
|
|
|
+ *
|
|
|
+ * The OpenCL 1.x versions of the C++ bindings included a size_t wrapper
|
|
|
+ * class to interface with kernel enqueue. This caused unpleasant interactions
|
|
|
+ * with the standard size_t declaration and led to namespacing bugs.
|
|
|
+ * In the 2.0 version we have replaced this with a std::array-based interface.
|
|
|
+ * However, the old behaviour can be regained for backward compatibility
|
|
|
+ * using the CL_HPP_ENABLE_SIZE_T_COMPATIBILITY macro.
|
|
|
+ *
|
|
|
+ * Finally, the program construction interface used a clumsy vector-of-pairs
|
|
|
+ * design in the earlier versions. We have replaced that with a cleaner
|
|
|
+ * vector-of-vectors and vector-of-strings design. However, for backward
|
|
|
+ * compatibility old behaviour can be regained with the
|
|
|
+ * CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY macro.
|
|
|
+ *
|
|
|
+ * In OpenCL 2.0 OpenCL C is not entirely backward compatibility with
|
|
|
+ * earlier versions. As a result a flag must be passed to the OpenCL C
|
|
|
+ * compiled to request OpenCL 2.0 compilation of kernels with 1.2 as
|
|
|
+ * the default in the absence of the flag.
|
|
|
+ * In some cases the C++ bindings automatically compile code for ease.
|
|
|
+ * For those cases the compilation defaults to OpenCL C 2.0.
|
|
|
+ * If this is not wanted, the CL_HPP_CL_1_2_DEFAULT_BUILD macro may
|
|
|
+ * be specified to assume 1.2 compilation.
|
|
|
+ * If more fine-grained decisions on a per-kernel bases are required
|
|
|
+ * then explicit build operations that take the flag should be used.
|
|
|
+ *
|
|
|
+ *
|
|
|
+ * \section parameterization Parameters
|
|
|
+ * This header may be parameterized by a set of preprocessor macros.
|
|
|
+ *
|
|
|
+ * - CL_HPP_TARGET_OPENCL_VERSION
|
|
|
+ *
|
|
|
+ * Defines the target OpenCL runtime version to build the header
|
|
|
+ * against. Defaults to 200, representing OpenCL 2.0.
|
|
|
+ *
|
|
|
+ * - CL_HPP_NO_STD_STRING
|
|
|
+ *
|
|
|
+ * Do not use the standard library string class. cl::string is not
|
|
|
+ * defined and may be defined by the user before cl2.hpp is
|
|
|
+ * included.
|
|
|
+ *
|
|
|
+ * - CL_HPP_NO_STD_VECTOR
|
|
|
+ *
|
|
|
+ * Do not use the standard library vector class. cl::vector is not
|
|
|
+ * defined and may be defined by the user before cl2.hpp is
|
|
|
+ * included.
|
|
|
+ *
|
|
|
+ * - CL_HPP_NO_STD_ARRAY
|
|
|
+ *
|
|
|
+ * Do not use the standard library array class. cl::array is not
|
|
|
+ * defined and may be defined by the user before cl2.hpp is
|
|
|
+ * included.
|
|
|
+ *
|
|
|
+ * - CL_HPP_NO_STD_UNIQUE_PTR
|
|
|
+ *
|
|
|
+ * Do not use the standard library unique_ptr class. cl::pointer and
|
|
|
+ * the cl::allocate_pointer functions are not defined and may be
|
|
|
+ * defined by the user before cl2.hpp is included.
|
|
|
+ *
|
|
|
+ * - CL_HPP_ENABLE_DEVICE_FISSION
|
|
|
+ *
|
|
|
+ * Enables device fission for OpenCL 1.2 platforms.
|
|
|
+ *
|
|
|
+ * - CL_HPP_ENABLE_EXCEPTIONS
|
|
|
+ *
|
|
|
+ * Enable exceptions for use in the C++ bindings header. This is the
|
|
|
+ * preferred error handling mechanism but is not required.
|
|
|
+ *
|
|
|
+ * - CL_HPP_ENABLE_SIZE_T_COMPATIBILITY
|
|
|
+ *
|
|
|
+ * Backward compatibility option to support cl.hpp-style size_t
|
|
|
+ * class. Replaces the updated std::array derived version and
|
|
|
+ * removal of size_t from the namespace. Note that in this case the
|
|
|
+ * new size_t class is placed in the cl::compatibility namespace and
|
|
|
+ * thus requires an additional using declaration for direct backward
|
|
|
+ * compatibility.
|
|
|
+ *
|
|
|
+ * - CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY
|
|
|
+ *
|
|
|
+ * Enable older vector of pairs interface for construction of
|
|
|
+ * programs.
|
|
|
+ *
|
|
|
+ * - CL_HPP_CL_1_2_DEFAULT_BUILD
|
|
|
+ *
|
|
|
+ * Default to OpenCL C 1.2 compilation rather than OpenCL C 2.0
|
|
|
+ * applies to use of cl::Program construction and other program
|
|
|
+ * build variants.
|
|
|
+ *
|
|
|
+ * - CL_HPP_USE_CL_SUB_GROUPS_KHR
|
|
|
+ *
|
|
|
+ * Enable the cl_khr_subgroups extension.
|
|
|
+ *
|
|
|
+ * - CL_HPP_USE_IL_KHR
|
|
|
+ *
|
|
|
+ * Enable the cl_khr_il_program extension.
|
|
|
+ *
|
|
|
+ *
|
|
|
+ * \section example Example
|
|
|
+ *
|
|
|
+ * The following example shows a general use case for the C++
|
|
|
+ * bindings, including support for the optional exception feature and
|
|
|
+ * also the supplied vector and string classes, see following sections for
|
|
|
+ * decriptions of these features.
|
|
|
+ *
|
|
|
+ * \code
|
|
|
+ #define CL_HPP_ENABLE_EXCEPTIONS
|
|
|
+ #define CL_HPP_TARGET_OPENCL_VERSION 200
|
|
|
+
|
|
|
+ #include <CL/cl2.hpp>
|
|
|
+ #include <iostream>
|
|
|
+ #include <vector>
|
|
|
+ #include <memory>
|
|
|
+ #include <algorithm>
|
|
|
+
|
|
|
+ const int numElements = 32;
|
|
|
+
|
|
|
+ int main(void)
|
|
|
+ {
|
|
|
+ // Filter for a 2.0 platform and set it as the default
|
|
|
+ std::vector<cl::Platform> platforms;
|
|
|
+ cl::Platform::get(&platforms);
|
|
|
+ cl::Platform plat;
|
|
|
+ for (auto &p : platforms) {
|
|
|
+ std::string platver = p.getInfo<CL_PLATFORM_VERSION>();
|
|
|
+ if (platver.find("OpenCL 2.") != std::string::npos) {
|
|
|
+ plat = p;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ if (plat() == 0) {
|
|
|
+ std::cout << "No OpenCL 2.0 platform found.";
|
|
|
+ return -1;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl::Platform newP = cl::Platform::setDefault(plat);
|
|
|
+ if (newP != plat) {
|
|
|
+ std::cout << "Error setting default platform.";
|
|
|
+ return -1;
|
|
|
+ }
|
|
|
+
|
|
|
+ // Use C++11 raw string literals for kernel source code
|
|
|
+ std::string kernel1{R"CLC(
|
|
|
+ global int globalA;
|
|
|
+ kernel void updateGlobal()
|
|
|
+ {
|
|
|
+ globalA = 75;
|
|
|
+ }
|
|
|
+ )CLC"};
|
|
|
+ std::string kernel2{R"CLC(
|
|
|
+ typedef struct { global int *bar; } Foo;
|
|
|
+ kernel void vectorAdd(global const Foo* aNum, global const int *inputA, global const int *inputB,
|
|
|
+ global int *output, int val, write_only pipe int outPipe, queue_t childQueue)
|
|
|
+ {
|
|
|
+ output[get_global_id(0)] = inputA[get_global_id(0)] + inputB[get_global_id(0)] + val + *(aNum->bar);
|
|
|
+ write_pipe(outPipe, &val);
|
|
|
+ queue_t default_queue = get_default_queue();
|
|
|
+ ndrange_t ndrange = ndrange_1D(get_global_size(0)/2, get_global_size(0)/2);
|
|
|
+
|
|
|
+ // Have a child kernel write into third quarter of output
|
|
|
+ enqueue_kernel(default_queue, CLK_ENQUEUE_FLAGS_WAIT_KERNEL, ndrange,
|
|
|
+ ^{
|
|
|
+ output[get_global_size(0)*2 + get_global_id(0)] =
|
|
|
+ inputA[get_global_size(0)*2 + get_global_id(0)] + inputB[get_global_size(0)*2 + get_global_id(0)] + globalA;
|
|
|
+ });
|
|
|
+
|
|
|
+ // Have a child kernel write into last quarter of output
|
|
|
+ enqueue_kernel(childQueue, CLK_ENQUEUE_FLAGS_WAIT_KERNEL, ndrange,
|
|
|
+ ^{
|
|
|
+ output[get_global_size(0)*3 + get_global_id(0)] =
|
|
|
+ inputA[get_global_size(0)*3 + get_global_id(0)] + inputB[get_global_size(0)*3 + get_global_id(0)] + globalA + 2;
|
|
|
+ });
|
|
|
+ }
|
|
|
+ )CLC"};
|
|
|
+
|
|
|
+ // New simpler string interface style
|
|
|
+ std::vector<std::string> programStrings {kernel1, kernel2};
|
|
|
+
|
|
|
+ cl::Program vectorAddProgram(programStrings);
|
|
|
+ try {
|
|
|
+ vectorAddProgram.build("-cl-std=CL2.0");
|
|
|
+ }
|
|
|
+ catch (...) {
|
|
|
+ // Print build info for all devices
|
|
|
+ cl_int buildErr = CL_SUCCESS;
|
|
|
+ auto buildInfo = vectorAddProgram.getBuildInfo<CL_PROGRAM_BUILD_LOG>(&buildErr);
|
|
|
+ for (auto &pair : buildInfo) {
|
|
|
+ std::cerr << pair.second << std::endl << std::endl;
|
|
|
+ }
|
|
|
+
|
|
|
+ return 1;
|
|
|
+ }
|
|
|
+
|
|
|
+ typedef struct { int *bar; } Foo;
|
|
|
+
|
|
|
+ // Get and run kernel that initializes the program-scope global
|
|
|
+ // A test for kernels that take no arguments
|
|
|
+ auto program2Kernel =
|
|
|
+ cl::KernelFunctor<>(vectorAddProgram, "updateGlobal");
|
|
|
+ program2Kernel(
|
|
|
+ cl::EnqueueArgs(
|
|
|
+ cl::NDRange(1)));
|
|
|
+
|
|
|
+ //////////////////
|
|
|
+ // SVM allocations
|
|
|
+
|
|
|
+ auto anSVMInt = cl::allocate_svm<int, cl::SVMTraitCoarse<>>();
|
|
|
+ *anSVMInt = 5;
|
|
|
+ cl::SVMAllocator<Foo, cl::SVMTraitCoarse<cl::SVMTraitReadOnly<>>> svmAllocReadOnly;
|
|
|
+ auto fooPointer = cl::allocate_pointer<Foo>(svmAllocReadOnly);
|
|
|
+ fooPointer->bar = anSVMInt.get();
|
|
|
+ cl::SVMAllocator<int, cl::SVMTraitCoarse<>> svmAlloc;
|
|
|
+ std::vector<int, cl::SVMAllocator<int, cl::SVMTraitCoarse<>>> inputA(numElements, 1, svmAlloc);
|
|
|
+ cl::coarse_svm_vector<int> inputB(numElements, 2, svmAlloc);
|
|
|
+
|
|
|
+ //
|
|
|
+ //////////////
|
|
|
+
|
|
|
+ // Traditional cl_mem allocations
|
|
|
+ std::vector<int> output(numElements, 0xdeadbeef);
|
|
|
+ cl::Buffer outputBuffer(begin(output), end(output), false);
|
|
|
+ cl::Pipe aPipe(sizeof(cl_int), numElements / 2);
|
|
|
+
|
|
|
+ // Default command queue, also passed in as a parameter
|
|
|
+ cl::DeviceCommandQueue defaultDeviceQueue = cl::DeviceCommandQueue::makeDefault(
|
|
|
+ cl::Context::getDefault(), cl::Device::getDefault());
|
|
|
+
|
|
|
+ auto vectorAddKernel =
|
|
|
+ cl::KernelFunctor<
|
|
|
+ decltype(fooPointer)&,
|
|
|
+ int*,
|
|
|
+ cl::coarse_svm_vector<int>&,
|
|
|
+ cl::Buffer,
|
|
|
+ int,
|
|
|
+ cl::Pipe&,
|
|
|
+ cl::DeviceCommandQueue
|
|
|
+ >(vectorAddProgram, "vectorAdd");
|
|
|
+
|
|
|
+ // Ensure that the additional SVM pointer is available to the kernel
|
|
|
+ // This one was not passed as a parameter
|
|
|
+ vectorAddKernel.setSVMPointers(anSVMInt);
|
|
|
+
|
|
|
+ // Hand control of coarse allocations to runtime
|
|
|
+ cl::enqueueUnmapSVM(anSVMInt);
|
|
|
+ cl::enqueueUnmapSVM(fooPointer);
|
|
|
+ cl::unmapSVM(inputB);
|
|
|
+ cl::unmapSVM(output2);
|
|
|
+
|
|
|
+ cl_int error;
|
|
|
+ vectorAddKernel(
|
|
|
+ cl::EnqueueArgs(
|
|
|
+ cl::NDRange(numElements/2),
|
|
|
+ cl::NDRange(numElements/2)),
|
|
|
+ fooPointer,
|
|
|
+ inputA.data(),
|
|
|
+ inputB,
|
|
|
+ outputBuffer,
|
|
|
+ 3,
|
|
|
+ aPipe,
|
|
|
+ defaultDeviceQueue,
|
|
|
+ error
|
|
|
+ );
|
|
|
+
|
|
|
+ cl::copy(outputBuffer, begin(output), end(output));
|
|
|
+ // Grab the SVM output vector using a map
|
|
|
+ cl::mapSVM(output2);
|
|
|
+
|
|
|
+ cl::Device d = cl::Device::getDefault();
|
|
|
+
|
|
|
+ std::cout << "Output:\n";
|
|
|
+ for (int i = 1; i < numElements; ++i) {
|
|
|
+ std::cout << "\t" << output[i] << "\n";
|
|
|
+ }
|
|
|
+ std::cout << "\n\n";
|
|
|
+
|
|
|
+ return 0;
|
|
|
+ }
|
|
|
+ *
|
|
|
+ * \endcode
|
|
|
+ *
|
|
|
+ */
|
|
|
+#ifndef CL_HPP_
|
|
|
+#define CL_HPP_
|
|
|
+
|
|
|
+/* Handle deprecated preprocessor definitions. In each case, we only check for
|
|
|
+ * the old name if the new name is not defined, so that user code can define
|
|
|
+ * both and hence work with either version of the bindings.
|
|
|
+ */
|
|
|
+#if !defined(CL_HPP_USE_DX_INTEROP) && defined(USE_DX_INTEROP)
|
|
|
+# pragma message("cl2.hpp: USE_DX_INTEROP is deprecated. Define CL_HPP_USE_DX_INTEROP instead")
|
|
|
+# define CL_HPP_USE_DX_INTEROP
|
|
|
+#endif
|
|
|
+#if !defined(CL_HPP_USE_CL_DEVICE_FISSION) && defined(USE_CL_DEVICE_FISSION)
|
|
|
+# pragma message("cl2.hpp: USE_CL_DEVICE_FISSION is deprecated. Define CL_HPP_USE_CL_DEVICE_FISSION instead")
|
|
|
+# define CL_HPP_USE_CL_DEVICE_FISSION
|
|
|
+#endif
|
|
|
+#if !defined(CL_HPP_ENABLE_EXCEPTIONS) && defined(__CL_ENABLE_EXCEPTIONS)
|
|
|
+# pragma message("cl2.hpp: __CL_ENABLE_EXCEPTIONS is deprecated. Define CL_HPP_ENABLE_EXCEPTIONS instead")
|
|
|
+# define CL_HPP_ENABLE_EXCEPTIONS
|
|
|
+#endif
|
|
|
+#if !defined(CL_HPP_NO_STD_VECTOR) && defined(__NO_STD_VECTOR)
|
|
|
+# pragma message("cl2.hpp: __NO_STD_VECTOR is deprecated. Define CL_HPP_NO_STD_VECTOR instead")
|
|
|
+# define CL_HPP_NO_STD_VECTOR
|
|
|
+#endif
|
|
|
+#if !defined(CL_HPP_NO_STD_STRING) && defined(__NO_STD_STRING)
|
|
|
+# pragma message("cl2.hpp: __NO_STD_STRING is deprecated. Define CL_HPP_NO_STD_STRING instead")
|
|
|
+# define CL_HPP_NO_STD_STRING
|
|
|
+#endif
|
|
|
+#if defined(VECTOR_CLASS)
|
|
|
+# pragma message("cl2.hpp: VECTOR_CLASS is deprecated. Alias cl::vector instead")
|
|
|
+#endif
|
|
|
+#if defined(STRING_CLASS)
|
|
|
+# pragma message("cl2.hpp: STRING_CLASS is deprecated. Alias cl::string instead.")
|
|
|
+#endif
|
|
|
+#if !defined(CL_HPP_USER_OVERRIDE_ERROR_STRINGS) && defined(__CL_USER_OVERRIDE_ERROR_STRINGS)
|
|
|
+# pragma message("cl2.hpp: __CL_USER_OVERRIDE_ERROR_STRINGS is deprecated. Define CL_HPP_USER_OVERRIDE_ERROR_STRINGS instead")
|
|
|
+# define CL_HPP_USER_OVERRIDE_ERROR_STRINGS
|
|
|
+#endif
|
|
|
+
|
|
|
+/* Warn about features that are no longer supported
|
|
|
+ */
|
|
|
+#if defined(__USE_DEV_VECTOR)
|
|
|
+# pragma message("cl2.hpp: __USE_DEV_VECTOR is no longer supported. Expect compilation errors")
|
|
|
+#endif
|
|
|
+#if defined(__USE_DEV_STRING)
|
|
|
+# pragma message("cl2.hpp: __USE_DEV_STRING is no longer supported. Expect compilation errors")
|
|
|
+#endif
|
|
|
+
|
|
|
+/* Detect which version to target */
|
|
|
+#if !defined(CL_HPP_TARGET_OPENCL_VERSION)
|
|
|
+# pragma message("cl2.hpp: CL_HPP_TARGET_OPENCL_VERSION is not defined. It will default to 220 (OpenCL 2.2)")
|
|
|
+# define CL_HPP_TARGET_OPENCL_VERSION 220
|
|
|
+#endif
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION != 100 && \
|
|
|
+ CL_HPP_TARGET_OPENCL_VERSION != 110 && \
|
|
|
+ CL_HPP_TARGET_OPENCL_VERSION != 120 && \
|
|
|
+ CL_HPP_TARGET_OPENCL_VERSION != 200 && \
|
|
|
+ CL_HPP_TARGET_OPENCL_VERSION != 210 && \
|
|
|
+ CL_HPP_TARGET_OPENCL_VERSION != 220
|
|
|
+# pragma message("cl2.hpp: CL_HPP_TARGET_OPENCL_VERSION is not a valid value (100, 110, 120, 200, 210 or 220). It will be set to 220")
|
|
|
+# undef CL_HPP_TARGET_OPENCL_VERSION
|
|
|
+# define CL_HPP_TARGET_OPENCL_VERSION 220
|
|
|
+#endif
|
|
|
+
|
|
|
+/* Forward target OpenCL version to C headers if necessary */
|
|
|
+#if defined(CL_TARGET_OPENCL_VERSION)
|
|
|
+/* Warn if prior definition of CL_TARGET_OPENCL_VERSION is lower than
|
|
|
+ * requested C++ bindings version */
|
|
|
+#if CL_TARGET_OPENCL_VERSION < CL_HPP_TARGET_OPENCL_VERSION
|
|
|
+# pragma message("CL_TARGET_OPENCL_VERSION is already defined as is lower than CL_HPP_TARGET_OPENCL_VERSION")
|
|
|
+#endif
|
|
|
+#else
|
|
|
+# define CL_TARGET_OPENCL_VERSION CL_HPP_TARGET_OPENCL_VERSION
|
|
|
+#endif
|
|
|
+
|
|
|
+#if !defined(CL_HPP_MINIMUM_OPENCL_VERSION)
|
|
|
+# define CL_HPP_MINIMUM_OPENCL_VERSION 200
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION != 100 && \
|
|
|
+ CL_HPP_MINIMUM_OPENCL_VERSION != 110 && \
|
|
|
+ CL_HPP_MINIMUM_OPENCL_VERSION != 120 && \
|
|
|
+ CL_HPP_MINIMUM_OPENCL_VERSION != 200 && \
|
|
|
+ CL_HPP_MINIMUM_OPENCL_VERSION != 210 && \
|
|
|
+ CL_HPP_MINIMUM_OPENCL_VERSION != 220
|
|
|
+# pragma message("cl2.hpp: CL_HPP_MINIMUM_OPENCL_VERSION is not a valid value (100, 110, 120, 200, 210 or 220). It will be set to 100")
|
|
|
+# undef CL_HPP_MINIMUM_OPENCL_VERSION
|
|
|
+# define CL_HPP_MINIMUM_OPENCL_VERSION 100
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION > CL_HPP_TARGET_OPENCL_VERSION
|
|
|
+# error "CL_HPP_MINIMUM_OPENCL_VERSION must not be greater than CL_HPP_TARGET_OPENCL_VERSION"
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 100 && !defined(CL_USE_DEPRECATED_OPENCL_1_0_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_1_0_APIS
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 110 && !defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_1_1_APIS
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 120 && !defined(CL_USE_DEPRECATED_OPENCL_1_2_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_1_2_APIS
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 200 && !defined(CL_USE_DEPRECATED_OPENCL_2_0_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_2_0_APIS
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 210 && !defined(CL_USE_DEPRECATED_OPENCL_2_1_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_2_1_APIS
|
|
|
+#endif
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION <= 220 && !defined(CL_USE_DEPRECATED_OPENCL_2_2_APIS)
|
|
|
+# define CL_USE_DEPRECATED_OPENCL_2_2_APIS
|
|
|
+#endif
|
|
|
+
|
|
|
+#ifdef _WIN32
|
|
|
+
|
|
|
+#include <malloc.h>
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_DX_INTEROP)
|
|
|
+#include <CL/cl_d3d10.h>
|
|
|
+#include <CL/cl_dx9_media_sharing.h>
|
|
|
+#endif
|
|
|
+#endif // _WIN32
|
|
|
+
|
|
|
+#if defined(_MSC_VER)
|
|
|
+#include <intrin.h>
|
|
|
+#endif // _MSC_VER
|
|
|
+
|
|
|
+ // Check for a valid C++ version
|
|
|
+
|
|
|
+// Need to do both tests here because for some reason __cplusplus is not
|
|
|
+// updated in visual studio
|
|
|
+#if (!defined(_MSC_VER) && __cplusplus < 201103L) || (defined(_MSC_VER) && _MSC_VER < 1700)
|
|
|
+#error Visual studio 2013 or another C++11-supporting compiler required
|
|
|
+#endif
|
|
|
+
|
|
|
+//
|
|
|
+#if defined(CL_HPP_USE_CL_DEVICE_FISSION) || defined(CL_HPP_USE_CL_SUB_GROUPS_KHR)
|
|
|
+#include <CL/cl_ext.h>
|
|
|
+#endif
|
|
|
+
|
|
|
+#if defined(__APPLE__) || defined(__MACOSX)
|
|
|
+#include <OpenCL/opencl.h>
|
|
|
+#else
|
|
|
+#include <CL/opencl.h>
|
|
|
+#endif // !__APPLE__
|
|
|
+
|
|
|
+#if (__cplusplus >= 201103L)
|
|
|
+#define CL_HPP_NOEXCEPT_ noexcept
|
|
|
+#else
|
|
|
+#define CL_HPP_NOEXCEPT_
|
|
|
+#endif
|
|
|
+
|
|
|
+#if defined(_MSC_VER)
|
|
|
+# define CL_HPP_DEFINE_STATIC_MEMBER_ __declspec(selectany)
|
|
|
+#elif defined(__MINGW32__)
|
|
|
+# define CL_HPP_DEFINE_STATIC_MEMBER_ __attribute__((selectany))
|
|
|
+#else
|
|
|
+# define CL_HPP_DEFINE_STATIC_MEMBER_ __attribute__((weak))
|
|
|
+#endif // !_MSC_VER
|
|
|
+
|
|
|
+// Define deprecated prefixes and suffixes to ensure compilation
|
|
|
+// in case they are not pre-defined
|
|
|
+#if !defined(CL_EXT_PREFIX__VERSION_1_1_DEPRECATED)
|
|
|
+#define CL_EXT_PREFIX__VERSION_1_1_DEPRECATED
|
|
|
+#endif // #if !defined(CL_EXT_PREFIX__VERSION_1_1_DEPRECATED)
|
|
|
+#if !defined(CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED)
|
|
|
+#define CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED
|
|
|
+#endif // #if !defined(CL_EXT_PREFIX__VERSION_1_1_DEPRECATED)
|
|
|
+
|
|
|
+#if !defined(CL_EXT_PREFIX__VERSION_1_2_DEPRECATED)
|
|
|
+#define CL_EXT_PREFIX__VERSION_1_2_DEPRECATED
|
|
|
+#endif // #if !defined(CL_EXT_PREFIX__VERSION_1_2_DEPRECATED)
|
|
|
+#if !defined(CL_EXT_SUFFIX__VERSION_1_2_DEPRECATED)
|
|
|
+#define CL_EXT_SUFFIX__VERSION_1_2_DEPRECATED
|
|
|
+#endif // #if !defined(CL_EXT_PREFIX__VERSION_1_2_DEPRECATED)
|
|
|
+
|
|
|
+#if !defined(CL_CALLBACK)
|
|
|
+#define CL_CALLBACK
|
|
|
+#endif //CL_CALLBACK
|
|
|
+
|
|
|
+#include <utility>
|
|
|
+#include <limits>
|
|
|
+#include <iterator>
|
|
|
+#include <mutex>
|
|
|
+#include <cstring>
|
|
|
+#include <functional>
|
|
|
+
|
|
|
+
|
|
|
+// Define a size_type to represent a correctly resolved size_t
|
|
|
+#if defined(CL_HPP_ENABLE_SIZE_T_COMPATIBILITY)
|
|
|
+namespace cl {
|
|
|
+ using size_type = ::size_t;
|
|
|
+} // namespace cl
|
|
|
+#else // #if defined(CL_HPP_ENABLE_SIZE_T_COMPATIBILITY)
|
|
|
+namespace cl {
|
|
|
+ using size_type = size_t;
|
|
|
+} // namespace cl
|
|
|
+#endif // #if defined(CL_HPP_ENABLE_SIZE_T_COMPATIBILITY)
|
|
|
+
|
|
|
+
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+#include <exception>
|
|
|
+#endif // #if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+
|
|
|
+#if !defined(CL_HPP_NO_STD_VECTOR)
|
|
|
+#include <vector>
|
|
|
+namespace cl {
|
|
|
+ template < class T, class Alloc = std::allocator<T> >
|
|
|
+ using vector = std::vector<T, Alloc>;
|
|
|
+} // namespace cl
|
|
|
+#endif // #if !defined(CL_HPP_NO_STD_VECTOR)
|
|
|
+
|
|
|
+#if !defined(CL_HPP_NO_STD_STRING)
|
|
|
+#include <string>
|
|
|
+namespace cl {
|
|
|
+ using string = std::string;
|
|
|
+} // namespace cl
|
|
|
+#endif // #if !defined(CL_HPP_NO_STD_STRING)
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+#if !defined(CL_HPP_NO_STD_UNIQUE_PTR)
|
|
|
+#include <memory>
|
|
|
+namespace cl {
|
|
|
+ // Replace unique_ptr and allocate_pointer for internal use
|
|
|
+ // to allow user to replace them
|
|
|
+ template<class T, class D>
|
|
|
+ using pointer = std::unique_ptr<T, D>;
|
|
|
+} // namespace cl
|
|
|
+#endif
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if !defined(CL_HPP_NO_STD_ARRAY)
|
|
|
+#include <array>
|
|
|
+namespace cl {
|
|
|
+ template < class T, size_type N >
|
|
|
+ using array = std::array<T, N>;
|
|
|
+} // namespace cl
|
|
|
+#endif // #if !defined(CL_HPP_NO_STD_ARRAY)
|
|
|
+
|
|
|
+// Define size_type appropriately to allow backward-compatibility
|
|
|
+// use of the old size_t interface class
|
|
|
+#if defined(CL_HPP_ENABLE_SIZE_T_COMPATIBILITY)
|
|
|
+namespace cl {
|
|
|
+ namespace compatibility {
|
|
|
+ /*! \brief class used to interface between C++ and
|
|
|
+ * OpenCL C calls that require arrays of size_t values, whose
|
|
|
+ * size is known statically.
|
|
|
+ */
|
|
|
+ template <int N>
|
|
|
+ class size_t
|
|
|
+ {
|
|
|
+ private:
|
|
|
+ size_type data_[N];
|
|
|
+
|
|
|
+ public:
|
|
|
+ //! \brief Initialize size_t to all 0s
|
|
|
+ size_t()
|
|
|
+ {
|
|
|
+ for (int i = 0; i < N; ++i) {
|
|
|
+ data_[i] = 0;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ size_t(const array<size_type, N> &rhs)
|
|
|
+ {
|
|
|
+ for (int i = 0; i < N; ++i) {
|
|
|
+ data_[i] = rhs[i];
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type& operator[](int index)
|
|
|
+ {
|
|
|
+ return data_[index];
|
|
|
+ }
|
|
|
+
|
|
|
+ const size_type& operator[](int index) const
|
|
|
+ {
|
|
|
+ return data_[index];
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Conversion operator to T*.
|
|
|
+ operator size_type* () { return data_; }
|
|
|
+
|
|
|
+ //! \brief Conversion operator to const T*.
|
|
|
+ operator const size_type* () const { return data_; }
|
|
|
+
|
|
|
+ operator array<size_type, N>() const
|
|
|
+ {
|
|
|
+ array<size_type, N> ret;
|
|
|
+
|
|
|
+ for (int i = 0; i < N; ++i) {
|
|
|
+ ret[i] = data_[i];
|
|
|
+ }
|
|
|
+ return ret;
|
|
|
+ }
|
|
|
+ };
|
|
|
+ } // namespace compatibility
|
|
|
+
|
|
|
+ template<int N>
|
|
|
+ using size_t = compatibility::size_t<N>;
|
|
|
+} // namespace cl
|
|
|
+#endif // #if defined(CL_HPP_ENABLE_SIZE_T_COMPATIBILITY)
|
|
|
+
|
|
|
+// Helper alias to avoid confusing the macros
|
|
|
+namespace cl {
|
|
|
+ namespace detail {
|
|
|
+ using size_t_array = array<size_type, 3>;
|
|
|
+ } // namespace detail
|
|
|
+} // namespace cl
|
|
|
+
|
|
|
+
|
|
|
+/*! \namespace cl
|
|
|
+ *
|
|
|
+ * \brief The OpenCL C++ bindings are defined within this namespace.
|
|
|
+ *
|
|
|
+ */
|
|
|
+namespace cl {
|
|
|
+ class Memory;
|
|
|
+
|
|
|
+#define CL_HPP_INIT_CL_EXT_FCN_PTR_(name) \
|
|
|
+ if (!pfn_##name) { \
|
|
|
+ pfn_##name = (PFN_##name) \
|
|
|
+ clGetExtensionFunctionAddress(#name); \
|
|
|
+ if (!pfn_##name) { \
|
|
|
+ } \
|
|
|
+ }
|
|
|
+
|
|
|
+#define CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_(platform, name) \
|
|
|
+ if (!pfn_##name) { \
|
|
|
+ pfn_##name = (PFN_##name) \
|
|
|
+ clGetExtensionFunctionAddressForPlatform(platform, #name); \
|
|
|
+ if (!pfn_##name) { \
|
|
|
+ } \
|
|
|
+ }
|
|
|
+
|
|
|
+ class Program;
|
|
|
+ class Device;
|
|
|
+ class Context;
|
|
|
+ class CommandQueue;
|
|
|
+ class DeviceCommandQueue;
|
|
|
+ class Memory;
|
|
|
+ class Buffer;
|
|
|
+ class Pipe;
|
|
|
+
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ /*! \brief Exception class
|
|
|
+ *
|
|
|
+ * This may be thrown by API functions when CL_HPP_ENABLE_EXCEPTIONS is defined.
|
|
|
+ */
|
|
|
+ class Error : public std::exception
|
|
|
+ {
|
|
|
+ private:
|
|
|
+ cl_int err_;
|
|
|
+ const char * errStr_;
|
|
|
+ public:
|
|
|
+ /*! \brief Create a new CL error exception for a given error code
|
|
|
+ * and corresponding message.
|
|
|
+ *
|
|
|
+ * \param err error code value.
|
|
|
+ *
|
|
|
+ * \param errStr a descriptive string that must remain in scope until
|
|
|
+ * handling of the exception has concluded. If set, it
|
|
|
+ * will be returned by what().
|
|
|
+ */
|
|
|
+ Error(cl_int err, const char * errStr = NULL) : err_(err), errStr_(errStr)
|
|
|
+ {}
|
|
|
+
|
|
|
+ ~Error() throw() {}
|
|
|
+
|
|
|
+ /*! \brief Get error string associated with exception
|
|
|
+ *
|
|
|
+ * \return A memory pointer to the error message string.
|
|
|
+ */
|
|
|
+ virtual const char * what() const throw ()
|
|
|
+ {
|
|
|
+ if (errStr_ == NULL) {
|
|
|
+ return "empty";
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ return errStr_;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Get error code associated with exception
|
|
|
+ *
|
|
|
+ * \return The error code.
|
|
|
+ */
|
|
|
+ cl_int err(void) const { return err_; }
|
|
|
+ };
|
|
|
+#define CL_HPP_ERR_STR_(x) #x
|
|
|
+#else
|
|
|
+#define CL_HPP_ERR_STR_(x) NULL
|
|
|
+#endif // CL_HPP_ENABLE_EXCEPTIONS
|
|
|
+
|
|
|
+
|
|
|
+namespace detail
|
|
|
+{
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+static inline cl_int errHandler (
|
|
|
+ cl_int err,
|
|
|
+ const char * errStr = NULL)
|
|
|
+{
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ throw Error(err, errStr);
|
|
|
+ }
|
|
|
+ return err;
|
|
|
+}
|
|
|
+#else
|
|
|
+static inline cl_int errHandler (cl_int err, const char * errStr = NULL)
|
|
|
+{
|
|
|
+ (void) errStr; // suppress unused variable warning
|
|
|
+ return err;
|
|
|
+}
|
|
|
+#endif // CL_HPP_ENABLE_EXCEPTIONS
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+//! \cond DOXYGEN_DETAIL
|
|
|
+#if !defined(CL_HPP_USER_OVERRIDE_ERROR_STRINGS)
|
|
|
+#define __GET_DEVICE_INFO_ERR CL_HPP_ERR_STR_(clGetDeviceInfo)
|
|
|
+#define __GET_PLATFORM_INFO_ERR CL_HPP_ERR_STR_(clGetPlatformInfo)
|
|
|
+#define __GET_DEVICE_IDS_ERR CL_HPP_ERR_STR_(clGetDeviceIDs)
|
|
|
+#define __GET_PLATFORM_IDS_ERR CL_HPP_ERR_STR_(clGetPlatformIDs)
|
|
|
+#define __GET_CONTEXT_INFO_ERR CL_HPP_ERR_STR_(clGetContextInfo)
|
|
|
+#define __GET_EVENT_INFO_ERR CL_HPP_ERR_STR_(clGetEventInfo)
|
|
|
+#define __GET_EVENT_PROFILE_INFO_ERR CL_HPP_ERR_STR_(clGetEventProfileInfo)
|
|
|
+#define __GET_MEM_OBJECT_INFO_ERR CL_HPP_ERR_STR_(clGetMemObjectInfo)
|
|
|
+#define __GET_IMAGE_INFO_ERR CL_HPP_ERR_STR_(clGetImageInfo)
|
|
|
+#define __GET_SAMPLER_INFO_ERR CL_HPP_ERR_STR_(clGetSamplerInfo)
|
|
|
+#define __GET_KERNEL_INFO_ERR CL_HPP_ERR_STR_(clGetKernelInfo)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __GET_KERNEL_ARG_INFO_ERR CL_HPP_ERR_STR_(clGetKernelArgInfo)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __GET_KERNEL_SUB_GROUP_INFO_ERR CL_HPP_ERR_STR_(clGetKernelSubGroupInfo)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __GET_KERNEL_WORK_GROUP_INFO_ERR CL_HPP_ERR_STR_(clGetKernelWorkGroupInfo)
|
|
|
+#define __GET_PROGRAM_INFO_ERR CL_HPP_ERR_STR_(clGetProgramInfo)
|
|
|
+#define __GET_PROGRAM_BUILD_INFO_ERR CL_HPP_ERR_STR_(clGetProgramBuildInfo)
|
|
|
+#define __GET_COMMAND_QUEUE_INFO_ERR CL_HPP_ERR_STR_(clGetCommandQueueInfo)
|
|
|
+
|
|
|
+#define __CREATE_CONTEXT_ERR CL_HPP_ERR_STR_(clCreateContext)
|
|
|
+#define __CREATE_CONTEXT_FROM_TYPE_ERR CL_HPP_ERR_STR_(clCreateContextFromType)
|
|
|
+#define __GET_SUPPORTED_IMAGE_FORMATS_ERR CL_HPP_ERR_STR_(clGetSupportedImageFormats)
|
|
|
+
|
|
|
+#define __CREATE_BUFFER_ERR CL_HPP_ERR_STR_(clCreateBuffer)
|
|
|
+#define __COPY_ERR CL_HPP_ERR_STR_(cl::copy)
|
|
|
+#define __CREATE_SUBBUFFER_ERR CL_HPP_ERR_STR_(clCreateSubBuffer)
|
|
|
+#define __CREATE_GL_BUFFER_ERR CL_HPP_ERR_STR_(clCreateFromGLBuffer)
|
|
|
+#define __CREATE_GL_RENDER_BUFFER_ERR CL_HPP_ERR_STR_(clCreateFromGLBuffer)
|
|
|
+#define __GET_GL_OBJECT_INFO_ERR CL_HPP_ERR_STR_(clGetGLObjectInfo)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __CREATE_IMAGE_ERR CL_HPP_ERR_STR_(clCreateImage)
|
|
|
+#define __CREATE_GL_TEXTURE_ERR CL_HPP_ERR_STR_(clCreateFromGLTexture)
|
|
|
+#define __IMAGE_DIMENSION_ERR CL_HPP_ERR_STR_(Incorrect image dimensions)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __SET_MEM_OBJECT_DESTRUCTOR_CALLBACK_ERR CL_HPP_ERR_STR_(clSetMemObjectDestructorCallback)
|
|
|
+
|
|
|
+#define __CREATE_USER_EVENT_ERR CL_HPP_ERR_STR_(clCreateUserEvent)
|
|
|
+#define __SET_USER_EVENT_STATUS_ERR CL_HPP_ERR_STR_(clSetUserEventStatus)
|
|
|
+#define __SET_EVENT_CALLBACK_ERR CL_HPP_ERR_STR_(clSetEventCallback)
|
|
|
+#define __WAIT_FOR_EVENTS_ERR CL_HPP_ERR_STR_(clWaitForEvents)
|
|
|
+
|
|
|
+#define __CREATE_KERNEL_ERR CL_HPP_ERR_STR_(clCreateKernel)
|
|
|
+#define __SET_KERNEL_ARGS_ERR CL_HPP_ERR_STR_(clSetKernelArg)
|
|
|
+#define __CREATE_PROGRAM_WITH_SOURCE_ERR CL_HPP_ERR_STR_(clCreateProgramWithSource)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __CREATE_PROGRAM_WITH_IL_ERR CL_HPP_ERR_STR_(clCreateProgramWithIL)
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __CREATE_PROGRAM_WITH_BINARY_ERR CL_HPP_ERR_STR_(clCreateProgramWithBinary)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#define __CREATE_PROGRAM_WITH_IL_ERR CL_HPP_ERR_STR_(clCreateProgramWithIL)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __CREATE_PROGRAM_WITH_BUILT_IN_KERNELS_ERR CL_HPP_ERR_STR_(clCreateProgramWithBuiltInKernels)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __BUILD_PROGRAM_ERR CL_HPP_ERR_STR_(clBuildProgram)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __COMPILE_PROGRAM_ERR CL_HPP_ERR_STR_(clCompileProgram)
|
|
|
+#define __LINK_PROGRAM_ERR CL_HPP_ERR_STR_(clLinkProgram)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __CREATE_KERNELS_IN_PROGRAM_ERR CL_HPP_ERR_STR_(clCreateKernelsInProgram)
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR CL_HPP_ERR_STR_(clCreateCommandQueueWithProperties)
|
|
|
+#define __CREATE_SAMPLER_WITH_PROPERTIES_ERR CL_HPP_ERR_STR_(clCreateSamplerWithProperties)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#define __SET_COMMAND_QUEUE_PROPERTY_ERR CL_HPP_ERR_STR_(clSetCommandQueueProperty)
|
|
|
+#define __ENQUEUE_READ_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueReadBuffer)
|
|
|
+#define __ENQUEUE_READ_BUFFER_RECT_ERR CL_HPP_ERR_STR_(clEnqueueReadBufferRect)
|
|
|
+#define __ENQUEUE_WRITE_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueWriteBuffer)
|
|
|
+#define __ENQUEUE_WRITE_BUFFER_RECT_ERR CL_HPP_ERR_STR_(clEnqueueWriteBufferRect)
|
|
|
+#define __ENQEUE_COPY_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueCopyBuffer)
|
|
|
+#define __ENQEUE_COPY_BUFFER_RECT_ERR CL_HPP_ERR_STR_(clEnqueueCopyBufferRect)
|
|
|
+#define __ENQUEUE_FILL_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueFillBuffer)
|
|
|
+#define __ENQUEUE_READ_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueReadImage)
|
|
|
+#define __ENQUEUE_WRITE_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueWriteImage)
|
|
|
+#define __ENQUEUE_COPY_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueCopyImage)
|
|
|
+#define __ENQUEUE_FILL_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueFillImage)
|
|
|
+#define __ENQUEUE_COPY_IMAGE_TO_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueCopyImageToBuffer)
|
|
|
+#define __ENQUEUE_COPY_BUFFER_TO_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueCopyBufferToImage)
|
|
|
+#define __ENQUEUE_MAP_BUFFER_ERR CL_HPP_ERR_STR_(clEnqueueMapBuffer)
|
|
|
+#define __ENQUEUE_MAP_IMAGE_ERR CL_HPP_ERR_STR_(clEnqueueMapImage)
|
|
|
+#define __ENQUEUE_UNMAP_MEM_OBJECT_ERR CL_HPP_ERR_STR_(clEnqueueUnMapMemObject)
|
|
|
+#define __ENQUEUE_NDRANGE_KERNEL_ERR CL_HPP_ERR_STR_(clEnqueueNDRangeKernel)
|
|
|
+#define __ENQUEUE_NATIVE_KERNEL CL_HPP_ERR_STR_(clEnqueueNativeKernel)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __ENQUEUE_MIGRATE_MEM_OBJECTS_ERR CL_HPP_ERR_STR_(clEnqueueMigrateMemObjects)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#define __ENQUEUE_MIGRATE_SVM_ERR CL_HPP_ERR_STR_(clEnqueueSVMMigrateMem)
|
|
|
+#define __SET_DEFAULT_DEVICE_COMMAND_QUEUE_ERR CL_HPP_ERR_STR_(clSetDefaultDeviceCommandQueue)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+
|
|
|
+#define __ENQUEUE_ACQUIRE_GL_ERR CL_HPP_ERR_STR_(clEnqueueAcquireGLObjects)
|
|
|
+#define __ENQUEUE_RELEASE_GL_ERR CL_HPP_ERR_STR_(clEnqueueReleaseGLObjects)
|
|
|
+
|
|
|
+#define __CREATE_PIPE_ERR CL_HPP_ERR_STR_(clCreatePipe)
|
|
|
+#define __GET_PIPE_INFO_ERR CL_HPP_ERR_STR_(clGetPipeInfo)
|
|
|
+
|
|
|
+
|
|
|
+#define __RETAIN_ERR CL_HPP_ERR_STR_(Retain Object)
|
|
|
+#define __RELEASE_ERR CL_HPP_ERR_STR_(Release Object)
|
|
|
+#define __FLUSH_ERR CL_HPP_ERR_STR_(clFlush)
|
|
|
+#define __FINISH_ERR CL_HPP_ERR_STR_(clFinish)
|
|
|
+#define __VECTOR_CAPACITY_ERR CL_HPP_ERR_STR_(Vector capacity error)
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#define __GET_HOST_TIMER_ERR CL_HPP_ERR_STR_(clGetHostTimer)
|
|
|
+#define __GET_DEVICE_AND_HOST_TIMER_ERR CL_HPP_ERR_STR_(clGetDeviceAndHostTimer)
|
|
|
+#endif
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+#define __SET_PROGRAM_RELEASE_CALLBACK_ERR CL_HPP_ERR_STR_(clSetProgramReleaseCallback)
|
|
|
+#define __SET_PROGRAM_SPECIALIZATION_CONSTANT_ERR CL_HPP_ERR_STR_(clSetProgramSpecializationConstant)
|
|
|
+#endif
|
|
|
+
|
|
|
+
|
|
|
+/**
|
|
|
+ * CL 1.2 version that uses device fission.
|
|
|
+ */
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __CREATE_SUB_DEVICES_ERR CL_HPP_ERR_STR_(clCreateSubDevices)
|
|
|
+#else
|
|
|
+#define __CREATE_SUB_DEVICES_ERR CL_HPP_ERR_STR_(clCreateSubDevicesEXT)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+/**
|
|
|
+ * Deprecated APIs for 1.2
|
|
|
+ */
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+#define __ENQUEUE_MARKER_ERR CL_HPP_ERR_STR_(clEnqueueMarker)
|
|
|
+#define __ENQUEUE_WAIT_FOR_EVENTS_ERR CL_HPP_ERR_STR_(clEnqueueWaitForEvents)
|
|
|
+#define __ENQUEUE_BARRIER_ERR CL_HPP_ERR_STR_(clEnqueueBarrier)
|
|
|
+#define __UNLOAD_COMPILER_ERR CL_HPP_ERR_STR_(clUnloadCompiler)
|
|
|
+#define __CREATE_GL_TEXTURE_2D_ERR CL_HPP_ERR_STR_(clCreateFromGLTexture2D)
|
|
|
+#define __CREATE_GL_TEXTURE_3D_ERR CL_HPP_ERR_STR_(clCreateFromGLTexture3D)
|
|
|
+#define __CREATE_IMAGE2D_ERR CL_HPP_ERR_STR_(clCreateImage2D)
|
|
|
+#define __CREATE_IMAGE3D_ERR CL_HPP_ERR_STR_(clCreateImage3D)
|
|
|
+#endif // #if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+
|
|
|
+/**
|
|
|
+ * Deprecated APIs for 2.0
|
|
|
+ */
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_2_APIS)
|
|
|
+#define __CREATE_COMMAND_QUEUE_ERR CL_HPP_ERR_STR_(clCreateCommandQueue)
|
|
|
+#define __ENQUEUE_TASK_ERR CL_HPP_ERR_STR_(clEnqueueTask)
|
|
|
+#define __CREATE_SAMPLER_ERR CL_HPP_ERR_STR_(clCreateSampler)
|
|
|
+#endif // #if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+
|
|
|
+/**
|
|
|
+ * CL 1.2 marker and barrier commands
|
|
|
+ */
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#define __ENQUEUE_MARKER_WAIT_LIST_ERR CL_HPP_ERR_STR_(clEnqueueMarkerWithWaitList)
|
|
|
+#define __ENQUEUE_BARRIER_WAIT_LIST_ERR CL_HPP_ERR_STR_(clEnqueueBarrierWithWaitList)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#define __CLONE_KERNEL_ERR CL_HPP_ERR_STR_(clCloneKernel)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+#endif // CL_HPP_USER_OVERRIDE_ERROR_STRINGS
|
|
|
+//! \endcond
|
|
|
+
|
|
|
+
|
|
|
+namespace detail {
|
|
|
+
|
|
|
+// Generic getInfoHelper. The final parameter is used to guide overload
|
|
|
+// resolution: the actual parameter passed is an int, which makes this
|
|
|
+// a worse conversion sequence than a specialization that declares the
|
|
|
+// parameter as an int.
|
|
|
+template<typename Functor, typename T>
|
|
|
+inline cl_int getInfoHelper(Functor f, cl_uint name, T* param, long)
|
|
|
+{
|
|
|
+ return f(name, sizeof(T), param, NULL);
|
|
|
+}
|
|
|
+
|
|
|
+// Specialized for getInfo<CL_PROGRAM_BINARIES>
|
|
|
+// Assumes that the output vector was correctly resized on the way in
|
|
|
+template <typename Func>
|
|
|
+inline cl_int getInfoHelper(Func f, cl_uint name, vector<vector<unsigned char>>* param, int)
|
|
|
+{
|
|
|
+ if (name != CL_PROGRAM_BINARIES) {
|
|
|
+ return CL_INVALID_VALUE;
|
|
|
+ }
|
|
|
+ if (param) {
|
|
|
+ // Create array of pointers, calculate total size and pass pointer array in
|
|
|
+ size_type numBinaries = param->size();
|
|
|
+ vector<unsigned char*> binariesPointers(numBinaries);
|
|
|
+
|
|
|
+ for (size_type i = 0; i < numBinaries; ++i)
|
|
|
+ {
|
|
|
+ binariesPointers[i] = (*param)[i].data();
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int err = f(name, numBinaries * sizeof(unsigned char*), binariesPointers.data(), NULL);
|
|
|
+
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+// Specialized getInfoHelper for vector params
|
|
|
+template <typename Func, typename T>
|
|
|
+inline cl_int getInfoHelper(Func f, cl_uint name, vector<T>* param, long)
|
|
|
+{
|
|
|
+ size_type required;
|
|
|
+ cl_int err = f(name, 0, NULL, &required);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ const size_type elements = required / sizeof(T);
|
|
|
+
|
|
|
+ // Temporary to avoid changing param on an error
|
|
|
+ vector<T> localData(elements);
|
|
|
+ err = f(name, required, localData.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ if (param) {
|
|
|
+ *param = std::move(localData);
|
|
|
+ }
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+/* Specialization for reference-counted types. This depends on the
|
|
|
+ * existence of Wrapper<T>::cl_type, and none of the other types having the
|
|
|
+ * cl_type member. Note that simplify specifying the parameter as Wrapper<T>
|
|
|
+ * does not work, because when using a derived type (e.g. Context) the generic
|
|
|
+ * template will provide a better match.
|
|
|
+ */
|
|
|
+template <typename Func, typename T>
|
|
|
+inline cl_int getInfoHelper(
|
|
|
+ Func f, cl_uint name, vector<T>* param, int, typename T::cl_type = 0)
|
|
|
+{
|
|
|
+ size_type required;
|
|
|
+ cl_int err = f(name, 0, NULL, &required);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ const size_type elements = required / sizeof(typename T::cl_type);
|
|
|
+
|
|
|
+ vector<typename T::cl_type> value(elements);
|
|
|
+ err = f(name, required, value.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (param) {
|
|
|
+ // Assign to convert CL type to T for each element
|
|
|
+ param->resize(elements);
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < elements; i++) {
|
|
|
+ (*param)[i] = T(value[i], true);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+// Specialized GetInfoHelper for string params
|
|
|
+template <typename Func>
|
|
|
+inline cl_int getInfoHelper(Func f, cl_uint name, string* param, long)
|
|
|
+{
|
|
|
+ size_type required;
|
|
|
+ cl_int err = f(name, 0, NULL, &required);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ // std::string has a constant data member
|
|
|
+ // a char vector does not
|
|
|
+ if (required > 0) {
|
|
|
+ vector<char> value(required);
|
|
|
+ err = f(name, required, value.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ if (param) {
|
|
|
+ param->assign(begin(value), prev(end(value)));
|
|
|
+ }
|
|
|
+ }
|
|
|
+ else if (param) {
|
|
|
+ param->assign("");
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+// Specialized GetInfoHelper for clsize_t params
|
|
|
+template <typename Func, size_type N>
|
|
|
+inline cl_int getInfoHelper(Func f, cl_uint name, array<size_type, N>* param, long)
|
|
|
+{
|
|
|
+ size_type required;
|
|
|
+ cl_int err = f(name, 0, NULL, &required);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type elements = required / sizeof(size_type);
|
|
|
+ vector<size_type> value(elements, 0);
|
|
|
+
|
|
|
+ err = f(name, required, value.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ // Bound the copy with N to prevent overruns
|
|
|
+ // if passed N > than the amount copied
|
|
|
+ if (elements > N) {
|
|
|
+ elements = N;
|
|
|
+ }
|
|
|
+ for (size_type i = 0; i < elements; ++i) {
|
|
|
+ (*param)[i] = value[i];
|
|
|
+ }
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+template<typename T> struct ReferenceHandler;
|
|
|
+
|
|
|
+/* Specialization for reference-counted types. This depends on the
|
|
|
+ * existence of Wrapper<T>::cl_type, and none of the other types having the
|
|
|
+ * cl_type member. Note that simplify specifying the parameter as Wrapper<T>
|
|
|
+ * does not work, because when using a derived type (e.g. Context) the generic
|
|
|
+ * template will provide a better match.
|
|
|
+ */
|
|
|
+template<typename Func, typename T>
|
|
|
+inline cl_int getInfoHelper(Func f, cl_uint name, T* param, int, typename T::cl_type = 0)
|
|
|
+{
|
|
|
+ typename T::cl_type value;
|
|
|
+ cl_int err = f(name, sizeof(value), &value, NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ *param = value;
|
|
|
+ if (value != NULL)
|
|
|
+ {
|
|
|
+ err = param->retain();
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_0_(F) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_PROFILE, string) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_VERSION, string) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_NAME, string) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_VENDOR, string) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_EXTENSIONS, string) \
|
|
|
+ \
|
|
|
+ F(cl_device_info, CL_DEVICE_TYPE, cl_device_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_VENDOR_ID, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_COMPUTE_UNITS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_WORK_ITEM_DIMENSIONS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_WORK_GROUP_SIZE, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_WORK_ITEM_SIZES, cl::vector<size_type>) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_CHAR, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_SHORT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_INT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_LONG, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_FLOAT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_DOUBLE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_CLOCK_FREQUENCY, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_ADDRESS_BITS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_READ_IMAGE_ARGS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_WRITE_IMAGE_ARGS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_MEM_ALLOC_SIZE, cl_ulong) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE2D_MAX_WIDTH, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE2D_MAX_HEIGHT, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE3D_MAX_WIDTH, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE3D_MAX_HEIGHT, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE3D_MAX_DEPTH, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IMAGE_SUPPORT, cl_bool) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_PARAMETER_SIZE, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_SAMPLERS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MEM_BASE_ADDR_ALIGN, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MIN_DATA_TYPE_ALIGN_SIZE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_SINGLE_FP_CONFIG, cl_device_fp_config) \
|
|
|
+ F(cl_device_info, CL_DEVICE_DOUBLE_FP_CONFIG, cl_device_fp_config) \
|
|
|
+ F(cl_device_info, CL_DEVICE_HALF_FP_CONFIG, cl_device_fp_config) \
|
|
|
+ F(cl_device_info, CL_DEVICE_GLOBAL_MEM_CACHE_TYPE, cl_device_mem_cache_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_GLOBAL_MEM_CACHELINE_SIZE, cl_uint)\
|
|
|
+ F(cl_device_info, CL_DEVICE_GLOBAL_MEM_CACHE_SIZE, cl_ulong) \
|
|
|
+ F(cl_device_info, CL_DEVICE_GLOBAL_MEM_SIZE, cl_ulong) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_CONSTANT_BUFFER_SIZE, cl_ulong) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_CONSTANT_ARGS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_LOCAL_MEM_TYPE, cl_device_local_mem_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_LOCAL_MEM_SIZE, cl_ulong) \
|
|
|
+ F(cl_device_info, CL_DEVICE_ERROR_CORRECTION_SUPPORT, cl_bool) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PROFILING_TIMER_RESOLUTION, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_ENDIAN_LITTLE, cl_bool) \
|
|
|
+ F(cl_device_info, CL_DEVICE_AVAILABLE, cl_bool) \
|
|
|
+ F(cl_device_info, CL_DEVICE_COMPILER_AVAILABLE, cl_bool) \
|
|
|
+ F(cl_device_info, CL_DEVICE_EXECUTION_CAPABILITIES, cl_device_exec_capabilities) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PLATFORM, cl_platform_id) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NAME, string) \
|
|
|
+ F(cl_device_info, CL_DEVICE_VENDOR, string) \
|
|
|
+ F(cl_device_info, CL_DRIVER_VERSION, string) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PROFILE, string) \
|
|
|
+ F(cl_device_info, CL_DEVICE_VERSION, string) \
|
|
|
+ F(cl_device_info, CL_DEVICE_EXTENSIONS, string) \
|
|
|
+ \
|
|
|
+ F(cl_context_info, CL_CONTEXT_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_context_info, CL_CONTEXT_DEVICES, cl::vector<Device>) \
|
|
|
+ F(cl_context_info, CL_CONTEXT_PROPERTIES, cl::vector<cl_context_properties>) \
|
|
|
+ \
|
|
|
+ F(cl_event_info, CL_EVENT_COMMAND_QUEUE, cl::CommandQueue) \
|
|
|
+ F(cl_event_info, CL_EVENT_COMMAND_TYPE, cl_command_type) \
|
|
|
+ F(cl_event_info, CL_EVENT_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_event_info, CL_EVENT_COMMAND_EXECUTION_STATUS, cl_int) \
|
|
|
+ \
|
|
|
+ F(cl_profiling_info, CL_PROFILING_COMMAND_QUEUED, cl_ulong) \
|
|
|
+ F(cl_profiling_info, CL_PROFILING_COMMAND_SUBMIT, cl_ulong) \
|
|
|
+ F(cl_profiling_info, CL_PROFILING_COMMAND_START, cl_ulong) \
|
|
|
+ F(cl_profiling_info, CL_PROFILING_COMMAND_END, cl_ulong) \
|
|
|
+ \
|
|
|
+ F(cl_mem_info, CL_MEM_TYPE, cl_mem_object_type) \
|
|
|
+ F(cl_mem_info, CL_MEM_FLAGS, cl_mem_flags) \
|
|
|
+ F(cl_mem_info, CL_MEM_SIZE, size_type) \
|
|
|
+ F(cl_mem_info, CL_MEM_HOST_PTR, void*) \
|
|
|
+ F(cl_mem_info, CL_MEM_MAP_COUNT, cl_uint) \
|
|
|
+ F(cl_mem_info, CL_MEM_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_mem_info, CL_MEM_CONTEXT, cl::Context) \
|
|
|
+ \
|
|
|
+ F(cl_image_info, CL_IMAGE_FORMAT, cl_image_format) \
|
|
|
+ F(cl_image_info, CL_IMAGE_ELEMENT_SIZE, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_ROW_PITCH, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_SLICE_PITCH, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_WIDTH, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_HEIGHT, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_DEPTH, size_type) \
|
|
|
+ \
|
|
|
+ F(cl_sampler_info, CL_SAMPLER_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_sampler_info, CL_SAMPLER_CONTEXT, cl::Context) \
|
|
|
+ F(cl_sampler_info, CL_SAMPLER_NORMALIZED_COORDS, cl_bool) \
|
|
|
+ F(cl_sampler_info, CL_SAMPLER_ADDRESSING_MODE, cl_addressing_mode) \
|
|
|
+ F(cl_sampler_info, CL_SAMPLER_FILTER_MODE, cl_filter_mode) \
|
|
|
+ \
|
|
|
+ F(cl_program_info, CL_PROGRAM_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_CONTEXT, cl::Context) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_NUM_DEVICES, cl_uint) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_DEVICES, cl::vector<Device>) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_SOURCE, string) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_BINARY_SIZES, cl::vector<size_type>) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_BINARIES, cl::vector<cl::vector<unsigned char>>) \
|
|
|
+ \
|
|
|
+ F(cl_program_build_info, CL_PROGRAM_BUILD_STATUS, cl_build_status) \
|
|
|
+ F(cl_program_build_info, CL_PROGRAM_BUILD_OPTIONS, string) \
|
|
|
+ F(cl_program_build_info, CL_PROGRAM_BUILD_LOG, string) \
|
|
|
+ \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_FUNCTION_NAME, string) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_NUM_ARGS, cl_uint) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_CONTEXT, cl::Context) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_PROGRAM, cl::Program) \
|
|
|
+ \
|
|
|
+ F(cl_kernel_work_group_info, CL_KERNEL_WORK_GROUP_SIZE, size_type) \
|
|
|
+ F(cl_kernel_work_group_info, CL_KERNEL_COMPILE_WORK_GROUP_SIZE, cl::detail::size_t_array) \
|
|
|
+ F(cl_kernel_work_group_info, CL_KERNEL_LOCAL_MEM_SIZE, cl_ulong) \
|
|
|
+ \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_CONTEXT, cl::Context) \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_DEVICE, cl::Device) \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_PROPERTIES, cl_command_queue_properties)
|
|
|
+
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_1_(F) \
|
|
|
+ F(cl_context_info, CL_CONTEXT_NUM_DEVICES, cl_uint)\
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_VECTOR_WIDTH_HALF, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_CHAR, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_SHORT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_INT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_LONG, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_FLOAT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_DOUBLE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_NATIVE_VECTOR_WIDTH_HALF, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_OPENCL_C_VERSION, string) \
|
|
|
+ \
|
|
|
+ F(cl_mem_info, CL_MEM_ASSOCIATED_MEMOBJECT, cl::Memory) \
|
|
|
+ F(cl_mem_info, CL_MEM_OFFSET, size_type) \
|
|
|
+ \
|
|
|
+ F(cl_kernel_work_group_info, CL_KERNEL_PREFERRED_WORK_GROUP_SIZE_MULTIPLE, size_type) \
|
|
|
+ F(cl_kernel_work_group_info, CL_KERNEL_PRIVATE_MEM_SIZE, cl_ulong) \
|
|
|
+ \
|
|
|
+ F(cl_event_info, CL_EVENT_CONTEXT, cl::Context)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_2_(F) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_NUM_KERNELS, size_type) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_KERNEL_NAMES, string) \
|
|
|
+ \
|
|
|
+ F(cl_program_build_info, CL_PROGRAM_BINARY_TYPE, cl_program_binary_type) \
|
|
|
+ \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_ATTRIBUTES, string) \
|
|
|
+ \
|
|
|
+ F(cl_kernel_arg_info, CL_KERNEL_ARG_ADDRESS_QUALIFIER, cl_kernel_arg_address_qualifier) \
|
|
|
+ F(cl_kernel_arg_info, CL_KERNEL_ARG_ACCESS_QUALIFIER, cl_kernel_arg_access_qualifier) \
|
|
|
+ F(cl_kernel_arg_info, CL_KERNEL_ARG_TYPE_NAME, string) \
|
|
|
+ F(cl_kernel_arg_info, CL_KERNEL_ARG_NAME, string) \
|
|
|
+ F(cl_kernel_arg_info, CL_KERNEL_ARG_TYPE_QUALIFIER, cl_kernel_arg_type_qualifier) \
|
|
|
+ \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARENT_DEVICE, cl::Device) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARTITION_PROPERTIES, cl::vector<cl_device_partition_property>) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARTITION_TYPE, cl::vector<cl_device_partition_property>) \
|
|
|
+ F(cl_device_info, CL_DEVICE_REFERENCE_COUNT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_INTEROP_USER_SYNC, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARTITION_AFFINITY_DOMAIN, cl_device_affinity_domain) \
|
|
|
+ F(cl_device_info, CL_DEVICE_BUILT_IN_KERNELS, string) \
|
|
|
+ \
|
|
|
+ F(cl_image_info, CL_IMAGE_ARRAY_SIZE, size_type) \
|
|
|
+ F(cl_image_info, CL_IMAGE_NUM_MIP_LEVELS, cl_uint) \
|
|
|
+ F(cl_image_info, CL_IMAGE_NUM_SAMPLES, cl_uint)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_2_0_(F) \
|
|
|
+ F(cl_device_info, CL_DEVICE_QUEUE_ON_HOST_PROPERTIES, cl_command_queue_properties) \
|
|
|
+ F(cl_device_info, CL_DEVICE_QUEUE_ON_DEVICE_PROPERTIES, cl_command_queue_properties) \
|
|
|
+ F(cl_device_info, CL_DEVICE_QUEUE_ON_DEVICE_PREFERRED_SIZE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_QUEUE_ON_DEVICE_MAX_SIZE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_ON_DEVICE_QUEUES, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_ON_DEVICE_EVENTS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_PIPE_ARGS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PIPE_MAX_ACTIVE_RESERVATIONS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PIPE_MAX_PACKET_SIZE, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_SVM_CAPABILITIES, cl_device_svm_capabilities) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_PLATFORM_ATOMIC_ALIGNMENT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_GLOBAL_ATOMIC_ALIGNMENT, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PREFERRED_LOCAL_ATOMIC_ALIGNMENT, cl_uint) \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_SIZE, cl_uint) \
|
|
|
+ F(cl_mem_info, CL_MEM_USES_SVM_POINTER, cl_bool) \
|
|
|
+ F(cl_program_build_info, CL_PROGRAM_BUILD_GLOBAL_VARIABLE_TOTAL_SIZE, size_type) \
|
|
|
+ F(cl_pipe_info, CL_PIPE_PACKET_SIZE, cl_uint) \
|
|
|
+ F(cl_pipe_info, CL_PIPE_MAX_PACKETS, cl_uint)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_SUBGROUP_KHR_(F) \
|
|
|
+ F(cl_kernel_sub_group_info, CL_KERNEL_MAX_SUB_GROUP_SIZE_FOR_NDRANGE_KHR, size_type) \
|
|
|
+ F(cl_kernel_sub_group_info, CL_KERNEL_SUB_GROUP_COUNT_FOR_NDRANGE_KHR, size_type)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_IL_KHR_(F) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IL_VERSION_KHR, string) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_IL_KHR, cl::vector<unsigned char>)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_2_1_(F) \
|
|
|
+ F(cl_platform_info, CL_PLATFORM_HOST_TIMER_RESOLUTION, size_type) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_IL, cl::vector<unsigned char>) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_MAX_NUM_SUB_GROUPS, size_type) \
|
|
|
+ F(cl_kernel_info, CL_KERNEL_COMPILE_NUM_SUB_GROUPS, size_type) \
|
|
|
+ F(cl_device_info, CL_DEVICE_MAX_NUM_SUB_GROUPS, cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_IL_VERSION, string) \
|
|
|
+ F(cl_device_info, CL_DEVICE_SUB_GROUP_INDEPENDENT_FORWARD_PROGRESS, cl_bool) \
|
|
|
+ F(cl_command_queue_info, CL_QUEUE_DEVICE_DEFAULT, cl::DeviceCommandQueue) \
|
|
|
+ F(cl_kernel_sub_group_info, CL_KERNEL_MAX_SUB_GROUP_SIZE_FOR_NDRANGE, size_type) \
|
|
|
+ F(cl_kernel_sub_group_info, CL_KERNEL_SUB_GROUP_COUNT_FOR_NDRANGE, size_type) \
|
|
|
+ F(cl_kernel_sub_group_info, CL_KERNEL_LOCAL_SIZE_FOR_SUB_GROUP_COUNT, cl::detail::size_t_array)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_2_2_(F) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_SCOPE_GLOBAL_CTORS_PRESENT, cl_bool) \
|
|
|
+ F(cl_program_info, CL_PROGRAM_SCOPE_GLOBAL_DTORS_PRESENT, cl_bool)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_DEVICE_FISSION_(F) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARENT_DEVICE_EXT, cl_device_id) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARTITION_TYPES_EXT, cl::vector<cl_device_partition_property_ext>) \
|
|
|
+ F(cl_device_info, CL_DEVICE_AFFINITY_DOMAINS_EXT, cl::vector<cl_device_partition_property_ext>) \
|
|
|
+ F(cl_device_info, CL_DEVICE_REFERENCE_COUNT_EXT , cl_uint) \
|
|
|
+ F(cl_device_info, CL_DEVICE_PARTITION_STYLE_EXT, cl::vector<cl_device_partition_property_ext>)
|
|
|
+
|
|
|
+template <typename enum_type, cl_int Name>
|
|
|
+struct param_traits {};
|
|
|
+
|
|
|
+#define CL_HPP_DECLARE_PARAM_TRAITS_(token, param_name, T) \
|
|
|
+struct token; \
|
|
|
+template<> \
|
|
|
+struct param_traits<detail:: token,param_name> \
|
|
|
+{ \
|
|
|
+ enum { value = param_name }; \
|
|
|
+ typedef T param_type; \
|
|
|
+};
|
|
|
+
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_0_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_1_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_2_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+CL_HPP_PARAM_NAME_INFO_2_0_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+CL_HPP_PARAM_NAME_INFO_2_1_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+CL_HPP_PARAM_NAME_INFO_2_2_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_CL_SUB_GROUPS_KHR) && CL_HPP_TARGET_OPENCL_VERSION < 210
|
|
|
+CL_HPP_PARAM_NAME_INFO_SUBGROUP_KHR_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // #if defined(CL_HPP_USE_CL_SUB_GROUPS_KHR) && CL_HPP_TARGET_OPENCL_VERSION < 210
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_IL_KHR)
|
|
|
+CL_HPP_PARAM_NAME_INFO_IL_KHR_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // #if defined(CL_HPP_USE_IL_KHR)
|
|
|
+
|
|
|
+
|
|
|
+// Flags deprecated in OpenCL 2.0
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_0_DEPRECATED_IN_2_0_(F) \
|
|
|
+ F(cl_device_info, CL_DEVICE_QUEUE_PROPERTIES, cl_command_queue_properties)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_1_DEPRECATED_IN_2_0_(F) \
|
|
|
+ F(cl_device_info, CL_DEVICE_HOST_UNIFIED_MEMORY, cl_bool)
|
|
|
+
|
|
|
+#define CL_HPP_PARAM_NAME_INFO_1_2_DEPRECATED_IN_2_0_(F) \
|
|
|
+ F(cl_image_info, CL_IMAGE_BUFFER, cl::Buffer)
|
|
|
+
|
|
|
+// Include deprecated query flags based on versions
|
|
|
+// Only include deprecated 1.0 flags if 2.0 not active as there is an enum clash
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION > 100 && CL_HPP_MINIMUM_OPENCL_VERSION < 200 && CL_HPP_TARGET_OPENCL_VERSION < 200
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_0_DEPRECATED_IN_2_0_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 110
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION > 110 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_1_DEPRECATED_IN_2_0_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION > 120 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+CL_HPP_PARAM_NAME_INFO_1_2_DEPRECATED_IN_2_0_(CL_HPP_DECLARE_PARAM_TRAITS_)
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_CL_DEVICE_FISSION)
|
|
|
+CL_HPP_PARAM_NAME_DEVICE_FISSION_(CL_HPP_DECLARE_PARAM_TRAITS_);
|
|
|
+#endif // CL_HPP_USE_CL_DEVICE_FISSION
|
|
|
+
|
|
|
+#ifdef CL_PLATFORM_ICD_SUFFIX_KHR
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_platform_info, CL_PLATFORM_ICD_SUFFIX_KHR, string)
|
|
|
+#endif
|
|
|
+
|
|
|
+#ifdef CL_DEVICE_PROFILING_TIMER_OFFSET_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_PROFILING_TIMER_OFFSET_AMD, cl_ulong)
|
|
|
+#endif
|
|
|
+
|
|
|
+#ifdef CL_DEVICE_GLOBAL_FREE_MEMORY_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_GLOBAL_FREE_MEMORY_AMD, vector<size_type>)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_SIMD_PER_COMPUTE_UNIT_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_SIMD_PER_COMPUTE_UNIT_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_SIMD_WIDTH_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_SIMD_WIDTH_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_SIMD_INSTRUCTION_WIDTH_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_SIMD_INSTRUCTION_WIDTH_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_WAVEFRONT_WIDTH_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_WAVEFRONT_WIDTH_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_GLOBAL_MEM_CHANNELS_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_GLOBAL_MEM_CHANNELS_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_GLOBAL_MEM_CHANNEL_BANKS_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_GLOBAL_MEM_CHANNEL_BANKS_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_GLOBAL_MEM_CHANNEL_BANK_WIDTH_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_GLOBAL_MEM_CHANNEL_BANK_WIDTH_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_LOCAL_MEM_SIZE_PER_COMPUTE_UNIT_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_LOCAL_MEM_SIZE_PER_COMPUTE_UNIT_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_LOCAL_MEM_BANKS_AMD
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_LOCAL_MEM_BANKS_AMD, cl_uint)
|
|
|
+#endif
|
|
|
+
|
|
|
+#ifdef CL_DEVICE_COMPUTE_UNITS_BITFIELD_ARM
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_COMPUTE_UNITS_BITFIELD_ARM, cl_ulong)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_JOB_SLOTS_ARM
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_JOB_SLOTS_ARM, cl_uint)
|
|
|
+#endif
|
|
|
+
|
|
|
+#ifdef CL_DEVICE_COMPUTE_CAPABILITY_MAJOR_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_COMPUTE_CAPABILITY_MAJOR_NV, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_COMPUTE_CAPABILITY_MINOR_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_COMPUTE_CAPABILITY_MINOR_NV, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_REGISTERS_PER_BLOCK_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_REGISTERS_PER_BLOCK_NV, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_WARP_SIZE_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_WARP_SIZE_NV, cl_uint)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_GPU_OVERLAP_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_GPU_OVERLAP_NV, cl_bool)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_KERNEL_EXEC_TIMEOUT_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_KERNEL_EXEC_TIMEOUT_NV, cl_bool)
|
|
|
+#endif
|
|
|
+#ifdef CL_DEVICE_INTEGRATED_MEMORY_NV
|
|
|
+CL_HPP_DECLARE_PARAM_TRAITS_(cl_device_info, CL_DEVICE_INTEGRATED_MEMORY_NV, cl_bool)
|
|
|
+#endif
|
|
|
+
|
|
|
+// Convenience functions
|
|
|
+
|
|
|
+template <typename Func, typename T>
|
|
|
+inline cl_int
|
|
|
+getInfo(Func f, cl_uint name, T* param)
|
|
|
+{
|
|
|
+ return getInfoHelper(f, name, param, 0);
|
|
|
+}
|
|
|
+
|
|
|
+template <typename Func, typename Arg0>
|
|
|
+struct GetInfoFunctor0
|
|
|
+{
|
|
|
+ Func f_; const Arg0& arg0_;
|
|
|
+ cl_int operator ()(
|
|
|
+ cl_uint param, size_type size, void* value, size_type* size_ret)
|
|
|
+ { return f_(arg0_, param, size, value, size_ret); }
|
|
|
+};
|
|
|
+
|
|
|
+template <typename Func, typename Arg0, typename Arg1>
|
|
|
+struct GetInfoFunctor1
|
|
|
+{
|
|
|
+ Func f_; const Arg0& arg0_; const Arg1& arg1_;
|
|
|
+ cl_int operator ()(
|
|
|
+ cl_uint param, size_type size, void* value, size_type* size_ret)
|
|
|
+ { return f_(arg0_, arg1_, param, size, value, size_ret); }
|
|
|
+};
|
|
|
+
|
|
|
+template <typename Func, typename Arg0, typename T>
|
|
|
+inline cl_int
|
|
|
+getInfo(Func f, const Arg0& arg0, cl_uint name, T* param)
|
|
|
+{
|
|
|
+ GetInfoFunctor0<Func, Arg0> f0 = { f, arg0 };
|
|
|
+ return getInfoHelper(f0, name, param, 0);
|
|
|
+}
|
|
|
+
|
|
|
+template <typename Func, typename Arg0, typename Arg1, typename T>
|
|
|
+inline cl_int
|
|
|
+getInfo(Func f, const Arg0& arg0, const Arg1& arg1, cl_uint name, T* param)
|
|
|
+{
|
|
|
+ GetInfoFunctor1<Func, Arg0, Arg1> f0 = { f, arg0, arg1 };
|
|
|
+ return getInfoHelper(f0, name, param, 0);
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+template<typename T>
|
|
|
+struct ReferenceHandler
|
|
|
+{ };
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+/**
|
|
|
+ * OpenCL 1.2 devices do have retain/release.
|
|
|
+ */
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_device_id>
|
|
|
+{
|
|
|
+ /**
|
|
|
+ * Retain the device.
|
|
|
+ * \param device A valid device created using createSubDevices
|
|
|
+ * \return
|
|
|
+ * CL_SUCCESS if the function executed successfully.
|
|
|
+ * CL_INVALID_DEVICE if device was not a valid subdevice
|
|
|
+ * CL_OUT_OF_RESOURCES
|
|
|
+ * CL_OUT_OF_HOST_MEMORY
|
|
|
+ */
|
|
|
+ static cl_int retain(cl_device_id device)
|
|
|
+ { return ::clRetainDevice(device); }
|
|
|
+ /**
|
|
|
+ * Retain the device.
|
|
|
+ * \param device A valid device created using createSubDevices
|
|
|
+ * \return
|
|
|
+ * CL_SUCCESS if the function executed successfully.
|
|
|
+ * CL_INVALID_DEVICE if device was not a valid subdevice
|
|
|
+ * CL_OUT_OF_RESOURCES
|
|
|
+ * CL_OUT_OF_HOST_MEMORY
|
|
|
+ */
|
|
|
+ static cl_int release(cl_device_id device)
|
|
|
+ { return ::clReleaseDevice(device); }
|
|
|
+};
|
|
|
+#else // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+/**
|
|
|
+ * OpenCL 1.1 devices do not have retain/release.
|
|
|
+ */
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_device_id>
|
|
|
+{
|
|
|
+ // cl_device_id does not have retain().
|
|
|
+ static cl_int retain(cl_device_id)
|
|
|
+ { return CL_SUCCESS; }
|
|
|
+ // cl_device_id does not have release().
|
|
|
+ static cl_int release(cl_device_id)
|
|
|
+ { return CL_SUCCESS; }
|
|
|
+};
|
|
|
+#endif // ! (CL_HPP_TARGET_OPENCL_VERSION >= 120)
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_platform_id>
|
|
|
+{
|
|
|
+ // cl_platform_id does not have retain().
|
|
|
+ static cl_int retain(cl_platform_id)
|
|
|
+ { return CL_SUCCESS; }
|
|
|
+ // cl_platform_id does not have release().
|
|
|
+ static cl_int release(cl_platform_id)
|
|
|
+ { return CL_SUCCESS; }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_context>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_context context)
|
|
|
+ { return ::clRetainContext(context); }
|
|
|
+ static cl_int release(cl_context context)
|
|
|
+ { return ::clReleaseContext(context); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_command_queue>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_command_queue queue)
|
|
|
+ { return ::clRetainCommandQueue(queue); }
|
|
|
+ static cl_int release(cl_command_queue queue)
|
|
|
+ { return ::clReleaseCommandQueue(queue); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_mem>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_mem memory)
|
|
|
+ { return ::clRetainMemObject(memory); }
|
|
|
+ static cl_int release(cl_mem memory)
|
|
|
+ { return ::clReleaseMemObject(memory); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_sampler>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_sampler sampler)
|
|
|
+ { return ::clRetainSampler(sampler); }
|
|
|
+ static cl_int release(cl_sampler sampler)
|
|
|
+ { return ::clReleaseSampler(sampler); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_program>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_program program)
|
|
|
+ { return ::clRetainProgram(program); }
|
|
|
+ static cl_int release(cl_program program)
|
|
|
+ { return ::clReleaseProgram(program); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_kernel>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_kernel kernel)
|
|
|
+ { return ::clRetainKernel(kernel); }
|
|
|
+ static cl_int release(cl_kernel kernel)
|
|
|
+ { return ::clReleaseKernel(kernel); }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+struct ReferenceHandler<cl_event>
|
|
|
+{
|
|
|
+ static cl_int retain(cl_event event)
|
|
|
+ { return ::clRetainEvent(event); }
|
|
|
+ static cl_int release(cl_event event)
|
|
|
+ { return ::clReleaseEvent(event); }
|
|
|
+};
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120 && CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+// Extracts version number with major in the upper 16 bits, minor in the lower 16
|
|
|
+static cl_uint getVersion(const vector<char> &versionInfo)
|
|
|
+{
|
|
|
+ int highVersion = 0;
|
|
|
+ int lowVersion = 0;
|
|
|
+ int index = 7;
|
|
|
+ while(versionInfo[index] != '.' ) {
|
|
|
+ highVersion *= 10;
|
|
|
+ highVersion += versionInfo[index]-'0';
|
|
|
+ ++index;
|
|
|
+ }
|
|
|
+ ++index;
|
|
|
+ while(versionInfo[index] != ' ' && versionInfo[index] != '\0') {
|
|
|
+ lowVersion *= 10;
|
|
|
+ lowVersion += versionInfo[index]-'0';
|
|
|
+ ++index;
|
|
|
+ }
|
|
|
+ return (highVersion << 16) | lowVersion;
|
|
|
+}
|
|
|
+
|
|
|
+static cl_uint getPlatformVersion(cl_platform_id platform)
|
|
|
+{
|
|
|
+ size_type size = 0;
|
|
|
+ clGetPlatformInfo(platform, CL_PLATFORM_VERSION, 0, NULL, &size);
|
|
|
+
|
|
|
+ vector<char> versionInfo(size);
|
|
|
+ clGetPlatformInfo(platform, CL_PLATFORM_VERSION, size, versionInfo.data(), &size);
|
|
|
+ return getVersion(versionInfo);
|
|
|
+}
|
|
|
+
|
|
|
+static cl_uint getDevicePlatformVersion(cl_device_id device)
|
|
|
+{
|
|
|
+ cl_platform_id platform;
|
|
|
+ clGetDeviceInfo(device, CL_DEVICE_PLATFORM, sizeof(platform), &platform, NULL);
|
|
|
+ return getPlatformVersion(platform);
|
|
|
+}
|
|
|
+
|
|
|
+static cl_uint getContextPlatformVersion(cl_context context)
|
|
|
+{
|
|
|
+ // The platform cannot be queried directly, so we first have to grab a
|
|
|
+ // device and obtain its context
|
|
|
+ size_type size = 0;
|
|
|
+ clGetContextInfo(context, CL_CONTEXT_DEVICES, 0, NULL, &size);
|
|
|
+ if (size == 0)
|
|
|
+ return 0;
|
|
|
+ vector<cl_device_id> devices(size/sizeof(cl_device_id));
|
|
|
+ clGetContextInfo(context, CL_CONTEXT_DEVICES, size, devices.data(), NULL);
|
|
|
+ return getDevicePlatformVersion(devices[0]);
|
|
|
+}
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120 && CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+
|
|
|
+template <typename T>
|
|
|
+class Wrapper
|
|
|
+{
|
|
|
+public:
|
|
|
+ typedef T cl_type;
|
|
|
+
|
|
|
+protected:
|
|
|
+ cl_type object_;
|
|
|
+
|
|
|
+public:
|
|
|
+ Wrapper() : object_(NULL) { }
|
|
|
+
|
|
|
+ Wrapper(const cl_type &obj, bool retainObject) : object_(obj)
|
|
|
+ {
|
|
|
+ if (retainObject) {
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ ~Wrapper()
|
|
|
+ {
|
|
|
+ if (object_ != NULL) { release(); }
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper(const Wrapper<cl_type>& rhs)
|
|
|
+ {
|
|
|
+ object_ = rhs.object_;
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper(Wrapper<cl_type>&& rhs) CL_HPP_NOEXCEPT_
|
|
|
+ {
|
|
|
+ object_ = rhs.object_;
|
|
|
+ rhs.object_ = NULL;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (const Wrapper<cl_type>& rhs)
|
|
|
+ {
|
|
|
+ if (this != &rhs) {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs.object_;
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (Wrapper<cl_type>&& rhs)
|
|
|
+ {
|
|
|
+ if (this != &rhs) {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs.object_;
|
|
|
+ rhs.object_ = NULL;
|
|
|
+ }
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (const cl_type &rhs)
|
|
|
+ {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs;
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ const cl_type& operator ()() const { return object_; }
|
|
|
+
|
|
|
+ cl_type& operator ()() { return object_; }
|
|
|
+
|
|
|
+ cl_type get() const { return object_; }
|
|
|
+
|
|
|
+protected:
|
|
|
+ template<typename Func, typename U>
|
|
|
+ friend inline cl_int getInfoHelper(Func, cl_uint, U*, int, typename U::cl_type);
|
|
|
+
|
|
|
+ cl_int retain() const
|
|
|
+ {
|
|
|
+ if (object_ != nullptr) {
|
|
|
+ return ReferenceHandler<cl_type>::retain(object_);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int release() const
|
|
|
+ {
|
|
|
+ if (object_ != nullptr) {
|
|
|
+ return ReferenceHandler<cl_type>::release(object_);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template <>
|
|
|
+class Wrapper<cl_device_id>
|
|
|
+{
|
|
|
+public:
|
|
|
+ typedef cl_device_id cl_type;
|
|
|
+
|
|
|
+protected:
|
|
|
+ cl_type object_;
|
|
|
+ bool referenceCountable_;
|
|
|
+
|
|
|
+ static bool isReferenceCountable(cl_device_id device)
|
|
|
+ {
|
|
|
+ bool retVal = false;
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ if (device != NULL) {
|
|
|
+ int version = getDevicePlatformVersion(device);
|
|
|
+ if(version > ((1 << 16) + 1)) {
|
|
|
+ retVal = true;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#else // CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ retVal = true;
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ return retVal;
|
|
|
+ }
|
|
|
+
|
|
|
+public:
|
|
|
+ Wrapper() : object_(NULL), referenceCountable_(false)
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper(const cl_type &obj, bool retainObject) :
|
|
|
+ object_(obj),
|
|
|
+ referenceCountable_(false)
|
|
|
+ {
|
|
|
+ referenceCountable_ = isReferenceCountable(obj);
|
|
|
+
|
|
|
+ if (retainObject) {
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ ~Wrapper()
|
|
|
+ {
|
|
|
+ release();
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper(const Wrapper<cl_type>& rhs)
|
|
|
+ {
|
|
|
+ object_ = rhs.object_;
|
|
|
+ referenceCountable_ = isReferenceCountable(object_);
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper(Wrapper<cl_type>&& rhs) CL_HPP_NOEXCEPT_
|
|
|
+ {
|
|
|
+ object_ = rhs.object_;
|
|
|
+ referenceCountable_ = rhs.referenceCountable_;
|
|
|
+ rhs.object_ = NULL;
|
|
|
+ rhs.referenceCountable_ = false;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (const Wrapper<cl_type>& rhs)
|
|
|
+ {
|
|
|
+ if (this != &rhs) {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs.object_;
|
|
|
+ referenceCountable_ = rhs.referenceCountable_;
|
|
|
+ detail::errHandler(retain(), __RETAIN_ERR);
|
|
|
+ }
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (Wrapper<cl_type>&& rhs)
|
|
|
+ {
|
|
|
+ if (this != &rhs) {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs.object_;
|
|
|
+ referenceCountable_ = rhs.referenceCountable_;
|
|
|
+ rhs.object_ = NULL;
|
|
|
+ rhs.referenceCountable_ = false;
|
|
|
+ }
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ Wrapper<cl_type>& operator = (const cl_type &rhs)
|
|
|
+ {
|
|
|
+ detail::errHandler(release(), __RELEASE_ERR);
|
|
|
+ object_ = rhs;
|
|
|
+ referenceCountable_ = isReferenceCountable(object_);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ const cl_type& operator ()() const { return object_; }
|
|
|
+
|
|
|
+ cl_type& operator ()() { return object_; }
|
|
|
+
|
|
|
+ cl_type get() const { return object_; }
|
|
|
+
|
|
|
+protected:
|
|
|
+ template<typename Func, typename U>
|
|
|
+ friend inline cl_int getInfoHelper(Func, cl_uint, U*, int, typename U::cl_type);
|
|
|
+
|
|
|
+ template<typename Func, typename U>
|
|
|
+ friend inline cl_int getInfoHelper(Func, cl_uint, vector<U>*, int, typename U::cl_type);
|
|
|
+
|
|
|
+ cl_int retain() const
|
|
|
+ {
|
|
|
+ if( object_ != nullptr && referenceCountable_ ) {
|
|
|
+ return ReferenceHandler<cl_type>::retain(object_);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int release() const
|
|
|
+ {
|
|
|
+ if (object_ != nullptr && referenceCountable_) {
|
|
|
+ return ReferenceHandler<cl_type>::release(object_);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template <typename T>
|
|
|
+inline bool operator==(const Wrapper<T> &lhs, const Wrapper<T> &rhs)
|
|
|
+{
|
|
|
+ return lhs() == rhs();
|
|
|
+}
|
|
|
+
|
|
|
+template <typename T>
|
|
|
+inline bool operator!=(const Wrapper<T> &lhs, const Wrapper<T> &rhs)
|
|
|
+{
|
|
|
+ return !operator==(lhs, rhs);
|
|
|
+}
|
|
|
+
|
|
|
+} // namespace detail
|
|
|
+//! \endcond
|
|
|
+
|
|
|
+
|
|
|
+using BuildLogType = vector<std::pair<cl::Device, typename detail::param_traits<detail::cl_program_build_info, CL_PROGRAM_BUILD_LOG>::param_type>>;
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+/**
|
|
|
+* Exception class for build errors to carry build info
|
|
|
+*/
|
|
|
+class BuildError : public Error
|
|
|
+{
|
|
|
+private:
|
|
|
+ BuildLogType buildLogs;
|
|
|
+public:
|
|
|
+ BuildError(cl_int err, const char * errStr, const BuildLogType &vec) : Error(err, errStr), buildLogs(vec)
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ BuildLogType getBuildLog() const
|
|
|
+ {
|
|
|
+ return buildLogs;
|
|
|
+ }
|
|
|
+};
|
|
|
+namespace detail {
|
|
|
+ static inline cl_int buildErrHandler(
|
|
|
+ cl_int err,
|
|
|
+ const char * errStr,
|
|
|
+ const BuildLogType &buildLogs)
|
|
|
+ {
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ throw BuildError(err, errStr, buildLogs);
|
|
|
+ }
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+} // namespace detail
|
|
|
+
|
|
|
+#else
|
|
|
+namespace detail {
|
|
|
+ static inline cl_int buildErrHandler(
|
|
|
+ cl_int err,
|
|
|
+ const char * errStr,
|
|
|
+ const BuildLogType &buildLogs)
|
|
|
+ {
|
|
|
+ (void)buildLogs; // suppress unused variable warning
|
|
|
+ (void)errStr;
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+} // namespace detail
|
|
|
+#endif // #if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+
|
|
|
+
|
|
|
+/*! \stuct ImageFormat
|
|
|
+ * \brief Adds constructors and member functions for cl_image_format.
|
|
|
+ *
|
|
|
+ * \see cl_image_format
|
|
|
+ */
|
|
|
+struct ImageFormat : public cl_image_format
|
|
|
+{
|
|
|
+ //! \brief Default constructor - performs no initialization.
|
|
|
+ ImageFormat(){}
|
|
|
+
|
|
|
+ //! \brief Initializing constructor.
|
|
|
+ ImageFormat(cl_channel_order order, cl_channel_type type)
|
|
|
+ {
|
|
|
+ image_channel_order = order;
|
|
|
+ image_channel_data_type = type;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Assignment operator.
|
|
|
+ ImageFormat& operator = (const ImageFormat& rhs)
|
|
|
+ {
|
|
|
+ if (this != &rhs) {
|
|
|
+ this->image_channel_data_type = rhs.image_channel_data_type;
|
|
|
+ this->image_channel_order = rhs.image_channel_order;
|
|
|
+ }
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_device_id.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are inexpensive, since they don't 'own'
|
|
|
+ * any underlying resources or data structures.
|
|
|
+ *
|
|
|
+ * \see cl_device_id
|
|
|
+ */
|
|
|
+class Device : public detail::Wrapper<cl_device_id>
|
|
|
+{
|
|
|
+private:
|
|
|
+ static std::once_flag default_initialized_;
|
|
|
+ static Device default_;
|
|
|
+ static cl_int default_error_;
|
|
|
+
|
|
|
+ /*! \brief Create the default context.
|
|
|
+ *
|
|
|
+ * This sets @c default_ and @c default_error_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefault();
|
|
|
+
|
|
|
+ /*! \brief Create the default platform from a provided platform.
|
|
|
+ *
|
|
|
+ * This sets @c default_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefaultProvided(const Device &p) {
|
|
|
+ default_ = p;
|
|
|
+ }
|
|
|
+
|
|
|
+public:
|
|
|
+#ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+ /*! \brief Reset the default.
|
|
|
+ *
|
|
|
+ * This sets @c default_ to an empty value to support cleanup in
|
|
|
+ * the unit test framework.
|
|
|
+ * This function is not thread safe.
|
|
|
+ */
|
|
|
+ static void unitTestClearDefault() {
|
|
|
+ default_ = Device();
|
|
|
+ }
|
|
|
+#endif // #ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Device() : detail::Wrapper<cl_type>() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_device_id.
|
|
|
+ *
|
|
|
+ * This simply copies the device ID value, which is an inexpensive operation.
|
|
|
+ */
|
|
|
+ explicit Device(const cl_device_id &device, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(device, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Returns the first device on the default context.
|
|
|
+ *
|
|
|
+ * \see Context::getDefault()
|
|
|
+ */
|
|
|
+ static Device getDefault(
|
|
|
+ cl_int *errResult = NULL)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefault);
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ if (errResult != NULL) {
|
|
|
+ *errResult = default_error_;
|
|
|
+ }
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Modify the default device to be used by
|
|
|
+ * subsequent operations.
|
|
|
+ * Will only set the default if no default was previously created.
|
|
|
+ * @return updated default device.
|
|
|
+ * Should be compared to the passed value to ensure that it was updated.
|
|
|
+ */
|
|
|
+ static Device setDefault(const Device &default_device)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefaultProvided, std::cref(default_device));
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_device_id.
|
|
|
+ *
|
|
|
+ * This simply copies the device ID value, which is an inexpensive operation.
|
|
|
+ */
|
|
|
+ Device& operator = (const cl_device_id& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Device(const Device& dev) : detail::Wrapper<cl_type>(dev) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Device& operator = (const Device &dev)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(dev);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Device(Device&& dev) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(dev)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Device& operator = (Device &&dev)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(dev));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetDeviceInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_device_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetDeviceInfo, object_, name, param),
|
|
|
+ __GET_DEVICE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetDeviceInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_device_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_device_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ /**
|
|
|
+ * Return the current value of the host clock as seen by the device.
|
|
|
+ * The resolution of the device timer may be queried with the
|
|
|
+ * CL_DEVICE_PROFILING_TIMER_RESOLUTION query.
|
|
|
+ * @return The host timer value.
|
|
|
+ */
|
|
|
+ cl_ulong getHostTimer(cl_int *error = nullptr)
|
|
|
+ {
|
|
|
+ cl_ulong retVal = 0;
|
|
|
+ cl_int err =
|
|
|
+ clGetHostTimer(this->get(), &retVal);
|
|
|
+ detail::errHandler(
|
|
|
+ err,
|
|
|
+ __GET_HOST_TIMER_ERR);
|
|
|
+ if (error) {
|
|
|
+ *error = err;
|
|
|
+ }
|
|
|
+ return retVal;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Return a synchronized pair of host and device timestamps as seen by device.
|
|
|
+ * Use to correlate the clocks and get the host timer only using getHostTimer
|
|
|
+ * as a lower cost mechanism in between calls.
|
|
|
+ * The resolution of the host timer may be queried with the
|
|
|
+ * CL_PLATFORM_HOST_TIMER_RESOLUTION query.
|
|
|
+ * The resolution of the device timer may be queried with the
|
|
|
+ * CL_DEVICE_PROFILING_TIMER_RESOLUTION query.
|
|
|
+ * @return A pair of (device timer, host timer) timer values.
|
|
|
+ */
|
|
|
+ std::pair<cl_ulong, cl_ulong> getDeviceAndHostTimer(cl_int *error = nullptr)
|
|
|
+ {
|
|
|
+ std::pair<cl_ulong, cl_ulong> retVal;
|
|
|
+ cl_int err =
|
|
|
+ clGetDeviceAndHostTimer(this->get(), &(retVal.first), &(retVal.second));
|
|
|
+ detail::errHandler(
|
|
|
+ err,
|
|
|
+ __GET_DEVICE_AND_HOST_TIMER_ERR);
|
|
|
+ if (error) {
|
|
|
+ *error = err;
|
|
|
+ }
|
|
|
+ return retVal;
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ /**
|
|
|
+ * CL 1.2 version
|
|
|
+ */
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ //! \brief Wrapper for clCreateSubDevices().
|
|
|
+ cl_int createSubDevices(
|
|
|
+ const cl_device_partition_property * properties,
|
|
|
+ vector<Device>* devices)
|
|
|
+ {
|
|
|
+ cl_uint n = 0;
|
|
|
+ cl_int err = clCreateSubDevices(object_, properties, 0, NULL, &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_SUB_DEVICES_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_device_id> ids(n);
|
|
|
+ err = clCreateSubDevices(object_, properties, n, ids.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_SUB_DEVICES_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ // Cannot trivially assign because we need to capture intermediates
|
|
|
+ // with safe construction
|
|
|
+ if (devices) {
|
|
|
+ devices->resize(ids.size());
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < ids.size(); i++) {
|
|
|
+ // We do not need to retain because this device is being created
|
|
|
+ // by the runtime
|
|
|
+ (*devices)[i] = Device(ids[i], false);
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+#elif defined(CL_HPP_USE_CL_DEVICE_FISSION)
|
|
|
+
|
|
|
+/**
|
|
|
+ * CL 1.1 version that uses device fission extension.
|
|
|
+ */
|
|
|
+ cl_int createSubDevices(
|
|
|
+ const cl_device_partition_property_ext * properties,
|
|
|
+ vector<Device>* devices)
|
|
|
+ {
|
|
|
+ typedef CL_API_ENTRY cl_int
|
|
|
+ ( CL_API_CALL * PFN_clCreateSubDevicesEXT)(
|
|
|
+ cl_device_id /*in_device*/,
|
|
|
+ const cl_device_partition_property_ext * /* properties */,
|
|
|
+ cl_uint /*num_entries*/,
|
|
|
+ cl_device_id * /*out_devices*/,
|
|
|
+ cl_uint * /*num_devices*/ ) CL_EXT_SUFFIX__VERSION_1_1;
|
|
|
+
|
|
|
+ static PFN_clCreateSubDevicesEXT pfn_clCreateSubDevicesEXT = NULL;
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clCreateSubDevicesEXT);
|
|
|
+
|
|
|
+ cl_uint n = 0;
|
|
|
+ cl_int err = pfn_clCreateSubDevicesEXT(object_, properties, 0, NULL, &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_SUB_DEVICES_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_device_id> ids(n);
|
|
|
+ err = pfn_clCreateSubDevicesEXT(object_, properties, n, ids.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_SUB_DEVICES_ERR);
|
|
|
+ }
|
|
|
+ // Cannot trivially assign because we need to capture intermediates
|
|
|
+ // with safe construction
|
|
|
+ if (devices) {
|
|
|
+ devices->resize(ids.size());
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < ids.size(); i++) {
|
|
|
+ // We do not need to retain because this device is being created
|
|
|
+ // by the runtime
|
|
|
+ (*devices)[i] = Device(ids[i], false);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+#endif // defined(CL_HPP_USE_CL_DEVICE_FISSION)
|
|
|
+};
|
|
|
+
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ std::once_flag Device::default_initialized_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ Device Device::default_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ cl_int Device::default_error_ = CL_SUCCESS;
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_platform_id.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are inexpensive, since they don't 'own'
|
|
|
+ * any underlying resources or data structures.
|
|
|
+ *
|
|
|
+ * \see cl_platform_id
|
|
|
+ */
|
|
|
+class Platform : public detail::Wrapper<cl_platform_id>
|
|
|
+{
|
|
|
+private:
|
|
|
+ static std::once_flag default_initialized_;
|
|
|
+ static Platform default_;
|
|
|
+ static cl_int default_error_;
|
|
|
+
|
|
|
+ /*! \brief Create the default context.
|
|
|
+ *
|
|
|
+ * This sets @c default_ and @c default_error_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefault() {
|
|
|
+ /* Throwing an exception from a call_once invocation does not do
|
|
|
+ * what we wish, so we catch it and save the error.
|
|
|
+ */
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ try
|
|
|
+#endif
|
|
|
+ {
|
|
|
+ // If default wasn't passed ,generate one
|
|
|
+ // Otherwise set it
|
|
|
+ cl_uint n = 0;
|
|
|
+
|
|
|
+ cl_int err = ::clGetPlatformIDs(0, NULL, &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ default_error_ = err;
|
|
|
+ return;
|
|
|
+ }
|
|
|
+ if (n == 0) {
|
|
|
+ default_error_ = CL_INVALID_PLATFORM;
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_platform_id> ids(n);
|
|
|
+ err = ::clGetPlatformIDs(n, ids.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ default_error_ = err;
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ default_ = Platform(ids[0]);
|
|
|
+ }
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ catch (cl::Error &e) {
|
|
|
+ default_error_ = e.err();
|
|
|
+ }
|
|
|
+#endif
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Create the default platform from a provided platform.
|
|
|
+ *
|
|
|
+ * This sets @c default_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefaultProvided(const Platform &p) {
|
|
|
+ default_ = p;
|
|
|
+ }
|
|
|
+
|
|
|
+public:
|
|
|
+#ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+ /*! \brief Reset the default.
|
|
|
+ *
|
|
|
+ * This sets @c default_ to an empty value to support cleanup in
|
|
|
+ * the unit test framework.
|
|
|
+ * This function is not thread safe.
|
|
|
+ */
|
|
|
+ static void unitTestClearDefault() {
|
|
|
+ default_ = Platform();
|
|
|
+ }
|
|
|
+#endif // #ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Platform() : detail::Wrapper<cl_type>() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_platform_id.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * This simply copies the platform ID value, which is an inexpensive operation.
|
|
|
+ */
|
|
|
+ explicit Platform(const cl_platform_id &platform, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(platform, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_platform_id.
|
|
|
+ *
|
|
|
+ * This simply copies the platform ID value, which is an inexpensive operation.
|
|
|
+ */
|
|
|
+ Platform& operator = (const cl_platform_id& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ static Platform getDefault(
|
|
|
+ cl_int *errResult = NULL)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefault);
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ if (errResult != NULL) {
|
|
|
+ *errResult = default_error_;
|
|
|
+ }
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Modify the default platform to be used by
|
|
|
+ * subsequent operations.
|
|
|
+ * Will only set the default if no default was previously created.
|
|
|
+ * @return updated default platform.
|
|
|
+ * Should be compared to the passed value to ensure that it was updated.
|
|
|
+ */
|
|
|
+ static Platform setDefault(const Platform &default_platform)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefaultProvided, std::cref(default_platform));
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetPlatformInfo().
|
|
|
+ cl_int getInfo(cl_platform_info name, string* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetPlatformInfo, object_, name, param),
|
|
|
+ __GET_PLATFORM_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetPlatformInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_platform_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_platform_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Gets a list of devices for this platform.
|
|
|
+ *
|
|
|
+ * Wraps clGetDeviceIDs().
|
|
|
+ */
|
|
|
+ cl_int getDevices(
|
|
|
+ cl_device_type type,
|
|
|
+ vector<Device>* devices) const
|
|
|
+ {
|
|
|
+ cl_uint n = 0;
|
|
|
+ if( devices == NULL ) {
|
|
|
+ return detail::errHandler(CL_INVALID_ARG_VALUE, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+ cl_int err = ::clGetDeviceIDs(object_, type, 0, NULL, &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_device_id> ids(n);
|
|
|
+ err = ::clGetDeviceIDs(object_, type, n, ids.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ // Cannot trivially assign because we need to capture intermediates
|
|
|
+ // with safe construction
|
|
|
+ // We must retain things we obtain from the API to avoid releasing
|
|
|
+ // API-owned objects.
|
|
|
+ if (devices) {
|
|
|
+ devices->resize(ids.size());
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < ids.size(); i++) {
|
|
|
+ (*devices)[i] = Device(ids[i], true);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_DX_INTEROP)
|
|
|
+ /*! \brief Get the list of available D3D10 devices.
|
|
|
+ *
|
|
|
+ * \param d3d_device_source.
|
|
|
+ *
|
|
|
+ * \param d3d_object.
|
|
|
+ *
|
|
|
+ * \param d3d_device_set.
|
|
|
+ *
|
|
|
+ * \param devices returns a vector of OpenCL D3D10 devices found. The cl::Device
|
|
|
+ * values returned in devices can be used to identify a specific OpenCL
|
|
|
+ * device. If \a devices argument is NULL, this argument is ignored.
|
|
|
+ *
|
|
|
+ * \return One of the following values:
|
|
|
+ * - CL_SUCCESS if the function is executed successfully.
|
|
|
+ *
|
|
|
+ * The application can query specific capabilities of the OpenCL device(s)
|
|
|
+ * returned by cl::getDevices. This can be used by the application to
|
|
|
+ * determine which device(s) to use.
|
|
|
+ *
|
|
|
+ * \note In the case that exceptions are enabled and a return value
|
|
|
+ * other than CL_SUCCESS is generated, then cl::Error exception is
|
|
|
+ * generated.
|
|
|
+ */
|
|
|
+ cl_int getDevices(
|
|
|
+ cl_d3d10_device_source_khr d3d_device_source,
|
|
|
+ void * d3d_object,
|
|
|
+ cl_d3d10_device_set_khr d3d_device_set,
|
|
|
+ vector<Device>* devices) const
|
|
|
+ {
|
|
|
+ typedef CL_API_ENTRY cl_int (CL_API_CALL *PFN_clGetDeviceIDsFromD3D10KHR)(
|
|
|
+ cl_platform_id platform,
|
|
|
+ cl_d3d10_device_source_khr d3d_device_source,
|
|
|
+ void * d3d_object,
|
|
|
+ cl_d3d10_device_set_khr d3d_device_set,
|
|
|
+ cl_uint num_entries,
|
|
|
+ cl_device_id * devices,
|
|
|
+ cl_uint* num_devices);
|
|
|
+
|
|
|
+ if( devices == NULL ) {
|
|
|
+ return detail::errHandler(CL_INVALID_ARG_VALUE, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ static PFN_clGetDeviceIDsFromD3D10KHR pfn_clGetDeviceIDsFromD3D10KHR = NULL;
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_(object_, clGetDeviceIDsFromD3D10KHR);
|
|
|
+
|
|
|
+ cl_uint n = 0;
|
|
|
+ cl_int err = pfn_clGetDeviceIDsFromD3D10KHR(
|
|
|
+ object_,
|
|
|
+ d3d_device_source,
|
|
|
+ d3d_object,
|
|
|
+ d3d_device_set,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_device_id> ids(n);
|
|
|
+ err = pfn_clGetDeviceIDsFromD3D10KHR(
|
|
|
+ object_,
|
|
|
+ d3d_device_source,
|
|
|
+ d3d_object,
|
|
|
+ d3d_device_set,
|
|
|
+ n,
|
|
|
+ ids.data(),
|
|
|
+ NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_DEVICE_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ // Cannot trivially assign because we need to capture intermediates
|
|
|
+ // with safe construction
|
|
|
+ // We must retain things we obtain from the API to avoid releasing
|
|
|
+ // API-owned objects.
|
|
|
+ if (devices) {
|
|
|
+ devices->resize(ids.size());
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < ids.size(); i++) {
|
|
|
+ (*devices)[i] = Device(ids[i], true);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+#endif
|
|
|
+
|
|
|
+ /*! \brief Gets a list of available platforms.
|
|
|
+ *
|
|
|
+ * Wraps clGetPlatformIDs().
|
|
|
+ */
|
|
|
+ static cl_int get(
|
|
|
+ vector<Platform>* platforms)
|
|
|
+ {
|
|
|
+ cl_uint n = 0;
|
|
|
+
|
|
|
+ if( platforms == NULL ) {
|
|
|
+ return detail::errHandler(CL_INVALID_ARG_VALUE, __GET_PLATFORM_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int err = ::clGetPlatformIDs(0, NULL, &n);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_PLATFORM_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_platform_id> ids(n);
|
|
|
+ err = ::clGetPlatformIDs(n, ids.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_PLATFORM_IDS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (platforms) {
|
|
|
+ platforms->resize(ids.size());
|
|
|
+
|
|
|
+ // Platforms don't reference count
|
|
|
+ for (size_type i = 0; i < ids.size(); i++) {
|
|
|
+ (*platforms)[i] = Platform(ids[i]);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Gets the first available platform.
|
|
|
+ *
|
|
|
+ * Wraps clGetPlatformIDs(), returning the first result.
|
|
|
+ */
|
|
|
+ static cl_int get(
|
|
|
+ Platform * platform)
|
|
|
+ {
|
|
|
+ cl_int err;
|
|
|
+ Platform default_platform = Platform::getDefault(&err);
|
|
|
+ if (platform) {
|
|
|
+ *platform = default_platform;
|
|
|
+ }
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Gets the first available platform, returning it by value.
|
|
|
+ *
|
|
|
+ * \return Returns a valid platform if one is available.
|
|
|
+ * If no platform is available will return a null platform.
|
|
|
+ * Throws an exception if no platforms are available
|
|
|
+ * or an error condition occurs.
|
|
|
+ * Wraps clGetPlatformIDs(), returning the first result.
|
|
|
+ */
|
|
|
+ static Platform get(
|
|
|
+ cl_int * errResult = NULL)
|
|
|
+ {
|
|
|
+ cl_int err;
|
|
|
+ Platform default_platform = Platform::getDefault(&err);
|
|
|
+ if (errResult) {
|
|
|
+ *errResult = err;
|
|
|
+ }
|
|
|
+ return default_platform;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ //! \brief Wrapper for clUnloadCompiler().
|
|
|
+ cl_int
|
|
|
+ unloadCompiler()
|
|
|
+ {
|
|
|
+ return ::clUnloadPlatformCompiler(object_);
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+}; // class Platform
|
|
|
+
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ std::once_flag Platform::default_initialized_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ Platform Platform::default_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ cl_int Platform::default_error_ = CL_SUCCESS;
|
|
|
+
|
|
|
+
|
|
|
+/**
|
|
|
+ * Deprecated APIs for 1.2
|
|
|
+ */
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+/**
|
|
|
+ * Unload the OpenCL compiler.
|
|
|
+ * \note Deprecated for OpenCL 1.2. Use Platform::unloadCompiler instead.
|
|
|
+ */
|
|
|
+inline CL_EXT_PREFIX__VERSION_1_1_DEPRECATED cl_int
|
|
|
+UnloadCompiler() CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED;
|
|
|
+inline cl_int
|
|
|
+UnloadCompiler()
|
|
|
+{
|
|
|
+ return ::clUnloadCompiler();
|
|
|
+}
|
|
|
+#endif // #if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_context.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are shallow, meaning that the copy will refer
|
|
|
+ * to the same underlying cl_context as the original. For details, see
|
|
|
+ * clRetainContext() and clReleaseContext().
|
|
|
+ *
|
|
|
+ * \see cl_context
|
|
|
+ */
|
|
|
+class Context
|
|
|
+ : public detail::Wrapper<cl_context>
|
|
|
+{
|
|
|
+private:
|
|
|
+ static std::once_flag default_initialized_;
|
|
|
+ static Context default_;
|
|
|
+ static cl_int default_error_;
|
|
|
+
|
|
|
+ /*! \brief Create the default context from the default device type in the default platform.
|
|
|
+ *
|
|
|
+ * This sets @c default_ and @c default_error_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefault() {
|
|
|
+ /* Throwing an exception from a call_once invocation does not do
|
|
|
+ * what we wish, so we catch it and save the error.
|
|
|
+ */
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ try
|
|
|
+#endif
|
|
|
+ {
|
|
|
+#if !defined(__APPLE__) && !defined(__MACOS)
|
|
|
+ const Platform &p = Platform::getDefault();
|
|
|
+ cl_platform_id defaultPlatform = p();
|
|
|
+ cl_context_properties properties[3] = {
|
|
|
+ CL_CONTEXT_PLATFORM, (cl_context_properties)defaultPlatform, 0
|
|
|
+ };
|
|
|
+#else // #if !defined(__APPLE__) && !defined(__MACOS)
|
|
|
+ cl_context_properties *properties = nullptr;
|
|
|
+#endif // #if !defined(__APPLE__) && !defined(__MACOS)
|
|
|
+
|
|
|
+ default_ = Context(
|
|
|
+ CL_DEVICE_TYPE_DEFAULT,
|
|
|
+ properties,
|
|
|
+ NULL,
|
|
|
+ NULL,
|
|
|
+ &default_error_);
|
|
|
+ }
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ catch (cl::Error &e) {
|
|
|
+ default_error_ = e.err();
|
|
|
+ }
|
|
|
+#endif
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ /*! \brief Create the default context from a provided Context.
|
|
|
+ *
|
|
|
+ * This sets @c default_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefaultProvided(const Context &c) {
|
|
|
+ default_ = c;
|
|
|
+ }
|
|
|
+
|
|
|
+public:
|
|
|
+#ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+ /*! \brief Reset the default.
|
|
|
+ *
|
|
|
+ * This sets @c default_ to an empty value to support cleanup in
|
|
|
+ * the unit test framework.
|
|
|
+ * This function is not thread safe.
|
|
|
+ */
|
|
|
+ static void unitTestClearDefault() {
|
|
|
+ default_ = Context();
|
|
|
+ }
|
|
|
+#endif // #ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+
|
|
|
+ /*! \brief Constructs a context including a list of specified devices.
|
|
|
+ *
|
|
|
+ * Wraps clCreateContext().
|
|
|
+ */
|
|
|
+ Context(
|
|
|
+ const vector<Device>& devices,
|
|
|
+ cl_context_properties* properties = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(
|
|
|
+ const char *,
|
|
|
+ const void *,
|
|
|
+ size_type,
|
|
|
+ void *) = NULL,
|
|
|
+ void* data = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ size_type numDevices = devices.size();
|
|
|
+ vector<cl_device_id> deviceIDs(numDevices);
|
|
|
+
|
|
|
+ for( size_type deviceIndex = 0; deviceIndex < numDevices; ++deviceIndex ) {
|
|
|
+ deviceIDs[deviceIndex] = (devices[deviceIndex])();
|
|
|
+ }
|
|
|
+
|
|
|
+ object_ = ::clCreateContext(
|
|
|
+ properties, (cl_uint) numDevices,
|
|
|
+ deviceIDs.data(),
|
|
|
+ notifyFptr, data, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ Context(
|
|
|
+ const Device& device,
|
|
|
+ cl_context_properties* properties = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(
|
|
|
+ const char *,
|
|
|
+ const void *,
|
|
|
+ size_type,
|
|
|
+ void *) = NULL,
|
|
|
+ void* data = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_device_id deviceID = device();
|
|
|
+
|
|
|
+ object_ = ::clCreateContext(
|
|
|
+ properties, 1,
|
|
|
+ &deviceID,
|
|
|
+ notifyFptr, data, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Constructs a context including all or a subset of devices of a specified type.
|
|
|
+ *
|
|
|
+ * Wraps clCreateContextFromType().
|
|
|
+ */
|
|
|
+ Context(
|
|
|
+ cl_device_type type,
|
|
|
+ cl_context_properties* properties = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(
|
|
|
+ const char *,
|
|
|
+ const void *,
|
|
|
+ size_type,
|
|
|
+ void *) = NULL,
|
|
|
+ void* data = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+#if !defined(__APPLE__) && !defined(__MACOS)
|
|
|
+ cl_context_properties prop[4] = {CL_CONTEXT_PLATFORM, 0, 0, 0 };
|
|
|
+
|
|
|
+ if (properties == NULL) {
|
|
|
+ // Get a valid platform ID as we cannot send in a blank one
|
|
|
+ vector<Platform> platforms;
|
|
|
+ error = Platform::get(&platforms);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_FROM_TYPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ // Check the platforms we found for a device of our specified type
|
|
|
+ cl_context_properties platform_id = 0;
|
|
|
+ for (unsigned int i = 0; i < platforms.size(); i++) {
|
|
|
+
|
|
|
+ vector<Device> devices;
|
|
|
+
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ try {
|
|
|
+#endif
|
|
|
+
|
|
|
+ error = platforms[i].getDevices(type, &devices);
|
|
|
+
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ } catch (cl::Error& e) {
|
|
|
+ error = e.err();
|
|
|
+ }
|
|
|
+ // Catch if exceptions are enabled as we don't want to exit if first platform has no devices of type
|
|
|
+ // We do error checking next anyway, and can throw there if needed
|
|
|
+#endif
|
|
|
+
|
|
|
+ // Only squash CL_SUCCESS and CL_DEVICE_NOT_FOUND
|
|
|
+ if (error != CL_SUCCESS && error != CL_DEVICE_NOT_FOUND) {
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_FROM_TYPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if (devices.size() > 0) {
|
|
|
+ platform_id = (cl_context_properties)platforms[i]();
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if (platform_id == 0) {
|
|
|
+ detail::errHandler(CL_DEVICE_NOT_FOUND, __CREATE_CONTEXT_FROM_TYPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = CL_DEVICE_NOT_FOUND;
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ prop[1] = platform_id;
|
|
|
+ properties = &prop[0];
|
|
|
+ }
|
|
|
+#endif
|
|
|
+ object_ = ::clCreateContextFromType(
|
|
|
+ properties, type, notifyFptr, data, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_FROM_TYPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Context(const Context& ctx) : detail::Wrapper<cl_type>(ctx) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Context& operator = (const Context &ctx)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(ctx);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Context(Context&& ctx) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(ctx)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Context& operator = (Context &&ctx)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(ctx));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ /*! \brief Returns a singleton context including all devices of CL_DEVICE_TYPE_DEFAULT.
|
|
|
+ *
|
|
|
+ * \note All calls to this function return the same cl_context as the first.
|
|
|
+ */
|
|
|
+ static Context getDefault(cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefault);
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = default_error_;
|
|
|
+ }
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Modify the default context to be used by
|
|
|
+ * subsequent operations.
|
|
|
+ * Will only set the default if no default was previously created.
|
|
|
+ * @return updated default context.
|
|
|
+ * Should be compared to the passed value to ensure that it was updated.
|
|
|
+ */
|
|
|
+ static Context setDefault(const Context &default_context)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefaultProvided, std::cref(default_context));
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Context() : detail::Wrapper<cl_type>() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_context - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the cl_context
|
|
|
+ * into the new Context object.
|
|
|
+ */
|
|
|
+ explicit Context(const cl_context& context, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(context, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_context - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the rhs and calls
|
|
|
+ * clReleaseContext() on the value previously held by this instance.
|
|
|
+ */
|
|
|
+ Context& operator = (const cl_context& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetContextInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_context_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetContextInfo, object_, name, param),
|
|
|
+ __GET_CONTEXT_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetContextInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_context_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_context_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Gets a list of supported image formats.
|
|
|
+ *
|
|
|
+ * Wraps clGetSupportedImageFormats().
|
|
|
+ */
|
|
|
+ cl_int getSupportedImageFormats(
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_mem_object_type type,
|
|
|
+ vector<ImageFormat>* formats) const
|
|
|
+ {
|
|
|
+ cl_uint numEntries;
|
|
|
+
|
|
|
+ if (!formats) {
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int err = ::clGetSupportedImageFormats(
|
|
|
+ object_,
|
|
|
+ flags,
|
|
|
+ type,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ &numEntries);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_SUPPORTED_IMAGE_FORMATS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (numEntries > 0) {
|
|
|
+ vector<ImageFormat> value(numEntries);
|
|
|
+ err = ::clGetSupportedImageFormats(
|
|
|
+ object_,
|
|
|
+ flags,
|
|
|
+ type,
|
|
|
+ numEntries,
|
|
|
+ (cl_image_format*)value.data(),
|
|
|
+ NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __GET_SUPPORTED_IMAGE_FORMATS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ formats->assign(begin(value), end(value));
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ // If no values are being returned, ensure an empty vector comes back
|
|
|
+ formats->clear();
|
|
|
+ }
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+inline void Device::makeDefault()
|
|
|
+{
|
|
|
+ /* Throwing an exception from a call_once invocation does not do
|
|
|
+ * what we wish, so we catch it and save the error.
|
|
|
+ */
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ try
|
|
|
+#endif
|
|
|
+ {
|
|
|
+ cl_int error = 0;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(&error);
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ default_error_ = error;
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ default_ = context.getInfo<CL_CONTEXT_DEVICES>()[0];
|
|
|
+ default_error_ = CL_SUCCESS;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ catch (cl::Error &e) {
|
|
|
+ default_error_ = e.err();
|
|
|
+ }
|
|
|
+#endif
|
|
|
+}
|
|
|
+
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ std::once_flag Context::default_initialized_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ Context Context::default_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ cl_int Context::default_error_ = CL_SUCCESS;
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_event.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are shallow, meaning that the copy will refer
|
|
|
+ * to the same underlying cl_event as the original. For details, see
|
|
|
+ * clRetainEvent() and clReleaseEvent().
|
|
|
+ *
|
|
|
+ * \see cl_event
|
|
|
+ */
|
|
|
+class Event : public detail::Wrapper<cl_event>
|
|
|
+{
|
|
|
+public:
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Event() : detail::Wrapper<cl_type>() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_event - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * This effectively transfers ownership of a refcount on the cl_event
|
|
|
+ * into the new Event object.
|
|
|
+ */
|
|
|
+ explicit Event(const cl_event& event, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(event, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_event - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the rhs and calls
|
|
|
+ * clReleaseEvent() on the value previously held by this instance.
|
|
|
+ */
|
|
|
+ Event& operator = (const cl_event& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetEventInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_event_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetEventInfo, object_, name, param),
|
|
|
+ __GET_EVENT_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetEventInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_event_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_event_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetEventProfilingInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getProfilingInfo(cl_profiling_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(detail::getInfo(
|
|
|
+ &::clGetEventProfilingInfo, object_, name, param),
|
|
|
+ __GET_EVENT_PROFILE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetEventProfilingInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_profiling_info, name>::param_type
|
|
|
+ getProfilingInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_profiling_info, name>::param_type param;
|
|
|
+ cl_int result = getProfilingInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Blocks the calling thread until this event completes.
|
|
|
+ *
|
|
|
+ * Wraps clWaitForEvents().
|
|
|
+ */
|
|
|
+ cl_int wait() const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clWaitForEvents(1, &object_),
|
|
|
+ __WAIT_FOR_EVENTS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ /*! \brief Registers a user callback function for a specific command execution status.
|
|
|
+ *
|
|
|
+ * Wraps clSetEventCallback().
|
|
|
+ */
|
|
|
+ cl_int setCallback(
|
|
|
+ cl_int type,
|
|
|
+ void (CL_CALLBACK * pfn_notify)(cl_event, cl_int, void *),
|
|
|
+ void * user_data = NULL)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetEventCallback(
|
|
|
+ object_,
|
|
|
+ type,
|
|
|
+ pfn_notify,
|
|
|
+ user_data),
|
|
|
+ __SET_EVENT_CALLBACK_ERR);
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+
|
|
|
+ /*! \brief Blocks the calling thread until every event specified is complete.
|
|
|
+ *
|
|
|
+ * Wraps clWaitForEvents().
|
|
|
+ */
|
|
|
+ static cl_int
|
|
|
+ waitForEvents(const vector<Event>& events)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clWaitForEvents(
|
|
|
+ (cl_uint) events.size(), (events.size() > 0) ? (cl_event*)&events.front() : NULL),
|
|
|
+ __WAIT_FOR_EVENTS_ERR);
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+/*! \brief Class interface for user events (a subset of cl_event's).
|
|
|
+ *
|
|
|
+ * See Event for details about copy semantics, etc.
|
|
|
+ */
|
|
|
+class UserEvent : public Event
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a user event on a given context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateUserEvent().
|
|
|
+ */
|
|
|
+ UserEvent(
|
|
|
+ const Context& context,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateUserEvent(
|
|
|
+ context(),
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_USER_EVENT_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ UserEvent() : Event() { }
|
|
|
+
|
|
|
+ /*! \brief Sets the execution status of a user event object.
|
|
|
+ *
|
|
|
+ * Wraps clSetUserEventStatus().
|
|
|
+ */
|
|
|
+ cl_int setStatus(cl_int status)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetUserEventStatus(object_,status),
|
|
|
+ __SET_USER_EVENT_STATUS_ERR);
|
|
|
+ }
|
|
|
+};
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+
|
|
|
+/*! \brief Blocks the calling thread until every event specified is complete.
|
|
|
+ *
|
|
|
+ * Wraps clWaitForEvents().
|
|
|
+ */
|
|
|
+inline static cl_int
|
|
|
+WaitForEvents(const vector<Event>& events)
|
|
|
+{
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clWaitForEvents(
|
|
|
+ (cl_uint) events.size(), (events.size() > 0) ? (cl_event*)&events.front() : NULL),
|
|
|
+ __WAIT_FOR_EVENTS_ERR);
|
|
|
+}
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_mem.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are shallow, meaning that the copy will refer
|
|
|
+ * to the same underlying cl_mem as the original. For details, see
|
|
|
+ * clRetainMemObject() and clReleaseMemObject().
|
|
|
+ *
|
|
|
+ * \see cl_mem
|
|
|
+ */
|
|
|
+class Memory : public detail::Wrapper<cl_mem>
|
|
|
+{
|
|
|
+public:
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Memory() : detail::Wrapper<cl_type>() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * Optionally transfer ownership of a refcount on the cl_mem
|
|
|
+ * into the new Memory object.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Memory(const cl_mem& memory, bool retainObject) :
|
|
|
+ detail::Wrapper<cl_type>(memory, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the rhs and calls
|
|
|
+ * clReleaseMemObject() on the value previously held by this instance.
|
|
|
+ */
|
|
|
+ Memory& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Memory(const Memory& mem) : detail::Wrapper<cl_type>(mem) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Memory& operator = (const Memory &mem)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(mem);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Memory(Memory&& mem) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(mem)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Memory& operator = (Memory &&mem)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(mem));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetMemObjectInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_mem_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetMemObjectInfo, object_, name, param),
|
|
|
+ __GET_MEM_OBJECT_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetMemObjectInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_mem_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_mem_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ /*! \brief Registers a callback function to be called when the memory object
|
|
|
+ * is no longer needed.
|
|
|
+ *
|
|
|
+ * Wraps clSetMemObjectDestructorCallback().
|
|
|
+ *
|
|
|
+ * Repeated calls to this function, for a given cl_mem value, will append
|
|
|
+ * to the list of functions called (in reverse order) when memory object's
|
|
|
+ * resources are freed and the memory object is deleted.
|
|
|
+ *
|
|
|
+ * \note
|
|
|
+ * The registered callbacks are associated with the underlying cl_mem
|
|
|
+ * value - not the Memory class instance.
|
|
|
+ */
|
|
|
+ cl_int setDestructorCallback(
|
|
|
+ void (CL_CALLBACK * pfn_notify)(cl_mem, void *),
|
|
|
+ void * user_data = NULL)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetMemObjectDestructorCallback(
|
|
|
+ object_,
|
|
|
+ pfn_notify,
|
|
|
+ user_data),
|
|
|
+ __SET_MEM_OBJECT_DESTRUCTOR_CALLBACK_ERR);
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+
|
|
|
+};
|
|
|
+
|
|
|
+// Pre-declare copy functions
|
|
|
+class Buffer;
|
|
|
+template< typename IteratorType >
|
|
|
+cl_int copy( IteratorType startIterator, IteratorType endIterator, cl::Buffer &buffer );
|
|
|
+template< typename IteratorType >
|
|
|
+cl_int copy( const cl::Buffer &buffer, IteratorType startIterator, IteratorType endIterator );
|
|
|
+template< typename IteratorType >
|
|
|
+cl_int copy( const CommandQueue &queue, IteratorType startIterator, IteratorType endIterator, cl::Buffer &buffer );
|
|
|
+template< typename IteratorType >
|
|
|
+cl_int copy( const CommandQueue &queue, const cl::Buffer &buffer, IteratorType startIterator, IteratorType endIterator );
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+namespace detail
|
|
|
+{
|
|
|
+ class SVMTraitNull
|
|
|
+ {
|
|
|
+ public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return 0;
|
|
|
+ }
|
|
|
+ };
|
|
|
+} // namespace detail
|
|
|
+
|
|
|
+template<class Trait = detail::SVMTraitNull>
|
|
|
+class SVMTraitReadWrite
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return CL_MEM_READ_WRITE |
|
|
|
+ Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template<class Trait = detail::SVMTraitNull>
|
|
|
+class SVMTraitReadOnly
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return CL_MEM_READ_ONLY |
|
|
|
+ Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template<class Trait = detail::SVMTraitNull>
|
|
|
+class SVMTraitWriteOnly
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return CL_MEM_WRITE_ONLY |
|
|
|
+ Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template<class Trait = SVMTraitReadWrite<>>
|
|
|
+class SVMTraitCoarse
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template<class Trait = SVMTraitReadWrite<>>
|
|
|
+class SVMTraitFine
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return CL_MEM_SVM_FINE_GRAIN_BUFFER |
|
|
|
+ Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+template<class Trait = SVMTraitReadWrite<>>
|
|
|
+class SVMTraitAtomic
|
|
|
+{
|
|
|
+public:
|
|
|
+ static cl_svm_mem_flags getSVMMemFlags()
|
|
|
+ {
|
|
|
+ return
|
|
|
+ CL_MEM_SVM_FINE_GRAIN_BUFFER |
|
|
|
+ CL_MEM_SVM_ATOMICS |
|
|
|
+ Trait::getSVMMemFlags();
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+// Pre-declare SVM map function
|
|
|
+template<typename T>
|
|
|
+inline cl_int enqueueMapSVM(
|
|
|
+ T* ptr,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL);
|
|
|
+
|
|
|
+/**
|
|
|
+ * STL-like allocator class for managing SVM objects provided for convenience.
|
|
|
+ *
|
|
|
+ * Note that while this behaves like an allocator for the purposes of constructing vectors and similar objects,
|
|
|
+ * care must be taken when using with smart pointers.
|
|
|
+ * The allocator should not be used to construct a unique_ptr if we are using coarse-grained SVM mode because
|
|
|
+ * the coarse-grained management behaviour would behave incorrectly with respect to reference counting.
|
|
|
+ *
|
|
|
+ * Instead the allocator embeds a Deleter which may be used with unique_ptr and is used
|
|
|
+ * with the allocate_shared and allocate_ptr supplied operations.
|
|
|
+ */
|
|
|
+template<typename T, class SVMTrait>
|
|
|
+class SVMAllocator {
|
|
|
+private:
|
|
|
+ Context context_;
|
|
|
+
|
|
|
+public:
|
|
|
+ typedef T value_type;
|
|
|
+ typedef value_type* pointer;
|
|
|
+ typedef const value_type* const_pointer;
|
|
|
+ typedef value_type& reference;
|
|
|
+ typedef const value_type& const_reference;
|
|
|
+ typedef std::size_t size_type;
|
|
|
+ typedef std::ptrdiff_t difference_type;
|
|
|
+
|
|
|
+ template<typename U>
|
|
|
+ struct rebind
|
|
|
+ {
|
|
|
+ typedef SVMAllocator<U, SVMTrait> other;
|
|
|
+ };
|
|
|
+
|
|
|
+ template<typename U, typename V>
|
|
|
+ friend class SVMAllocator;
|
|
|
+
|
|
|
+ SVMAllocator() :
|
|
|
+ context_(Context::getDefault())
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ explicit SVMAllocator(cl::Context context) :
|
|
|
+ context_(context)
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ SVMAllocator(const SVMAllocator &other) :
|
|
|
+ context_(other.context_)
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ template<typename U>
|
|
|
+ SVMAllocator(const SVMAllocator<U, SVMTrait> &other) :
|
|
|
+ context_(other.context_)
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ ~SVMAllocator()
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ pointer address(reference r) CL_HPP_NOEXCEPT_
|
|
|
+ {
|
|
|
+ return std::addressof(r);
|
|
|
+ }
|
|
|
+
|
|
|
+ const_pointer address(const_reference r) CL_HPP_NOEXCEPT_
|
|
|
+ {
|
|
|
+ return std::addressof(r);
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Allocate an SVM pointer.
|
|
|
+ *
|
|
|
+ * If the allocator is coarse-grained, this will take ownership to allow
|
|
|
+ * containers to correctly construct data in place.
|
|
|
+ */
|
|
|
+ pointer allocate(
|
|
|
+ size_type size,
|
|
|
+ typename cl::SVMAllocator<void, SVMTrait>::const_pointer = 0)
|
|
|
+ {
|
|
|
+ // Allocate memory with default alignment matching the size of the type
|
|
|
+ void* voidPointer =
|
|
|
+ clSVMAlloc(
|
|
|
+ context_(),
|
|
|
+ SVMTrait::getSVMMemFlags(),
|
|
|
+ size*sizeof(T),
|
|
|
+ 0);
|
|
|
+ pointer retValue = reinterpret_cast<pointer>(
|
|
|
+ voidPointer);
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ if (!retValue) {
|
|
|
+ std::bad_alloc excep;
|
|
|
+ throw excep;
|
|
|
+ }
|
|
|
+#endif // #if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+
|
|
|
+ // If allocation was coarse-grained then map it
|
|
|
+ if (!(SVMTrait::getSVMMemFlags() & CL_MEM_SVM_FINE_GRAIN_BUFFER)) {
|
|
|
+ cl_int err = enqueueMapSVM(retValue, CL_TRUE, CL_MAP_READ | CL_MAP_WRITE, size*sizeof(T));
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ std::bad_alloc excep;
|
|
|
+ throw excep;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ // If exceptions disabled, return null pointer from allocator
|
|
|
+ return retValue;
|
|
|
+ }
|
|
|
+
|
|
|
+ void deallocate(pointer p, size_type)
|
|
|
+ {
|
|
|
+ clSVMFree(context_(), p);
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Return the maximum possible allocation size.
|
|
|
+ * This is the minimum of the maximum sizes of all devices in the context.
|
|
|
+ */
|
|
|
+ size_type max_size() const CL_HPP_NOEXCEPT_
|
|
|
+ {
|
|
|
+ size_type maxSize = std::numeric_limits<size_type>::max() / sizeof(T);
|
|
|
+
|
|
|
+ for (const Device &d : context_.getInfo<CL_CONTEXT_DEVICES>()) {
|
|
|
+ maxSize = std::min(
|
|
|
+ maxSize,
|
|
|
+ static_cast<size_type>(d.getInfo<CL_DEVICE_MAX_MEM_ALLOC_SIZE>()));
|
|
|
+ }
|
|
|
+
|
|
|
+ return maxSize;
|
|
|
+ }
|
|
|
+
|
|
|
+ template< class U, class... Args >
|
|
|
+ void construct(U* p, Args&&... args)
|
|
|
+ {
|
|
|
+ new(p)T(args...);
|
|
|
+ }
|
|
|
+
|
|
|
+ template< class U >
|
|
|
+ void destroy(U* p)
|
|
|
+ {
|
|
|
+ p->~U();
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Returns true if the contexts match.
|
|
|
+ */
|
|
|
+ inline bool operator==(SVMAllocator const& rhs)
|
|
|
+ {
|
|
|
+ return (context_==rhs.context_);
|
|
|
+ }
|
|
|
+
|
|
|
+ inline bool operator!=(SVMAllocator const& a)
|
|
|
+ {
|
|
|
+ return !operator==(a);
|
|
|
+ }
|
|
|
+}; // class SVMAllocator return cl::pointer<T>(tmp, detail::Deleter<T, Alloc>{alloc, copies});
|
|
|
+
|
|
|
+
|
|
|
+template<class SVMTrait>
|
|
|
+class SVMAllocator<void, SVMTrait> {
|
|
|
+public:
|
|
|
+ typedef void value_type;
|
|
|
+ typedef value_type* pointer;
|
|
|
+ typedef const value_type* const_pointer;
|
|
|
+
|
|
|
+ template<typename U>
|
|
|
+ struct rebind
|
|
|
+ {
|
|
|
+ typedef SVMAllocator<U, SVMTrait> other;
|
|
|
+ };
|
|
|
+
|
|
|
+ template<typename U, typename V>
|
|
|
+ friend class SVMAllocator;
|
|
|
+};
|
|
|
+
|
|
|
+#if !defined(CL_HPP_NO_STD_UNIQUE_PTR)
|
|
|
+namespace detail
|
|
|
+{
|
|
|
+ template<class Alloc>
|
|
|
+ class Deleter {
|
|
|
+ private:
|
|
|
+ Alloc alloc_;
|
|
|
+ size_type copies_;
|
|
|
+
|
|
|
+ public:
|
|
|
+ typedef typename std::allocator_traits<Alloc>::pointer pointer;
|
|
|
+
|
|
|
+ Deleter(const Alloc &alloc, size_type copies) : alloc_{ alloc }, copies_{ copies }
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+ void operator()(pointer ptr) const {
|
|
|
+ Alloc tmpAlloc{ alloc_ };
|
|
|
+ std::allocator_traits<Alloc>::destroy(tmpAlloc, std::addressof(*ptr));
|
|
|
+ std::allocator_traits<Alloc>::deallocate(tmpAlloc, ptr, copies_);
|
|
|
+ }
|
|
|
+ };
|
|
|
+} // namespace detail
|
|
|
+
|
|
|
+/**
|
|
|
+ * Allocation operation compatible with std::allocate_ptr.
|
|
|
+ * Creates a unique_ptr<T> by default.
|
|
|
+ * This requirement is to ensure that the control block is not
|
|
|
+ * allocated in memory inaccessible to the host.
|
|
|
+ */
|
|
|
+template <class T, class Alloc, class... Args>
|
|
|
+cl::pointer<T, detail::Deleter<Alloc>> allocate_pointer(const Alloc &alloc_, Args&&... args)
|
|
|
+{
|
|
|
+ Alloc alloc(alloc_);
|
|
|
+ static const size_type copies = 1;
|
|
|
+
|
|
|
+ // Ensure that creation of the management block and the
|
|
|
+ // object are dealt with separately such that we only provide a deleter
|
|
|
+
|
|
|
+ T* tmp = std::allocator_traits<Alloc>::allocate(alloc, copies);
|
|
|
+ if (!tmp) {
|
|
|
+ std::bad_alloc excep;
|
|
|
+ throw excep;
|
|
|
+ }
|
|
|
+ try {
|
|
|
+ std::allocator_traits<Alloc>::construct(
|
|
|
+ alloc,
|
|
|
+ std::addressof(*tmp),
|
|
|
+ std::forward<Args>(args)...);
|
|
|
+
|
|
|
+ return cl::pointer<T, detail::Deleter<Alloc>>(tmp, detail::Deleter<Alloc>{alloc, copies});
|
|
|
+ }
|
|
|
+ catch (std::bad_alloc& b)
|
|
|
+ {
|
|
|
+ std::allocator_traits<Alloc>::deallocate(alloc, tmp, copies);
|
|
|
+ throw;
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+template< class T, class SVMTrait, class... Args >
|
|
|
+cl::pointer<T, detail::Deleter<SVMAllocator<T, SVMTrait>>> allocate_svm(Args... args)
|
|
|
+{
|
|
|
+ SVMAllocator<T, SVMTrait> alloc;
|
|
|
+ return cl::allocate_pointer<T>(alloc, args...);
|
|
|
+}
|
|
|
+
|
|
|
+template< class T, class SVMTrait, class... Args >
|
|
|
+cl::pointer<T, detail::Deleter<SVMAllocator<T, SVMTrait>>> allocate_svm(const cl::Context &c, Args... args)
|
|
|
+{
|
|
|
+ SVMAllocator<T, SVMTrait> alloc(c);
|
|
|
+ return cl::allocate_pointer<T>(alloc, args...);
|
|
|
+}
|
|
|
+#endif // #if !defined(CL_HPP_NO_STD_UNIQUE_PTR)
|
|
|
+
|
|
|
+/*! \brief Vector alias to simplify contruction of coarse-grained SVM containers.
|
|
|
+ *
|
|
|
+ */
|
|
|
+template < class T >
|
|
|
+using coarse_svm_vector = vector<T, cl::SVMAllocator<int, cl::SVMTraitCoarse<>>>;
|
|
|
+
|
|
|
+/*! \brief Vector alias to simplify contruction of fine-grained SVM containers.
|
|
|
+*
|
|
|
+*/
|
|
|
+template < class T >
|
|
|
+using fine_svm_vector = vector<T, cl::SVMAllocator<int, cl::SVMTraitFine<>>>;
|
|
|
+
|
|
|
+/*! \brief Vector alias to simplify contruction of fine-grained SVM containers that support platform atomics.
|
|
|
+*
|
|
|
+*/
|
|
|
+template < class T >
|
|
|
+using atomic_svm_vector = vector<T, cl::SVMAllocator<int, cl::SVMTraitAtomic<>>>;
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+
|
|
|
+/*! \brief Class interface for Buffer Memory Objects.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Buffer : public Memory
|
|
|
+{
|
|
|
+public:
|
|
|
+
|
|
|
+ /*! \brief Constructs a Buffer in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateBuffer().
|
|
|
+ *
|
|
|
+ * \param host_ptr Storage to be used if the CL_MEM_USE_HOST_PTR flag was
|
|
|
+ * specified. Note alignment & exclusivity requirements.
|
|
|
+ */
|
|
|
+ Buffer(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ size_type size,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, host_ptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Constructs a Buffer in the default context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateBuffer().
|
|
|
+ *
|
|
|
+ * \param host_ptr Storage to be used if the CL_MEM_USE_HOST_PTR flag was
|
|
|
+ * specified. Note alignment & exclusivity requirements.
|
|
|
+ *
|
|
|
+ * \see Context::getDefault()
|
|
|
+ */
|
|
|
+ Buffer(
|
|
|
+ cl_mem_flags flags,
|
|
|
+ size_type size,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, host_ptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Construct a Buffer from a host container via iterators.
|
|
|
+ * IteratorType must be random access.
|
|
|
+ * If useHostPtr is specified iterators must represent contiguous data.
|
|
|
+ */
|
|
|
+ template< typename IteratorType >
|
|
|
+ Buffer(
|
|
|
+ IteratorType startIterator,
|
|
|
+ IteratorType endIterator,
|
|
|
+ bool readOnly,
|
|
|
+ bool useHostPtr = false,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ typedef typename std::iterator_traits<IteratorType>::value_type DataType;
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_mem_flags flags = 0;
|
|
|
+ if( readOnly ) {
|
|
|
+ flags |= CL_MEM_READ_ONLY;
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ flags |= CL_MEM_READ_WRITE;
|
|
|
+ }
|
|
|
+ if( useHostPtr ) {
|
|
|
+ flags |= CL_MEM_USE_HOST_PTR;
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type size = sizeof(DataType)*(endIterator - startIterator);
|
|
|
+
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+ if( useHostPtr ) {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, static_cast<DataType*>(&*startIterator), &error);
|
|
|
+ } else {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, 0, &error);
|
|
|
+ }
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ if( !useHostPtr ) {
|
|
|
+ error = cl::copy(startIterator, endIterator, *this);
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Construct a Buffer from a host container via iterators using a specified context.
|
|
|
+ * IteratorType must be random access.
|
|
|
+ * If useHostPtr is specified iterators must represent contiguous data.
|
|
|
+ */
|
|
|
+ template< typename IteratorType >
|
|
|
+ Buffer(const Context &context, IteratorType startIterator, IteratorType endIterator,
|
|
|
+ bool readOnly, bool useHostPtr = false, cl_int* err = NULL);
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Construct a Buffer from a host container via iterators using a specified queue.
|
|
|
+ * If useHostPtr is specified iterators must be random access.
|
|
|
+ */
|
|
|
+ template< typename IteratorType >
|
|
|
+ Buffer(const CommandQueue &queue, IteratorType startIterator, IteratorType endIterator,
|
|
|
+ bool readOnly, bool useHostPtr = false, cl_int* err = NULL);
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Buffer() : Memory() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with earlier versions.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Buffer(const cl_mem& buffer, bool retainObject = false) :
|
|
|
+ Memory(buffer, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Buffer& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Memory::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Buffer(const Buffer& buf) : Memory(buf) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Buffer& operator = (const Buffer &buf)
|
|
|
+ {
|
|
|
+ Memory::operator=(buf);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Buffer(Buffer&& buf) CL_HPP_NOEXCEPT_ : Memory(std::move(buf)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Buffer& operator = (Buffer &&buf)
|
|
|
+ {
|
|
|
+ Memory::operator=(std::move(buf));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ /*! \brief Creates a new buffer object from this.
|
|
|
+ *
|
|
|
+ * Wraps clCreateSubBuffer().
|
|
|
+ */
|
|
|
+ Buffer createSubBuffer(
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_buffer_create_type buffer_create_type,
|
|
|
+ const void * buffer_create_info,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ Buffer result;
|
|
|
+ cl_int error;
|
|
|
+ result.object_ = ::clCreateSubBuffer(
|
|
|
+ object_,
|
|
|
+ flags,
|
|
|
+ buffer_create_type,
|
|
|
+ buffer_create_info,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_SUBBUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return result;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+};
|
|
|
+
|
|
|
+#if defined (CL_HPP_USE_DX_INTEROP)
|
|
|
+/*! \brief Class interface for creating OpenCL buffers from ID3D10Buffer's.
|
|
|
+ *
|
|
|
+ * This is provided to facilitate interoperability with Direct3D.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class BufferD3D10 : public Buffer
|
|
|
+{
|
|
|
+public:
|
|
|
+
|
|
|
+
|
|
|
+ /*! \brief Constructs a BufferD3D10, in a specified context, from a
|
|
|
+ * given ID3D10Buffer.
|
|
|
+ *
|
|
|
+ * Wraps clCreateFromD3D10BufferKHR().
|
|
|
+ */
|
|
|
+ BufferD3D10(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ID3D10Buffer* bufobj,
|
|
|
+ cl_int * err = NULL) : pfn_clCreateFromD3D10BufferKHR(nullptr)
|
|
|
+ {
|
|
|
+ typedef CL_API_ENTRY cl_mem (CL_API_CALL *PFN_clCreateFromD3D10BufferKHR)(
|
|
|
+ cl_context context, cl_mem_flags flags, ID3D10Buffer* buffer,
|
|
|
+ cl_int* errcode_ret);
|
|
|
+ PFN_clCreateFromD3D10BufferKHR pfn_clCreateFromD3D10BufferKHR;
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ vector<cl_context_properties> props = context.getInfo<CL_CONTEXT_PROPERTIES>();
|
|
|
+ cl_platform platform = -1;
|
|
|
+ for( int i = 0; i < props.size(); ++i ) {
|
|
|
+ if( props[i] == CL_CONTEXT_PLATFORM ) {
|
|
|
+ platform = props[i+1];
|
|
|
+ }
|
|
|
+ }
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_(platform, clCreateFromD3D10BufferKHR);
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clCreateFromD3D10BufferKHR);
|
|
|
+#endif
|
|
|
+
|
|
|
+ cl_int error;
|
|
|
+ object_ = pfn_clCreateFromD3D10BufferKHR(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ bufobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ BufferD3D10() : Buffer() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit BufferD3D10(const cl_mem& buffer, bool retainObject = false) :
|
|
|
+ Buffer(buffer, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ BufferD3D10& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Buffer::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferD3D10(const BufferD3D10& buf) :
|
|
|
+ Buffer(buf) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferD3D10& operator = (const BufferD3D10 &buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(buf);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferD3D10(BufferD3D10&& buf) CL_HPP_NOEXCEPT_ : Buffer(std::move(buf)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferD3D10& operator = (BufferD3D10 &&buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(std::move(buf));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+#endif
|
|
|
+
|
|
|
+/*! \brief Class interface for GL Buffer Memory Objects.
|
|
|
+ *
|
|
|
+ * This is provided to facilitate interoperability with OpenGL.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class BufferGL : public Buffer
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a BufferGL in a specified context, from a given
|
|
|
+ * GL buffer.
|
|
|
+ *
|
|
|
+ * Wraps clCreateFromGLBuffer().
|
|
|
+ */
|
|
|
+ BufferGL(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_GLuint bufobj,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateFromGLBuffer(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ bufobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ BufferGL() : Buffer() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit BufferGL(const cl_mem& buffer, bool retainObject = false) :
|
|
|
+ Buffer(buffer, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ BufferGL& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Buffer::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferGL(const BufferGL& buf) : Buffer(buf) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferGL& operator = (const BufferGL &buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(buf);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferGL(BufferGL&& buf) CL_HPP_NOEXCEPT_ : Buffer(std::move(buf)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferGL& operator = (BufferGL &&buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(std::move(buf));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetGLObjectInfo().
|
|
|
+ cl_int getObjectInfo(
|
|
|
+ cl_gl_object_type *type,
|
|
|
+ cl_GLuint * gl_object_name)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clGetGLObjectInfo(object_,type,gl_object_name),
|
|
|
+ __GET_GL_OBJECT_INFO_ERR);
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+/*! \brief Class interface for GL Render Buffer Memory Objects.
|
|
|
+ *
|
|
|
+ * This is provided to facilitate interoperability with OpenGL.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class BufferRenderGL : public Buffer
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a BufferRenderGL in a specified context, from a given
|
|
|
+ * GL Renderbuffer.
|
|
|
+ *
|
|
|
+ * Wraps clCreateFromGLRenderbuffer().
|
|
|
+ */
|
|
|
+ BufferRenderGL(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_GLuint bufobj,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateFromGLRenderbuffer(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ bufobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_RENDER_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ BufferRenderGL() : Buffer() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit BufferRenderGL(const cl_mem& buffer, bool retainObject = false) :
|
|
|
+ Buffer(buffer, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ BufferRenderGL& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Buffer::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferRenderGL(const BufferRenderGL& buf) : Buffer(buf) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferRenderGL& operator = (const BufferRenderGL &buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(buf);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferRenderGL(BufferRenderGL&& buf) CL_HPP_NOEXCEPT_ : Buffer(std::move(buf)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ BufferRenderGL& operator = (BufferRenderGL &&buf)
|
|
|
+ {
|
|
|
+ Buffer::operator=(std::move(buf));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetGLObjectInfo().
|
|
|
+ cl_int getObjectInfo(
|
|
|
+ cl_gl_object_type *type,
|
|
|
+ cl_GLuint * gl_object_name)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clGetGLObjectInfo(object_,type,gl_object_name),
|
|
|
+ __GET_GL_OBJECT_INFO_ERR);
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+/*! \brief C++ base class for Image Memory objects.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Image : public Memory
|
|
|
+{
|
|
|
+protected:
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image() : Memory() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image(const cl_mem& image, bool retainObject = false) :
|
|
|
+ Memory(image, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Memory::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image(const Image& img) : Memory(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image& operator = (const Image &img)
|
|
|
+ {
|
|
|
+ Memory::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image(Image&& img) CL_HPP_NOEXCEPT_ : Memory(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image& operator = (Image &&img)
|
|
|
+ {
|
|
|
+ Memory::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+public:
|
|
|
+ //! \brief Wrapper for clGetImageInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getImageInfo(cl_image_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetImageInfo, object_, name, param),
|
|
|
+ __GET_IMAGE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetImageInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_image_info, name>::param_type
|
|
|
+ getImageInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_image_info, name>::param_type param;
|
|
|
+ cl_int result = getImageInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+/*! \brief Class interface for 1D Image Memory objects.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Image1D : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a 1D Image in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateImage().
|
|
|
+ */
|
|
|
+ Image1D(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type width,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE1D,
|
|
|
+ width,
|
|
|
+ 0, 0, 0, 0, 0, 0, 0, 0
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ host_ptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image1D() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image1D(const cl_mem& image1D, bool retainObject = false) :
|
|
|
+ Image(image1D, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image1D& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1D(const Image1D& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1D& operator = (const Image1D &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1D(Image1D&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1D& operator = (Image1D &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+};
|
|
|
+
|
|
|
+/*! \class Image1DBuffer
|
|
|
+ * \brief Image interface for 1D buffer images.
|
|
|
+ */
|
|
|
+class Image1DBuffer : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ Image1DBuffer(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type width,
|
|
|
+ const Buffer &buffer,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE1D_BUFFER,
|
|
|
+ width,
|
|
|
+ 0, 0, 0, 0, 0, 0, 0,
|
|
|
+ buffer()
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ NULL,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ Image1DBuffer() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image1DBuffer(const cl_mem& image1D, bool retainObject = false) :
|
|
|
+ Image(image1D, retainObject) { }
|
|
|
+
|
|
|
+ Image1DBuffer& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DBuffer(const Image1DBuffer& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DBuffer& operator = (const Image1DBuffer &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DBuffer(Image1DBuffer&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DBuffer& operator = (Image1DBuffer &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+};
|
|
|
+
|
|
|
+/*! \class Image1DArray
|
|
|
+ * \brief Image interface for arrays of 1D images.
|
|
|
+ */
|
|
|
+class Image1DArray : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ Image1DArray(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type arraySize,
|
|
|
+ size_type width,
|
|
|
+ size_type rowPitch,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE1D_ARRAY,
|
|
|
+ width,
|
|
|
+ 0, 0, // height, depth (unused)
|
|
|
+ arraySize,
|
|
|
+ rowPitch,
|
|
|
+ 0, 0, 0, 0
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ host_ptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ Image1DArray() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image1DArray(const cl_mem& imageArray, bool retainObject = false) :
|
|
|
+ Image(imageArray, retainObject) { }
|
|
|
+
|
|
|
+
|
|
|
+ Image1DArray& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DArray(const Image1DArray& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DArray& operator = (const Image1DArray &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DArray(Image1DArray&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image1DArray& operator = (Image1DArray &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+};
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+
|
|
|
+/*! \brief Class interface for 2D Image Memory objects.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Image2D : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a 2D Image in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateImage().
|
|
|
+ */
|
|
|
+ Image2D(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type width,
|
|
|
+ size_type height,
|
|
|
+ size_type row_pitch = 0,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useCreateImage;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120 && CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useCreateImage = (version >= 0x10002); // OpenCL 1.2 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ useCreateImage = true;
|
|
|
+#else
|
|
|
+ useCreateImage = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ if (useCreateImage)
|
|
|
+ {
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE2D,
|
|
|
+ width,
|
|
|
+ height,
|
|
|
+ 0, 0, // depth, array size (unused)
|
|
|
+ row_pitch,
|
|
|
+ 0, 0, 0, 0
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ host_ptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ if (!useCreateImage)
|
|
|
+ {
|
|
|
+ object_ = ::clCreateImage2D(
|
|
|
+ context(), flags,&format, width, height, row_pitch, host_ptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE2D_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 || defined(CL_HPP_USE_CL_IMAGE2D_FROM_BUFFER_KHR)
|
|
|
+ /*! \brief Constructs a 2D Image from a buffer.
|
|
|
+ * \note This will share storage with the underlying buffer.
|
|
|
+ *
|
|
|
+ * Wraps clCreateImage().
|
|
|
+ */
|
|
|
+ Image2D(
|
|
|
+ const Context& context,
|
|
|
+ ImageFormat format,
|
|
|
+ const Buffer &sourceBuffer,
|
|
|
+ size_type width,
|
|
|
+ size_type height,
|
|
|
+ size_type row_pitch = 0,
|
|
|
+ cl_int* err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE2D,
|
|
|
+ width,
|
|
|
+ height,
|
|
|
+ 0, 0, // depth, array size (unused)
|
|
|
+ row_pitch,
|
|
|
+ 0, 0, 0,
|
|
|
+ // Use buffer as input to image
|
|
|
+ sourceBuffer()
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ 0, // flags inherited from buffer
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ nullptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != nullptr) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif //#if CL_HPP_TARGET_OPENCL_VERSION >= 200 || defined(CL_HPP_USE_CL_IMAGE2D_FROM_BUFFER_KHR)
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ /*! \brief Constructs a 2D Image from an image.
|
|
|
+ * \note This will share storage with the underlying image but may
|
|
|
+ * reinterpret the channel order and type.
|
|
|
+ *
|
|
|
+ * The image will be created matching with a descriptor matching the source.
|
|
|
+ *
|
|
|
+ * \param order is the channel order to reinterpret the image data as.
|
|
|
+ * The channel order may differ as described in the OpenCL
|
|
|
+ * 2.0 API specification.
|
|
|
+ *
|
|
|
+ * Wraps clCreateImage().
|
|
|
+ */
|
|
|
+ Image2D(
|
|
|
+ const Context& context,
|
|
|
+ cl_channel_order order,
|
|
|
+ const Image &sourceImage,
|
|
|
+ cl_int* err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ // Descriptor fields have to match source image
|
|
|
+ size_type sourceWidth =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_WIDTH>();
|
|
|
+ size_type sourceHeight =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_HEIGHT>();
|
|
|
+ size_type sourceRowPitch =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_ROW_PITCH>();
|
|
|
+ cl_uint sourceNumMIPLevels =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_NUM_MIP_LEVELS>();
|
|
|
+ cl_uint sourceNumSamples =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_NUM_SAMPLES>();
|
|
|
+ cl_image_format sourceFormat =
|
|
|
+ sourceImage.getImageInfo<CL_IMAGE_FORMAT>();
|
|
|
+
|
|
|
+ // Update only the channel order.
|
|
|
+ // Channel format inherited from source.
|
|
|
+ sourceFormat.image_channel_order = order;
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE2D,
|
|
|
+ sourceWidth,
|
|
|
+ sourceHeight,
|
|
|
+ 0, 0, // depth (unused), array size (unused)
|
|
|
+ sourceRowPitch,
|
|
|
+ 0, // slice pitch (unused)
|
|
|
+ sourceNumMIPLevels,
|
|
|
+ sourceNumSamples,
|
|
|
+ // Use buffer as input to image
|
|
|
+ sourceImage()
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ 0, // flags should be inherited from mem_object
|
|
|
+ &sourceFormat,
|
|
|
+ &desc,
|
|
|
+ nullptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != nullptr) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif //#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image2D() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image2D(const cl_mem& image2D, bool retainObject = false) :
|
|
|
+ Image(image2D, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image2D& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2D(const Image2D& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2D& operator = (const Image2D &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2D(Image2D&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2D& operator = (Image2D &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+};
|
|
|
+
|
|
|
+
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+/*! \brief Class interface for GL 2D Image Memory objects.
|
|
|
+ *
|
|
|
+ * This is provided to facilitate interoperability with OpenGL.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ * \note Deprecated for OpenCL 1.2. Please use ImageGL instead.
|
|
|
+ */
|
|
|
+class CL_EXT_PREFIX__VERSION_1_1_DEPRECATED Image2DGL : public Image2D
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs an Image2DGL in a specified context, from a given
|
|
|
+ * GL Texture.
|
|
|
+ *
|
|
|
+ * Wraps clCreateFromGLTexture2D().
|
|
|
+ */
|
|
|
+ Image2DGL(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_GLenum target,
|
|
|
+ cl_GLint miplevel,
|
|
|
+ cl_GLuint texobj,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateFromGLTexture2D(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ target,
|
|
|
+ miplevel,
|
|
|
+ texobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_TEXTURE_2D_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image2DGL() : Image2D() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image2DGL(const cl_mem& image, bool retainObject = false) :
|
|
|
+ Image2D(image, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *c
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image2DGL& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image2D::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DGL(const Image2DGL& img) : Image2D(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DGL& operator = (const Image2DGL &img)
|
|
|
+ {
|
|
|
+ Image2D::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DGL(Image2DGL&& img) CL_HPP_NOEXCEPT_ : Image2D(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DGL& operator = (Image2DGL &&img)
|
|
|
+ {
|
|
|
+ Image2D::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+} CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED;
|
|
|
+#endif // CL_USE_DEPRECATED_OPENCL_1_1_APIS
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+/*! \class Image2DArray
|
|
|
+ * \brief Image interface for arrays of 2D images.
|
|
|
+ */
|
|
|
+class Image2DArray : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ Image2DArray(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type arraySize,
|
|
|
+ size_type width,
|
|
|
+ size_type height,
|
|
|
+ size_type rowPitch,
|
|
|
+ size_type slicePitch,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE2D_ARRAY,
|
|
|
+ width,
|
|
|
+ height,
|
|
|
+ 0, // depth (unused)
|
|
|
+ arraySize,
|
|
|
+ rowPitch,
|
|
|
+ slicePitch,
|
|
|
+ 0, 0, 0
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ host_ptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ Image2DArray() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image2DArray(const cl_mem& imageArray, bool retainObject = false) : Image(imageArray, retainObject) { }
|
|
|
+
|
|
|
+ Image2DArray& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DArray(const Image2DArray& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DArray& operator = (const Image2DArray &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DArray(Image2DArray&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image2DArray& operator = (Image2DArray &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+/*! \brief Class interface for 3D Image Memory objects.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Image3D : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs a 3D Image in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateImage().
|
|
|
+ */
|
|
|
+ Image3D(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ ImageFormat format,
|
|
|
+ size_type width,
|
|
|
+ size_type height,
|
|
|
+ size_type depth,
|
|
|
+ size_type row_pitch = 0,
|
|
|
+ size_type slice_pitch = 0,
|
|
|
+ void* host_ptr = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useCreateImage;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120 && CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useCreateImage = (version >= 0x10002); // OpenCL 1.2 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ useCreateImage = true;
|
|
|
+#else
|
|
|
+ useCreateImage = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ if (useCreateImage)
|
|
|
+ {
|
|
|
+ cl_image_desc desc =
|
|
|
+ {
|
|
|
+ CL_MEM_OBJECT_IMAGE3D,
|
|
|
+ width,
|
|
|
+ height,
|
|
|
+ depth,
|
|
|
+ 0, // array size (unused)
|
|
|
+ row_pitch,
|
|
|
+ slice_pitch,
|
|
|
+ 0, 0, 0
|
|
|
+ };
|
|
|
+ object_ = ::clCreateImage(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ &format,
|
|
|
+ &desc,
|
|
|
+ host_ptr,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ if (!useCreateImage)
|
|
|
+ {
|
|
|
+ object_ = ::clCreateImage3D(
|
|
|
+ context(), flags, &format, width, height, depth, row_pitch,
|
|
|
+ slice_pitch, host_ptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_IMAGE3D_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 120
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image3D() : Image() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image3D(const cl_mem& image3D, bool retainObject = false) :
|
|
|
+ Image(image3D, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image3D& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3D(const Image3D& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3D& operator = (const Image3D &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3D(Image3D&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3D& operator = (Image3D &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+/*! \brief Class interface for GL 3D Image Memory objects.
|
|
|
+ *
|
|
|
+ * This is provided to facilitate interoperability with OpenGL.
|
|
|
+ *
|
|
|
+ * See Memory for details about copy semantics, etc.
|
|
|
+ *
|
|
|
+ * \see Memory
|
|
|
+ */
|
|
|
+class Image3DGL : public Image3D
|
|
|
+{
|
|
|
+public:
|
|
|
+ /*! \brief Constructs an Image3DGL in a specified context, from a given
|
|
|
+ * GL Texture.
|
|
|
+ *
|
|
|
+ * Wraps clCreateFromGLTexture3D().
|
|
|
+ */
|
|
|
+ Image3DGL(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_GLenum target,
|
|
|
+ cl_GLint miplevel,
|
|
|
+ cl_GLuint texobj,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateFromGLTexture3D(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ target,
|
|
|
+ miplevel,
|
|
|
+ texobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_TEXTURE_3D_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Image3DGL() : Image3D() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Image3DGL(const cl_mem& image, bool retainObject = false) :
|
|
|
+ Image3D(image, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Image3DGL& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image3D::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3DGL(const Image3DGL& img) : Image3D(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3DGL& operator = (const Image3DGL &img)
|
|
|
+ {
|
|
|
+ Image3D::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3DGL(Image3DGL&& img) CL_HPP_NOEXCEPT_ : Image3D(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Image3DGL& operator = (Image3DGL &&img)
|
|
|
+ {
|
|
|
+ Image3D::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+#endif // CL_USE_DEPRECATED_OPENCL_1_1_APIS
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+/*! \class ImageGL
|
|
|
+ * \brief general image interface for GL interop.
|
|
|
+ * We abstract the 2D and 3D GL images into a single instance here
|
|
|
+ * that wraps all GL sourced images on the grounds that setup information
|
|
|
+ * was performed by OpenCL anyway.
|
|
|
+ */
|
|
|
+class ImageGL : public Image
|
|
|
+{
|
|
|
+public:
|
|
|
+ ImageGL(
|
|
|
+ const Context& context,
|
|
|
+ cl_mem_flags flags,
|
|
|
+ cl_GLenum target,
|
|
|
+ cl_GLint miplevel,
|
|
|
+ cl_GLuint texobj,
|
|
|
+ cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ object_ = ::clCreateFromGLTexture(
|
|
|
+ context(),
|
|
|
+ flags,
|
|
|
+ target,
|
|
|
+ miplevel,
|
|
|
+ texobj,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_GL_TEXTURE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ ImageGL() : Image() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit ImageGL(const cl_mem& image, bool retainObject = false) :
|
|
|
+ Image(image, retainObject) { }
|
|
|
+
|
|
|
+ ImageGL& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Image::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ ImageGL(const ImageGL& img) : Image(img) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ ImageGL& operator = (const ImageGL &img)
|
|
|
+ {
|
|
|
+ Image::operator=(img);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ ImageGL(ImageGL&& img) CL_HPP_NOEXCEPT_ : Image(std::move(img)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ ImageGL& operator = (ImageGL &&img)
|
|
|
+ {
|
|
|
+ Image::operator=(std::move(img));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+};
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+/*! \brief Class interface for Pipe Memory Objects.
|
|
|
+*
|
|
|
+* See Memory for details about copy semantics, etc.
|
|
|
+*
|
|
|
+* \see Memory
|
|
|
+*/
|
|
|
+class Pipe : public Memory
|
|
|
+{
|
|
|
+public:
|
|
|
+
|
|
|
+ /*! \brief Constructs a Pipe in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreatePipe().
|
|
|
+ * @param context Context in which to create the pipe.
|
|
|
+ * @param flags Bitfield. Only CL_MEM_READ_WRITE and CL_MEM_HOST_NO_ACCESS are valid.
|
|
|
+ * @param packet_size Size in bytes of a single packet of the pipe.
|
|
|
+ * @param max_packets Number of packets that may be stored in the pipe.
|
|
|
+ *
|
|
|
+ */
|
|
|
+ Pipe(
|
|
|
+ const Context& context,
|
|
|
+ cl_uint packet_size,
|
|
|
+ cl_uint max_packets,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_mem_flags flags = CL_MEM_READ_WRITE | CL_MEM_HOST_NO_ACCESS;
|
|
|
+ object_ = ::clCreatePipe(context(), flags, packet_size, max_packets, nullptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PIPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Constructs a Pipe in a the default context.
|
|
|
+ *
|
|
|
+ * Wraps clCreatePipe().
|
|
|
+ * @param flags Bitfield. Only CL_MEM_READ_WRITE and CL_MEM_HOST_NO_ACCESS are valid.
|
|
|
+ * @param packet_size Size in bytes of a single packet of the pipe.
|
|
|
+ * @param max_packets Number of packets that may be stored in the pipe.
|
|
|
+ *
|
|
|
+ */
|
|
|
+ Pipe(
|
|
|
+ cl_uint packet_size,
|
|
|
+ cl_uint max_packets,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+ cl_mem_flags flags = CL_MEM_READ_WRITE | CL_MEM_HOST_NO_ACCESS;
|
|
|
+ object_ = ::clCreatePipe(context(), flags, packet_size, max_packets, nullptr, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PIPE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Pipe() : Memory() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with earlier versions.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ explicit Pipe(const cl_mem& pipe, bool retainObject = false) :
|
|
|
+ Memory(pipe, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment from cl_mem - performs shallow copy.
|
|
|
+ *
|
|
|
+ * See Memory for further details.
|
|
|
+ */
|
|
|
+ Pipe& operator = (const cl_mem& rhs)
|
|
|
+ {
|
|
|
+ Memory::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Pipe(const Pipe& pipe) : Memory(pipe) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Pipe& operator = (const Pipe &pipe)
|
|
|
+ {
|
|
|
+ Memory::operator=(pipe);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Pipe(Pipe&& pipe) CL_HPP_NOEXCEPT_ : Memory(std::move(pipe)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Pipe& operator = (Pipe &&pipe)
|
|
|
+ {
|
|
|
+ Memory::operator=(std::move(pipe));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetMemObjectInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_pipe_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetPipeInfo, object_, name, param),
|
|
|
+ __GET_PIPE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetMemObjectInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_pipe_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_pipe_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+}; // class Pipe
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_sampler.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are shallow, meaning that the copy will refer
|
|
|
+ * to the same underlying cl_sampler as the original. For details, see
|
|
|
+ * clRetainSampler() and clReleaseSampler().
|
|
|
+ *
|
|
|
+ * \see cl_sampler
|
|
|
+ */
|
|
|
+class Sampler : public detail::Wrapper<cl_sampler>
|
|
|
+{
|
|
|
+public:
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Sampler() { }
|
|
|
+
|
|
|
+ /*! \brief Constructs a Sampler in a specified context.
|
|
|
+ *
|
|
|
+ * Wraps clCreateSampler().
|
|
|
+ */
|
|
|
+ Sampler(
|
|
|
+ const Context& context,
|
|
|
+ cl_bool normalized_coords,
|
|
|
+ cl_addressing_mode addressing_mode,
|
|
|
+ cl_filter_mode filter_mode,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ cl_sampler_properties sampler_properties[] = {
|
|
|
+ CL_SAMPLER_NORMALIZED_COORDS, normalized_coords,
|
|
|
+ CL_SAMPLER_ADDRESSING_MODE, addressing_mode,
|
|
|
+ CL_SAMPLER_FILTER_MODE, filter_mode,
|
|
|
+ 0 };
|
|
|
+ object_ = ::clCreateSamplerWithProperties(
|
|
|
+ context(),
|
|
|
+ sampler_properties,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_SAMPLER_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+#else
|
|
|
+ object_ = ::clCreateSampler(
|
|
|
+ context(),
|
|
|
+ normalized_coords,
|
|
|
+ addressing_mode,
|
|
|
+ filter_mode,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_SAMPLER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+#endif
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_sampler - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * This effectively transfers ownership of a refcount on the cl_sampler
|
|
|
+ * into the new Sampler object.
|
|
|
+ */
|
|
|
+ explicit Sampler(const cl_sampler& sampler, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(sampler, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_sampler - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the rhs and calls
|
|
|
+ * clReleaseSampler() on the value previously held by this instance.
|
|
|
+ */
|
|
|
+ Sampler& operator = (const cl_sampler& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Sampler(const Sampler& sam) : detail::Wrapper<cl_type>(sam) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Sampler& operator = (const Sampler &sam)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(sam);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Sampler(Sampler&& sam) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(sam)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Sampler& operator = (Sampler &&sam)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(sam));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetSamplerInfo().
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_sampler_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetSamplerInfo, object_, name, param),
|
|
|
+ __GET_SAMPLER_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Wrapper for clGetSamplerInfo() that returns by value.
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_sampler_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_sampler_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+class Program;
|
|
|
+class CommandQueue;
|
|
|
+class DeviceCommandQueue;
|
|
|
+class Kernel;
|
|
|
+
|
|
|
+//! \brief Class interface for specifying NDRange values.
|
|
|
+class NDRange
|
|
|
+{
|
|
|
+private:
|
|
|
+ size_type sizes_[3];
|
|
|
+ cl_uint dimensions_;
|
|
|
+
|
|
|
+public:
|
|
|
+ //! \brief Default constructor - resulting range has zero dimensions.
|
|
|
+ NDRange()
|
|
|
+ : dimensions_(0)
|
|
|
+ {
|
|
|
+ sizes_[0] = 0;
|
|
|
+ sizes_[1] = 0;
|
|
|
+ sizes_[2] = 0;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Constructs one-dimensional range.
|
|
|
+ NDRange(size_type size0)
|
|
|
+ : dimensions_(1)
|
|
|
+ {
|
|
|
+ sizes_[0] = size0;
|
|
|
+ sizes_[1] = 1;
|
|
|
+ sizes_[2] = 1;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Constructs two-dimensional range.
|
|
|
+ NDRange(size_type size0, size_type size1)
|
|
|
+ : dimensions_(2)
|
|
|
+ {
|
|
|
+ sizes_[0] = size0;
|
|
|
+ sizes_[1] = size1;
|
|
|
+ sizes_[2] = 1;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Constructs three-dimensional range.
|
|
|
+ NDRange(size_type size0, size_type size1, size_type size2)
|
|
|
+ : dimensions_(3)
|
|
|
+ {
|
|
|
+ sizes_[0] = size0;
|
|
|
+ sizes_[1] = size1;
|
|
|
+ sizes_[2] = size2;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Conversion operator to const size_type *.
|
|
|
+ *
|
|
|
+ * \returns a pointer to the size of the first dimension.
|
|
|
+ */
|
|
|
+ operator const size_type*() const {
|
|
|
+ return sizes_;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Queries the number of dimensions in the range.
|
|
|
+ size_type dimensions() const
|
|
|
+ {
|
|
|
+ return dimensions_;
|
|
|
+ }
|
|
|
+
|
|
|
+ //! \brief Returns the size of the object in bytes based on the
|
|
|
+ // runtime number of dimensions
|
|
|
+ size_type size() const
|
|
|
+ {
|
|
|
+ return dimensions_*sizeof(size_type);
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type* get()
|
|
|
+ {
|
|
|
+ return sizes_;
|
|
|
+ }
|
|
|
+
|
|
|
+ const size_type* get() const
|
|
|
+ {
|
|
|
+ return sizes_;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+//! \brief A zero-dimensional range.
|
|
|
+static const NDRange NullRange;
|
|
|
+
|
|
|
+//! \brief Local address wrapper for use with Kernel::setArg
|
|
|
+struct LocalSpaceArg
|
|
|
+{
|
|
|
+ size_type size_;
|
|
|
+};
|
|
|
+
|
|
|
+namespace detail {
|
|
|
+
|
|
|
+template <typename T, class Enable = void>
|
|
|
+struct KernelArgumentHandler;
|
|
|
+
|
|
|
+// Enable for objects that are not subclasses of memory
|
|
|
+// Pointers, constants etc
|
|
|
+template <typename T>
|
|
|
+struct KernelArgumentHandler<T, typename std::enable_if<!std::is_base_of<cl::Memory, T>::value>::type>
|
|
|
+{
|
|
|
+ static size_type size(const T&) { return sizeof(T); }
|
|
|
+ static const T* ptr(const T& value) { return &value; }
|
|
|
+};
|
|
|
+
|
|
|
+// Enable for subclasses of memory where we want to get a reference to the cl_mem out
|
|
|
+// and pass that in for safety
|
|
|
+template <typename T>
|
|
|
+struct KernelArgumentHandler<T, typename std::enable_if<std::is_base_of<cl::Memory, T>::value>::type>
|
|
|
+{
|
|
|
+ static size_type size(const T&) { return sizeof(cl_mem); }
|
|
|
+ static const cl_mem* ptr(const T& value) { return &(value()); }
|
|
|
+};
|
|
|
+
|
|
|
+// Specialization for DeviceCommandQueue defined later
|
|
|
+
|
|
|
+template <>
|
|
|
+struct KernelArgumentHandler<LocalSpaceArg, void>
|
|
|
+{
|
|
|
+ static size_type size(const LocalSpaceArg& value) { return value.size_; }
|
|
|
+ static const void* ptr(const LocalSpaceArg&) { return NULL; }
|
|
|
+};
|
|
|
+
|
|
|
+}
|
|
|
+//! \endcond
|
|
|
+
|
|
|
+/*! Local
|
|
|
+ * \brief Helper function for generating LocalSpaceArg objects.
|
|
|
+ */
|
|
|
+inline LocalSpaceArg
|
|
|
+Local(size_type size)
|
|
|
+{
|
|
|
+ LocalSpaceArg ret = { size };
|
|
|
+ return ret;
|
|
|
+}
|
|
|
+
|
|
|
+/*! \brief Class interface for cl_kernel.
|
|
|
+ *
|
|
|
+ * \note Copies of these objects are shallow, meaning that the copy will refer
|
|
|
+ * to the same underlying cl_kernel as the original. For details, see
|
|
|
+ * clRetainKernel() and clReleaseKernel().
|
|
|
+ *
|
|
|
+ * \see cl_kernel
|
|
|
+ */
|
|
|
+class Kernel : public detail::Wrapper<cl_kernel>
|
|
|
+{
|
|
|
+public:
|
|
|
+ inline Kernel(const Program& program, const char* name, cl_int* err = NULL);
|
|
|
+
|
|
|
+ //! \brief Default constructor - initializes to NULL.
|
|
|
+ Kernel() { }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_kernel - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ * This effectively transfers ownership of a refcount on the cl_kernel
|
|
|
+ * into the new Kernel object.
|
|
|
+ */
|
|
|
+ explicit Kernel(const cl_kernel& kernel, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(kernel, retainObject) { }
|
|
|
+
|
|
|
+ /*! \brief Assignment operator from cl_kernel - takes ownership.
|
|
|
+ *
|
|
|
+ * This effectively transfers ownership of a refcount on the rhs and calls
|
|
|
+ * clReleaseKernel() on the value previously held by this instance.
|
|
|
+ */
|
|
|
+ Kernel& operator = (const cl_kernel& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Kernel(const Kernel& kernel) : detail::Wrapper<cl_type>(kernel) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Kernel& operator = (const Kernel &kernel)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(kernel);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Kernel(Kernel&& kernel) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(kernel)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Kernel& operator = (Kernel &&kernel)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(kernel));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_kernel_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetKernelInfo, object_, name, param),
|
|
|
+ __GET_KERNEL_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_kernel_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_kernel_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ template <typename T>
|
|
|
+ cl_int getArgInfo(cl_uint argIndex, cl_kernel_arg_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetKernelArgInfo, object_, argIndex, name, param),
|
|
|
+ __GET_KERNEL_ARG_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_kernel_arg_info, name>::param_type
|
|
|
+ getArgInfo(cl_uint argIndex, cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_kernel_arg_info, name>::param_type param;
|
|
|
+ cl_int result = getArgInfo(argIndex, name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getWorkGroupInfo(
|
|
|
+ const Device& device, cl_kernel_work_group_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(
|
|
|
+ &::clGetKernelWorkGroupInfo, object_, device(), name, param),
|
|
|
+ __GET_KERNEL_WORK_GROUP_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_kernel_work_group_info, name>::param_type
|
|
|
+ getWorkGroupInfo(const Device& device, cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_kernel_work_group_info, name>::param_type param;
|
|
|
+ cl_int result = getWorkGroupInfo(device, name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+#if (CL_HPP_TARGET_OPENCL_VERSION >= 200 && defined(CL_HPP_USE_CL_SUB_GROUPS_KHR)) || CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ cl_int getSubGroupInfo(const cl::Device &dev, cl_kernel_sub_group_info name, const cl::NDRange &range, size_type* param) const
|
|
|
+ {
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ return detail::errHandler(
|
|
|
+ clGetKernelSubGroupInfo(object_, dev(), name, range.size(), range.get(), sizeof(size_type), param, nullptr),
|
|
|
+ __GET_KERNEL_SUB_GROUP_INFO_ERR);
|
|
|
+
|
|
|
+#else // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ typedef clGetKernelSubGroupInfoKHR_fn PFN_clGetKernelSubGroupInfoKHR;
|
|
|
+ static PFN_clGetKernelSubGroupInfoKHR pfn_clGetKernelSubGroupInfoKHR = NULL;
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clGetKernelSubGroupInfoKHR);
|
|
|
+
|
|
|
+ return detail::errHandler(
|
|
|
+ pfn_clGetKernelSubGroupInfoKHR(object_, dev(), name, range.size(), range.get(), sizeof(size_type), param, nullptr),
|
|
|
+ __GET_KERNEL_SUB_GROUP_INFO_ERR);
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name>
|
|
|
+ size_type getSubGroupInfo(const cl::Device &dev, const cl::NDRange &range, cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ size_type param;
|
|
|
+ cl_int result = getSubGroupInfo(dev, name, range, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ /*! \brief setArg overload taking a shared_ptr type
|
|
|
+ */
|
|
|
+ template<typename T, class D>
|
|
|
+ cl_int setArg(cl_uint index, const cl::pointer<T, D> &argPtr)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelArgSVMPointer(object_, index, argPtr.get()),
|
|
|
+ __SET_KERNEL_ARGS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief setArg overload taking a vector type.
|
|
|
+ */
|
|
|
+ template<typename T, class Alloc>
|
|
|
+ cl_int setArg(cl_uint index, const cl::vector<T, Alloc> &argPtr)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelArgSVMPointer(object_, index, argPtr.data()),
|
|
|
+ __SET_KERNEL_ARGS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief setArg overload taking a pointer type
|
|
|
+ */
|
|
|
+ template<typename T>
|
|
|
+ typename std::enable_if<std::is_pointer<T>::value, cl_int>::type
|
|
|
+ setArg(cl_uint index, const T argPtr)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelArgSVMPointer(object_, index, argPtr),
|
|
|
+ __SET_KERNEL_ARGS_ERR);
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+ /*! \brief setArg overload taking a POD type
|
|
|
+ */
|
|
|
+ template <typename T>
|
|
|
+ typename std::enable_if<!std::is_pointer<T>::value, cl_int>::type
|
|
|
+ setArg(cl_uint index, const T &value)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelArg(
|
|
|
+ object_,
|
|
|
+ index,
|
|
|
+ detail::KernelArgumentHandler<T>::size(value),
|
|
|
+ detail::KernelArgumentHandler<T>::ptr(value)),
|
|
|
+ __SET_KERNEL_ARGS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int setArg(cl_uint index, size_type size, const void* argPtr)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelArg(object_, index, size, argPtr),
|
|
|
+ __SET_KERNEL_ARGS_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ /*!
|
|
|
+ * Specify a vector of SVM pointers that the kernel may access in
|
|
|
+ * addition to its arguments.
|
|
|
+ */
|
|
|
+ cl_int setSVMPointers(const vector<void*> &pointerList)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelExecInfo(
|
|
|
+ object_,
|
|
|
+ CL_KERNEL_EXEC_INFO_SVM_PTRS,
|
|
|
+ sizeof(void*)*pointerList.size(),
|
|
|
+ pointerList.data()));
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Specify a std::array of SVM pointers that the kernel may access in
|
|
|
+ * addition to its arguments.
|
|
|
+ */
|
|
|
+ template<int ArrayLength>
|
|
|
+ cl_int setSVMPointers(const std::array<void*, ArrayLength> &pointerList)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelExecInfo(
|
|
|
+ object_,
|
|
|
+ CL_KERNEL_EXEC_INFO_SVM_PTRS,
|
|
|
+ sizeof(void*)*pointerList.size(),
|
|
|
+ pointerList.data()));
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Enable fine-grained system SVM.
|
|
|
+ *
|
|
|
+ * \note It is only possible to enable fine-grained system SVM if all devices
|
|
|
+ * in the context associated with kernel support it.
|
|
|
+ *
|
|
|
+ * \param svmEnabled True if fine-grained system SVM is requested. False otherwise.
|
|
|
+ * \return CL_SUCCESS if the function was executed succesfully. CL_INVALID_OPERATION
|
|
|
+ * if no devices in the context support fine-grained system SVM.
|
|
|
+ *
|
|
|
+ * \see clSetKernelExecInfo
|
|
|
+ */
|
|
|
+ cl_int enableFineGrainedSystemSVM(bool svmEnabled)
|
|
|
+ {
|
|
|
+ cl_bool svmEnabled_ = svmEnabled ? CL_TRUE : CL_FALSE;
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelExecInfo(
|
|
|
+ object_,
|
|
|
+ CL_KERNEL_EXEC_INFO_SVM_FINE_GRAIN_SYSTEM,
|
|
|
+ sizeof(cl_bool),
|
|
|
+ &svmEnabled_
|
|
|
+ )
|
|
|
+ );
|
|
|
+ }
|
|
|
+
|
|
|
+ template<int index, int ArrayLength, class D, typename T0, typename T1, typename... Ts>
|
|
|
+ void setSVMPointersHelper(std::array<void*, ArrayLength> &pointerList, const pointer<T0, D> &t0, const pointer<T1, D> &t1, Ts & ... ts)
|
|
|
+ {
|
|
|
+ pointerList[index] = static_cast<void*>(t0.get());
|
|
|
+ setSVMPointersHelper<index + 1, ArrayLength>(pointerList, t1, ts...);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<int index, int ArrayLength, typename T0, typename T1, typename... Ts>
|
|
|
+ typename std::enable_if<std::is_pointer<T0>::value, void>::type
|
|
|
+ setSVMPointersHelper(std::array<void*, ArrayLength> &pointerList, T0 t0, T1 t1, Ts... ts)
|
|
|
+ {
|
|
|
+ pointerList[index] = static_cast<void*>(t0);
|
|
|
+ setSVMPointersHelper<index + 1, ArrayLength>(pointerList, t1, ts...);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<int index, int ArrayLength, typename T0, class D>
|
|
|
+ void setSVMPointersHelper(std::array<void*, ArrayLength> &pointerList, const pointer<T0, D> &t0)
|
|
|
+ {
|
|
|
+ pointerList[index] = static_cast<void*>(t0.get());
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ template<int index, int ArrayLength, typename T0>
|
|
|
+ typename std::enable_if<std::is_pointer<T0>::value, void>::type
|
|
|
+ setSVMPointersHelper(std::array<void*, ArrayLength> &pointerList, T0 t0)
|
|
|
+ {
|
|
|
+ pointerList[index] = static_cast<void*>(t0);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<typename T0, typename... Ts>
|
|
|
+ cl_int setSVMPointers(const T0 &t0, Ts & ... ts)
|
|
|
+ {
|
|
|
+ std::array<void*, 1 + sizeof...(Ts)> pointerList;
|
|
|
+
|
|
|
+ setSVMPointersHelper<0, 1 + sizeof...(Ts)>(pointerList, t0, ts...);
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetKernelExecInfo(
|
|
|
+ object_,
|
|
|
+ CL_KERNEL_EXEC_INFO_SVM_PTRS,
|
|
|
+ sizeof(void*)*(1 + sizeof...(Ts)),
|
|
|
+ pointerList.data()));
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ /**
|
|
|
+ * Make a deep copy of the kernel object including its arguments.
|
|
|
+ * @return A new kernel object with internal state entirely separate from that
|
|
|
+ * of the original but with any arguments set on the original intact.
|
|
|
+ */
|
|
|
+ Kernel clone()
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ Kernel retValue(clCloneKernel(this->get(), &error));
|
|
|
+
|
|
|
+ detail::errHandler(error, __CLONE_KERNEL_ERR);
|
|
|
+ return retValue;
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+};
|
|
|
+
|
|
|
+/*! \class Program
|
|
|
+ * \brief Program interface that implements cl_program.
|
|
|
+ */
|
|
|
+class Program : public detail::Wrapper<cl_program>
|
|
|
+{
|
|
|
+public:
|
|
|
+#if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ typedef vector<vector<unsigned char>> Binaries;
|
|
|
+ typedef vector<string> Sources;
|
|
|
+#else // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ typedef vector<std::pair<const void*, size_type> > Binaries;
|
|
|
+ typedef vector<std::pair<const char*, size_type> > Sources;
|
|
|
+#endif // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+
|
|
|
+ Program(
|
|
|
+ const string& source,
|
|
|
+ bool build = false,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ const char * strings = source.c_str();
|
|
|
+ const size_type length = source.size();
|
|
|
+
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithSource(
|
|
|
+ context(), (cl_uint)1, &strings, &length, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_SOURCE_ERR);
|
|
|
+
|
|
|
+ if (error == CL_SUCCESS && build) {
|
|
|
+
|
|
|
+ error = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+#if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ "-cl-std=CL2.0",
|
|
|
+#else
|
|
|
+ "",
|
|
|
+#endif // #if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ NULL,
|
|
|
+ NULL);
|
|
|
+
|
|
|
+ detail::buildErrHandler(error, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ Program(
|
|
|
+ const Context& context,
|
|
|
+ const string& source,
|
|
|
+ bool build = false,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ const char * strings = source.c_str();
|
|
|
+ const size_type length = source.size();
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithSource(
|
|
|
+ context(), (cl_uint)1, &strings, &length, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_SOURCE_ERR);
|
|
|
+
|
|
|
+ if (error == CL_SUCCESS && build) {
|
|
|
+ error = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+#if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ "-cl-std=CL2.0",
|
|
|
+#else
|
|
|
+ "",
|
|
|
+#endif // #if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ NULL,
|
|
|
+ NULL);
|
|
|
+
|
|
|
+ detail::buildErrHandler(error, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Create a program from a vector of source strings and the default context.
|
|
|
+ * Does not compile or link the program.
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const Sources& sources,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+ const size_type n = (size_type)sources.size();
|
|
|
+
|
|
|
+ vector<size_type> lengths(n);
|
|
|
+ vector<const char*> strings(n);
|
|
|
+
|
|
|
+ for (size_type i = 0; i < n; ++i) {
|
|
|
+#if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ strings[i] = sources[(int)i].data();
|
|
|
+ lengths[i] = sources[(int)i].length();
|
|
|
+#else // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ strings[i] = sources[(int)i].first;
|
|
|
+ lengths[i] = sources[(int)i].second;
|
|
|
+#endif // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ }
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithSource(
|
|
|
+ context(), (cl_uint)n, strings.data(), lengths.data(), &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_SOURCE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Create a program from a vector of source strings and a provided context.
|
|
|
+ * Does not compile or link the program.
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const Context& context,
|
|
|
+ const Sources& sources,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ const size_type n = (size_type)sources.size();
|
|
|
+
|
|
|
+ vector<size_type> lengths(n);
|
|
|
+ vector<const char*> strings(n);
|
|
|
+
|
|
|
+ for (size_type i = 0; i < n; ++i) {
|
|
|
+#if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ strings[i] = sources[(int)i].data();
|
|
|
+ lengths[i] = sources[(int)i].length();
|
|
|
+#else // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ strings[i] = sources[(int)i].first;
|
|
|
+ lengths[i] = sources[(int)i].second;
|
|
|
+#endif // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ }
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithSource(
|
|
|
+ context(), (cl_uint)n, strings.data(), lengths.data(), &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_SOURCE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210 || (CL_HPP_TARGET_OPENCL_VERSION==200 && defined(CL_HPP_USE_IL_KHR))
|
|
|
+ /**
|
|
|
+ * Program constructor to allow construction of program from SPIR-V or another IL.
|
|
|
+ * Valid for either OpenCL >= 2.1 or when CL_HPP_USE_IL_KHR is defined.
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const vector<char>& IL,
|
|
|
+ bool build = false,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(err);
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithIL(
|
|
|
+ context(), static_cast<const void*>(IL.data()), IL.size(), &error);
|
|
|
+
|
|
|
+#else // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ typedef clCreateProgramWithILKHR_fn PFN_clCreateProgramWithILKHR;
|
|
|
+ static PFN_clCreateProgramWithILKHR pfn_clCreateProgramWithILKHR = NULL;
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clCreateProgramWithILKHR);
|
|
|
+
|
|
|
+ return detail::errHandler(
|
|
|
+ pfn_clCreateProgramWithILKHR(
|
|
|
+ context(), static_cast<const void*>(IL.data()), IL.size(), &error);
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_IL_ERR);
|
|
|
+
|
|
|
+ if (error == CL_SUCCESS && build) {
|
|
|
+
|
|
|
+ error = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+#if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ "-cl-std=CL2.0",
|
|
|
+#else
|
|
|
+ "",
|
|
|
+#endif // #if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ NULL,
|
|
|
+ NULL);
|
|
|
+
|
|
|
+ detail::buildErrHandler(error, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Program constructor to allow construction of program from SPIR-V or another IL
|
|
|
+ * for a specific context.
|
|
|
+ * Valid for either OpenCL >= 2.1 or when CL_HPP_USE_IL_KHR is defined.
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const Context& context,
|
|
|
+ const vector<char>& IL,
|
|
|
+ bool build = false,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithIL(
|
|
|
+ context(), static_cast<const void*>(IL.data()), IL.size(), &error);
|
|
|
+
|
|
|
+#else // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ typedef clCreateProgramWithILKHR_fn PFN_clCreateProgramWithILKHR;
|
|
|
+ static PFN_clCreateProgramWithILKHR pfn_clCreateProgramWithILKHR = NULL;
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clCreateProgramWithILKHR);
|
|
|
+
|
|
|
+ return detail::errHandler(
|
|
|
+ pfn_clCreateProgramWithILKHR(
|
|
|
+ context(), static_cast<const void*>(IL.data()), IL.size(), &error);
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_IL_ERR);
|
|
|
+
|
|
|
+ if (error == CL_SUCCESS && build) {
|
|
|
+ error = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+#if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ "-cl-std=CL2.0",
|
|
|
+#else
|
|
|
+ "",
|
|
|
+#endif // #if !defined(CL_HPP_CL_1_2_DEFAULT_BUILD)
|
|
|
+ NULL,
|
|
|
+ NULL);
|
|
|
+
|
|
|
+ detail::buildErrHandler(error, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Construct a program object from a list of devices and a per-device list of binaries.
|
|
|
+ * \param context A valid OpenCL context in which to construct the program.
|
|
|
+ * \param devices A vector of OpenCL device objects for which the program will be created.
|
|
|
+ * \param binaries A vector of pairs of a pointer to a binary object and its length.
|
|
|
+ * \param binaryStatus An optional vector that on completion will be resized to
|
|
|
+ * match the size of binaries and filled with values to specify if each binary
|
|
|
+ * was successfully loaded.
|
|
|
+ * Set to CL_SUCCESS if the binary was successfully loaded.
|
|
|
+ * Set to CL_INVALID_VALUE if the length is 0 or the binary pointer is NULL.
|
|
|
+ * Set to CL_INVALID_BINARY if the binary provided is not valid for the matching device.
|
|
|
+ * \param err if non-NULL will be set to CL_SUCCESS on successful operation or one of the following errors:
|
|
|
+ * CL_INVALID_CONTEXT if context is not a valid context.
|
|
|
+ * CL_INVALID_VALUE if the length of devices is zero; or if the length of binaries does not match the length of devices;
|
|
|
+ * or if any entry in binaries is NULL or has length 0.
|
|
|
+ * CL_INVALID_DEVICE if OpenCL devices listed in devices are not in the list of devices associated with context.
|
|
|
+ * CL_INVALID_BINARY if an invalid program binary was encountered for any device. binaryStatus will return specific status for each device.
|
|
|
+ * CL_OUT_OF_HOST_MEMORY if there is a failure to allocate resources required by the OpenCL implementation on the host.
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const Context& context,
|
|
|
+ const vector<Device>& devices,
|
|
|
+ const Binaries& binaries,
|
|
|
+ vector<cl_int>* binaryStatus = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ const size_type numDevices = devices.size();
|
|
|
+
|
|
|
+ // Catch size mismatch early and return
|
|
|
+ if(binaries.size() != numDevices) {
|
|
|
+ error = CL_INVALID_VALUE;
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_BINARY_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ vector<size_type> lengths(numDevices);
|
|
|
+ vector<const unsigned char*> images(numDevices);
|
|
|
+#if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ for (size_type i = 0; i < numDevices; ++i) {
|
|
|
+ images[i] = binaries[i].data();
|
|
|
+ lengths[i] = binaries[(int)i].size();
|
|
|
+ }
|
|
|
+#else // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+ for (size_type i = 0; i < numDevices; ++i) {
|
|
|
+ images[i] = (const unsigned char*)binaries[i].first;
|
|
|
+ lengths[i] = binaries[(int)i].second;
|
|
|
+ }
|
|
|
+#endif // #if !defined(CL_HPP_ENABLE_PROGRAM_CONSTRUCTION_FROM_ARRAY_COMPATIBILITY)
|
|
|
+
|
|
|
+ vector<cl_device_id> deviceIDs(numDevices);
|
|
|
+ for( size_type deviceIndex = 0; deviceIndex < numDevices; ++deviceIndex ) {
|
|
|
+ deviceIDs[deviceIndex] = (devices[deviceIndex])();
|
|
|
+ }
|
|
|
+
|
|
|
+ if(binaryStatus) {
|
|
|
+ binaryStatus->resize(numDevices);
|
|
|
+ }
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithBinary(
|
|
|
+ context(), (cl_uint) devices.size(),
|
|
|
+ deviceIDs.data(),
|
|
|
+ lengths.data(), images.data(), (binaryStatus != NULL && numDevices > 0)
|
|
|
+ ? &binaryStatus->front()
|
|
|
+ : NULL, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_BINARY_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ /**
|
|
|
+ * Create program using builtin kernels.
|
|
|
+ * \param kernelNames Semi-colon separated list of builtin kernel names
|
|
|
+ */
|
|
|
+ Program(
|
|
|
+ const Context& context,
|
|
|
+ const vector<Device>& devices,
|
|
|
+ const string& kernelNames,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+
|
|
|
+ size_type numDevices = devices.size();
|
|
|
+ vector<cl_device_id> deviceIDs(numDevices);
|
|
|
+ for( size_type deviceIndex = 0; deviceIndex < numDevices; ++deviceIndex ) {
|
|
|
+ deviceIDs[deviceIndex] = (devices[deviceIndex])();
|
|
|
+ }
|
|
|
+
|
|
|
+ object_ = ::clCreateProgramWithBuiltInKernels(
|
|
|
+ context(),
|
|
|
+ (cl_uint) devices.size(),
|
|
|
+ deviceIDs.data(),
|
|
|
+ kernelNames.c_str(),
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_PROGRAM_WITH_BUILT_IN_KERNELS_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+ Program() { }
|
|
|
+
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ */
|
|
|
+ explicit Program(const cl_program& program, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(program, retainObject) { }
|
|
|
+
|
|
|
+ Program& operator = (const cl_program& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Program(const Program& program) : detail::Wrapper<cl_type>(program) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Program& operator = (const Program &program)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(program);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Program(Program&& program) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(program)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ Program& operator = (Program &&program)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(program));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int build(
|
|
|
+ const vector<Device>& devices,
|
|
|
+ const char* options = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(cl_program, void *) = NULL,
|
|
|
+ void* data = NULL) const
|
|
|
+ {
|
|
|
+ size_type numDevices = devices.size();
|
|
|
+ vector<cl_device_id> deviceIDs(numDevices);
|
|
|
+
|
|
|
+ for( size_type deviceIndex = 0; deviceIndex < numDevices; ++deviceIndex ) {
|
|
|
+ deviceIDs[deviceIndex] = (devices[deviceIndex])();
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int buildError = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ (cl_uint)
|
|
|
+ devices.size(),
|
|
|
+ deviceIDs.data(),
|
|
|
+ options,
|
|
|
+ notifyFptr,
|
|
|
+ data);
|
|
|
+
|
|
|
+ return detail::buildErrHandler(buildError, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int build(
|
|
|
+ const char* options = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(cl_program, void *) = NULL,
|
|
|
+ void* data = NULL) const
|
|
|
+ {
|
|
|
+ cl_int buildError = ::clBuildProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ options,
|
|
|
+ notifyFptr,
|
|
|
+ data);
|
|
|
+
|
|
|
+
|
|
|
+ return detail::buildErrHandler(buildError, __BUILD_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ cl_int compile(
|
|
|
+ const char* options = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(cl_program, void *) = NULL,
|
|
|
+ void* data = NULL) const
|
|
|
+ {
|
|
|
+ cl_int error = ::clCompileProgram(
|
|
|
+ object_,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ options,
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ NULL,
|
|
|
+ notifyFptr,
|
|
|
+ data);
|
|
|
+ return detail::buildErrHandler(error, __COMPILE_PROGRAM_ERR, getBuildInfo<CL_PROGRAM_BUILD_LOG>());
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_program_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetProgramInfo, object_, name, param),
|
|
|
+ __GET_PROGRAM_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_program_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_program_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getBuildInfo(
|
|
|
+ const Device& device, cl_program_build_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(
|
|
|
+ &::clGetProgramBuildInfo, object_, device(), name, param),
|
|
|
+ __GET_PROGRAM_BUILD_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_program_build_info, name>::param_type
|
|
|
+ getBuildInfo(const Device& device, cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_program_build_info, name>::param_type param;
|
|
|
+ cl_int result = getBuildInfo(device, name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Build info function that returns a vector of device/info pairs for the specified
|
|
|
+ * info type and for all devices in the program.
|
|
|
+ * On an error reading the info for any device, an empty vector of info will be returned.
|
|
|
+ */
|
|
|
+ template <cl_int name>
|
|
|
+ vector<std::pair<cl::Device, typename detail::param_traits<detail::cl_program_build_info, name>::param_type>>
|
|
|
+ getBuildInfo(cl_int *err = NULL) const
|
|
|
+ {
|
|
|
+ cl_int result = CL_SUCCESS;
|
|
|
+
|
|
|
+ auto devs = getInfo<CL_PROGRAM_DEVICES>(&result);
|
|
|
+ vector<std::pair<cl::Device, typename detail::param_traits<detail::cl_program_build_info, name>::param_type>>
|
|
|
+ devInfo;
|
|
|
+
|
|
|
+ // If there was an initial error from getInfo return the error
|
|
|
+ if (result != CL_SUCCESS) {
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return devInfo;
|
|
|
+ }
|
|
|
+
|
|
|
+ for (const cl::Device &d : devs) {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_program_build_info, name>::param_type param;
|
|
|
+ result = getBuildInfo(d, name, ¶m);
|
|
|
+ devInfo.push_back(
|
|
|
+ std::pair<cl::Device, typename detail::param_traits<detail::cl_program_build_info, name>::param_type>
|
|
|
+ (d, param));
|
|
|
+ if (result != CL_SUCCESS) {
|
|
|
+ // On error, leave the loop and return the error code
|
|
|
+ break;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ if (result != CL_SUCCESS) {
|
|
|
+ devInfo.clear();
|
|
|
+ }
|
|
|
+ return devInfo;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int createKernels(vector<Kernel>* kernels)
|
|
|
+ {
|
|
|
+ cl_uint numKernels;
|
|
|
+ cl_int err = ::clCreateKernelsInProgram(object_, 0, NULL, &numKernels);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_KERNELS_IN_PROGRAM_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ vector<cl_kernel> value(numKernels);
|
|
|
+
|
|
|
+ err = ::clCreateKernelsInProgram(
|
|
|
+ object_, numKernels, value.data(), NULL);
|
|
|
+ if (err != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(err, __CREATE_KERNELS_IN_PROGRAM_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ if (kernels) {
|
|
|
+ kernels->resize(value.size());
|
|
|
+
|
|
|
+ // Assign to param, constructing with retain behaviour
|
|
|
+ // to correctly capture each underlying CL object
|
|
|
+ for (size_type i = 0; i < value.size(); i++) {
|
|
|
+ // We do not need to retain because this kernel is being created
|
|
|
+ // by the runtime
|
|
|
+ (*kernels)[i] = Kernel(value[i], false);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return CL_SUCCESS;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+ /*! \brief Registers a callback function to be called when destructors for
|
|
|
+ * program scope global variables are complete and before the
|
|
|
+ * program is released.
|
|
|
+ *
|
|
|
+ * Wraps clSetProgramReleaseCallback().
|
|
|
+ *
|
|
|
+ * Each call to this function registers the specified user callback function
|
|
|
+ * on a callback stack associated with program. The registered user callback
|
|
|
+ * functions are called in the reverse order in which they were registered.
|
|
|
+ */
|
|
|
+ cl_int setReleaseCallback(
|
|
|
+ void (CL_CALLBACK * pfn_notify)(cl_program program, void * user_data),
|
|
|
+ void * user_data = NULL)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetProgramReleaseCallback(
|
|
|
+ object_,
|
|
|
+ pfn_notify,
|
|
|
+ user_data),
|
|
|
+ __SET_PROGRAM_RELEASE_CALLBACK_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Sets a SPIR-V specialization constant.
|
|
|
+ *
|
|
|
+ * Wraps clSetProgramSpecializationConstant().
|
|
|
+ */
|
|
|
+ template <typename T>
|
|
|
+ typename std::enable_if<!std::is_pointer<T>::value, cl_int>::type
|
|
|
+ setSpecializationConstant(cl_uint index, const T &value)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetProgramSpecializationConstant(
|
|
|
+ object_,
|
|
|
+ index,
|
|
|
+ sizeof(value),
|
|
|
+ &value),
|
|
|
+ __SET_PROGRAM_SPECIALIZATION_CONSTANT_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Sets a SPIR-V specialization constant.
|
|
|
+ *
|
|
|
+ * Wraps clSetProgramSpecializationConstant().
|
|
|
+ */
|
|
|
+ cl_int setSpecializationConstant(cl_uint index, size_type size, const void* value)
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetProgramSpecializationConstant(
|
|
|
+ object_,
|
|
|
+ index,
|
|
|
+ size,
|
|
|
+ value),
|
|
|
+ __SET_PROGRAM_SPECIALIZATION_CONSTANT_ERR);
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+};
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+inline Program linkProgram(
|
|
|
+ Program input1,
|
|
|
+ Program input2,
|
|
|
+ const char* options = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(cl_program, void *) = NULL,
|
|
|
+ void* data = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+{
|
|
|
+ cl_int error_local = CL_SUCCESS;
|
|
|
+
|
|
|
+ cl_program programs[2] = { input1(), input2() };
|
|
|
+
|
|
|
+ Context ctx = input1.getInfo<CL_PROGRAM_CONTEXT>(&error_local);
|
|
|
+ if(error_local!=CL_SUCCESS) {
|
|
|
+ detail::errHandler(error_local, __LINK_PROGRAM_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_program prog = ::clLinkProgram(
|
|
|
+ ctx(),
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ options,
|
|
|
+ 2,
|
|
|
+ programs,
|
|
|
+ notifyFptr,
|
|
|
+ data,
|
|
|
+ &error_local);
|
|
|
+
|
|
|
+ detail::errHandler(error_local,__COMPILE_PROGRAM_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error_local;
|
|
|
+ }
|
|
|
+
|
|
|
+ return Program(prog);
|
|
|
+}
|
|
|
+
|
|
|
+inline Program linkProgram(
|
|
|
+ vector<Program> inputPrograms,
|
|
|
+ const char* options = NULL,
|
|
|
+ void (CL_CALLBACK * notifyFptr)(cl_program, void *) = NULL,
|
|
|
+ void* data = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+{
|
|
|
+ cl_int error_local = CL_SUCCESS;
|
|
|
+
|
|
|
+ vector<cl_program> programs(inputPrograms.size());
|
|
|
+
|
|
|
+ for (unsigned int i = 0; i < inputPrograms.size(); i++) {
|
|
|
+ programs[i] = inputPrograms[i]();
|
|
|
+ }
|
|
|
+
|
|
|
+ Context ctx;
|
|
|
+ if(inputPrograms.size() > 0) {
|
|
|
+ ctx = inputPrograms[0].getInfo<CL_PROGRAM_CONTEXT>(&error_local);
|
|
|
+ if(error_local!=CL_SUCCESS) {
|
|
|
+ detail::errHandler(error_local, __LINK_PROGRAM_ERR);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ cl_program prog = ::clLinkProgram(
|
|
|
+ ctx(),
|
|
|
+ 0,
|
|
|
+ NULL,
|
|
|
+ options,
|
|
|
+ (cl_uint)inputPrograms.size(),
|
|
|
+ programs.data(),
|
|
|
+ notifyFptr,
|
|
|
+ data,
|
|
|
+ &error_local);
|
|
|
+
|
|
|
+ detail::errHandler(error_local,__COMPILE_PROGRAM_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error_local;
|
|
|
+ }
|
|
|
+
|
|
|
+ return Program(prog, false);
|
|
|
+}
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+// Template specialization for CL_PROGRAM_BINARIES
|
|
|
+template <>
|
|
|
+inline cl_int cl::Program::getInfo(cl_program_info name, vector<vector<unsigned char>>* param) const
|
|
|
+{
|
|
|
+ if (name != CL_PROGRAM_BINARIES) {
|
|
|
+ return CL_INVALID_VALUE;
|
|
|
+ }
|
|
|
+ if (param) {
|
|
|
+ // Resize the parameter array appropriately for each allocation
|
|
|
+ // and pass down to the helper
|
|
|
+
|
|
|
+ vector<size_type> sizes = getInfo<CL_PROGRAM_BINARY_SIZES>();
|
|
|
+ size_type numBinaries = sizes.size();
|
|
|
+
|
|
|
+ // Resize the parameter array and constituent arrays
|
|
|
+ param->resize(numBinaries);
|
|
|
+ for (size_type i = 0; i < numBinaries; ++i) {
|
|
|
+ (*param)[i].resize(sizes[i]);
|
|
|
+ }
|
|
|
+
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(&::clGetProgramInfo, object_, name, param),
|
|
|
+ __GET_PROGRAM_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+template<>
|
|
|
+inline vector<vector<unsigned char>> cl::Program::getInfo<CL_PROGRAM_BINARIES>(cl_int* err) const
|
|
|
+{
|
|
|
+ vector<vector<unsigned char>> binariesVectors;
|
|
|
+
|
|
|
+ cl_int result = getInfo(CL_PROGRAM_BINARIES, &binariesVectors);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return binariesVectors;
|
|
|
+}
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+// Template specialization for clSetProgramSpecializationConstant
|
|
|
+template <>
|
|
|
+inline cl_int cl::Program::setSpecializationConstant(cl_uint index, const bool &value)
|
|
|
+{
|
|
|
+ cl_uchar ucValue = value ? CL_UCHAR_MAX : 0;
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clSetProgramSpecializationConstant(
|
|
|
+ object_,
|
|
|
+ index,
|
|
|
+ sizeof(ucValue),
|
|
|
+ &ucValue),
|
|
|
+ __SET_PROGRAM_SPECIALIZATION_CONSTANT_ERR);
|
|
|
+}
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 220
|
|
|
+
|
|
|
+inline Kernel::Kernel(const Program& program, const char* name, cl_int* err)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ object_ = ::clCreateKernel(program(), name, &error);
|
|
|
+ detail::errHandler(error, __CREATE_KERNEL_ERR);
|
|
|
+
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+}
|
|
|
+
|
|
|
+enum class QueueProperties : cl_command_queue_properties
|
|
|
+{
|
|
|
+ None = 0,
|
|
|
+ Profiling = CL_QUEUE_PROFILING_ENABLE,
|
|
|
+ OutOfOrder = CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE,
|
|
|
+};
|
|
|
+
|
|
|
+inline QueueProperties operator|(QueueProperties lhs, QueueProperties rhs)
|
|
|
+{
|
|
|
+ return static_cast<QueueProperties>(static_cast<cl_command_queue_properties>(lhs) | static_cast<cl_command_queue_properties>(rhs));
|
|
|
+}
|
|
|
+
|
|
|
+/*! \class CommandQueue
|
|
|
+ * \brief CommandQueue interface for cl_command_queue.
|
|
|
+ */
|
|
|
+class CommandQueue : public detail::Wrapper<cl_command_queue>
|
|
|
+{
|
|
|
+private:
|
|
|
+ static std::once_flag default_initialized_;
|
|
|
+ static CommandQueue default_;
|
|
|
+ static cl_int default_error_;
|
|
|
+
|
|
|
+ /*! \brief Create the default command queue returned by @ref getDefault.
|
|
|
+ *
|
|
|
+ * It sets default_error_ to indicate success or failure. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefault()
|
|
|
+ {
|
|
|
+ /* We don't want to throw an error from this function, so we have to
|
|
|
+ * catch and set the error flag.
|
|
|
+ */
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ try
|
|
|
+#endif
|
|
|
+ {
|
|
|
+ int error;
|
|
|
+ Context context = Context::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ default_error_ = error;
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ Device device = Device::getDefault();
|
|
|
+ default_ = CommandQueue(context, device, 0, &default_error_);
|
|
|
+ }
|
|
|
+ }
|
|
|
+#if defined(CL_HPP_ENABLE_EXCEPTIONS)
|
|
|
+ catch (cl::Error &e) {
|
|
|
+ default_error_ = e.err();
|
|
|
+ }
|
|
|
+#endif
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Create the default command queue.
|
|
|
+ *
|
|
|
+ * This sets @c default_. It does not throw
|
|
|
+ * @c cl::Error.
|
|
|
+ */
|
|
|
+ static void makeDefaultProvided(const CommandQueue &c) {
|
|
|
+ default_ = c;
|
|
|
+ }
|
|
|
+
|
|
|
+public:
|
|
|
+#ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+ /*! \brief Reset the default.
|
|
|
+ *
|
|
|
+ * This sets @c default_ to an empty value to support cleanup in
|
|
|
+ * the unit test framework.
|
|
|
+ * This function is not thread safe.
|
|
|
+ */
|
|
|
+ static void unitTestClearDefault() {
|
|
|
+ default_ = CommandQueue();
|
|
|
+ }
|
|
|
+#endif // #ifdef CL_HPP_UNIT_TEST_ENABLE
|
|
|
+
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue based on passed properties.
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ CommandQueue(
|
|
|
+ cl_command_queue_properties properties,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(&error);
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ Device device = context.getInfo<CL_CONTEXT_DEVICES>()[0];
|
|
|
+ bool useWithProperties;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties, 0 };
|
|
|
+ if ((properties & CL_QUEUE_ON_DEVICE) == 0) {
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ error = CL_INVALID_QUEUE_PROPERTIES;
|
|
|
+ }
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), device(), properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue based on passed properties.
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ CommandQueue(
|
|
|
+ QueueProperties properties,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ Context context = Context::getDefault(&error);
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ Device device = context.getInfo<CL_CONTEXT_DEVICES>()[0];
|
|
|
+ bool useWithProperties;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, static_cast<cl_queue_properties>(properties), 0 };
|
|
|
+
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), device(), static_cast<cl_command_queue_properties>(properties), &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue for an implementation defined device in the given context
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ explicit CommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ cl_command_queue_properties properties = 0,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useWithProperties;
|
|
|
+ vector<cl::Device> devices;
|
|
|
+ error = context.getInfo(CL_CONTEXT_DEVICES, &devices);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS)
|
|
|
+ {
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties, 0 };
|
|
|
+ if ((properties & CL_QUEUE_ON_DEVICE) == 0) {
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), devices[0](), queue_properties, &error);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ error = CL_INVALID_QUEUE_PROPERTIES;
|
|
|
+ }
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), devices[0](), properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue for an implementation defined device in the given context
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ explicit CommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ QueueProperties properties,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useWithProperties;
|
|
|
+ vector<cl::Device> devices;
|
|
|
+ error = context.getInfo(CL_CONTEXT_DEVICES, &devices);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_CONTEXT_ERR);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS)
|
|
|
+ {
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, static_cast<cl_queue_properties>(properties), 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), devices[0](), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), devices[0](), static_cast<cl_command_queue_properties>(properties), &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue for a passed device and context
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ CommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ const Device& device,
|
|
|
+ cl_command_queue_properties properties = 0,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useWithProperties;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties, 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), device(), properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * \brief Constructs a CommandQueue for a passed device and context
|
|
|
+ * Will return an CL_INVALID_QUEUE_PROPERTIES error if CL_QUEUE_ON_DEVICE is specified.
|
|
|
+ */
|
|
|
+ CommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ const Device& device,
|
|
|
+ QueueProperties properties,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ bool useWithProperties;
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200 && CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ // Run-time decision based on the actual platform
|
|
|
+ {
|
|
|
+ cl_uint version = detail::getContextPlatformVersion(context());
|
|
|
+ useWithProperties = (version >= 0x20000); // OpenCL 2.0 or above
|
|
|
+ }
|
|
|
+#elif CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ useWithProperties = true;
|
|
|
+#else
|
|
|
+ useWithProperties = false;
|
|
|
+#endif
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (useWithProperties) {
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, static_cast<cl_queue_properties>(properties), 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+#if CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ if (!useWithProperties) {
|
|
|
+ object_ = ::clCreateCommandQueue(
|
|
|
+ context(), device(), static_cast<cl_command_queue_properties>(properties), &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+#endif // CL_HPP_MINIMUM_OPENCL_VERSION < 200
|
|
|
+ }
|
|
|
+
|
|
|
+ static CommandQueue getDefault(cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefault);
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ detail::errHandler(default_error_, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+#else // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ detail::errHandler(default_error_, __CREATE_COMMAND_QUEUE_ERR);
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = default_error_;
|
|
|
+ }
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Modify the default command queue to be used by
|
|
|
+ * subsequent operations.
|
|
|
+ * Will only set the default if no default was previously created.
|
|
|
+ * @return updated default command queue.
|
|
|
+ * Should be compared to the passed value to ensure that it was updated.
|
|
|
+ */
|
|
|
+ static CommandQueue setDefault(const CommandQueue &default_queue)
|
|
|
+ {
|
|
|
+ std::call_once(default_initialized_, makeDefaultProvided, std::cref(default_queue));
|
|
|
+ detail::errHandler(default_error_);
|
|
|
+ return default_;
|
|
|
+ }
|
|
|
+
|
|
|
+ CommandQueue() { }
|
|
|
+
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_mem - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ */
|
|
|
+ explicit CommandQueue(const cl_command_queue& commandQueue, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(commandQueue, retainObject) { }
|
|
|
+
|
|
|
+ CommandQueue& operator = (const cl_command_queue& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ CommandQueue(const CommandQueue& queue) : detail::Wrapper<cl_type>(queue) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ CommandQueue& operator = (const CommandQueue &queue)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(queue);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ CommandQueue(CommandQueue&& queue) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(queue)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ CommandQueue& operator = (CommandQueue &&queue)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(queue));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_command_queue_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(
|
|
|
+ &::clGetCommandQueueInfo, object_, name, param),
|
|
|
+ __GET_COMMAND_QUEUE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_command_queue_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_command_queue_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueReadBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueReadBuffer(
|
|
|
+ object_, buffer(), blocking, offset, size,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_READ_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueWriteBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ const void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueWriteBuffer(
|
|
|
+ object_, buffer(), blocking, offset, size,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_WRITE_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueCopyBuffer(
|
|
|
+ const Buffer& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ size_type src_offset,
|
|
|
+ size_type dst_offset,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueCopyBuffer(
|
|
|
+ object_, src(), dst(), src_offset, dst_offset, size,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQEUE_COPY_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ cl_int enqueueReadBufferRect(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& buffer_offset,
|
|
|
+ const array<size_type, 3>& host_offset,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type buffer_row_pitch,
|
|
|
+ size_type buffer_slice_pitch,
|
|
|
+ size_type host_row_pitch,
|
|
|
+ size_type host_slice_pitch,
|
|
|
+ void *ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueReadBufferRect(
|
|
|
+ object_,
|
|
|
+ buffer(),
|
|
|
+ blocking,
|
|
|
+ buffer_offset.data(),
|
|
|
+ host_offset.data(),
|
|
|
+ region.data(),
|
|
|
+ buffer_row_pitch,
|
|
|
+ buffer_slice_pitch,
|
|
|
+ host_row_pitch,
|
|
|
+ host_slice_pitch,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_READ_BUFFER_RECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueWriteBufferRect(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& buffer_offset,
|
|
|
+ const array<size_type, 3>& host_offset,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type buffer_row_pitch,
|
|
|
+ size_type buffer_slice_pitch,
|
|
|
+ size_type host_row_pitch,
|
|
|
+ size_type host_slice_pitch,
|
|
|
+ const void *ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueWriteBufferRect(
|
|
|
+ object_,
|
|
|
+ buffer(),
|
|
|
+ blocking,
|
|
|
+ buffer_offset.data(),
|
|
|
+ host_offset.data(),
|
|
|
+ region.data(),
|
|
|
+ buffer_row_pitch,
|
|
|
+ buffer_slice_pitch,
|
|
|
+ host_row_pitch,
|
|
|
+ host_slice_pitch,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_WRITE_BUFFER_RECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueCopyBufferRect(
|
|
|
+ const Buffer& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type src_row_pitch,
|
|
|
+ size_type src_slice_pitch,
|
|
|
+ size_type dst_row_pitch,
|
|
|
+ size_type dst_slice_pitch,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueCopyBufferRect(
|
|
|
+ object_,
|
|
|
+ src(),
|
|
|
+ dst(),
|
|
|
+ src_origin.data(),
|
|
|
+ dst_origin.data(),
|
|
|
+ region.data(),
|
|
|
+ src_row_pitch,
|
|
|
+ src_slice_pitch,
|
|
|
+ dst_row_pitch,
|
|
|
+ dst_slice_pitch,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQEUE_COPY_BUFFER_RECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ /**
|
|
|
+ * Enqueue a command to fill a buffer object with a pattern
|
|
|
+ * of a given size. The pattern is specified as a vector type.
|
|
|
+ * \tparam PatternType The datatype of the pattern field.
|
|
|
+ * The pattern type must be an accepted OpenCL data type.
|
|
|
+ * \tparam offset Is the offset in bytes into the buffer at
|
|
|
+ * which to start filling. This must be a multiple of
|
|
|
+ * the pattern size.
|
|
|
+ * \tparam size Is the size in bytes of the region to fill.
|
|
|
+ * This must be a multiple of the pattern size.
|
|
|
+ */
|
|
|
+ template<typename PatternType>
|
|
|
+ cl_int enqueueFillBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ PatternType pattern,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueFillBuffer(
|
|
|
+ object_,
|
|
|
+ buffer(),
|
|
|
+ static_cast<void*>(&pattern),
|
|
|
+ sizeof(PatternType),
|
|
|
+ offset,
|
|
|
+ size,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_FILL_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+ cl_int enqueueReadImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type row_pitch,
|
|
|
+ size_type slice_pitch,
|
|
|
+ void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueReadImage(
|
|
|
+ object_,
|
|
|
+ image(),
|
|
|
+ blocking,
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ row_pitch,
|
|
|
+ slice_pitch,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_READ_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueWriteImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type row_pitch,
|
|
|
+ size_type slice_pitch,
|
|
|
+ const void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueWriteImage(
|
|
|
+ object_,
|
|
|
+ image(),
|
|
|
+ blocking,
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ row_pitch,
|
|
|
+ slice_pitch,
|
|
|
+ ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_WRITE_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueCopyImage(
|
|
|
+ const Image& src,
|
|
|
+ const Image& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueCopyImage(
|
|
|
+ object_,
|
|
|
+ src(),
|
|
|
+ dst(),
|
|
|
+ src_origin.data(),
|
|
|
+ dst_origin.data(),
|
|
|
+ region.data(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_COPY_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ /**
|
|
|
+ * Enqueue a command to fill an image object with a specified color.
|
|
|
+ * \param fillColor is the color to use to fill the image.
|
|
|
+ * This is a four component RGBA floating-point color value if
|
|
|
+ * the image channel data type is not an unnormalized signed or
|
|
|
+ * unsigned data type.
|
|
|
+ */
|
|
|
+ cl_int enqueueFillImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_float4 fillColor,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueFillImage(
|
|
|
+ object_,
|
|
|
+ image(),
|
|
|
+ static_cast<void*>(&fillColor),
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_FILL_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueue a command to fill an image object with a specified color.
|
|
|
+ * \param fillColor is the color to use to fill the image.
|
|
|
+ * This is a four component RGBA signed integer color value if
|
|
|
+ * the image channel data type is an unnormalized signed integer
|
|
|
+ * type.
|
|
|
+ */
|
|
|
+ cl_int enqueueFillImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_int4 fillColor,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueFillImage(
|
|
|
+ object_,
|
|
|
+ image(),
|
|
|
+ static_cast<void*>(&fillColor),
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_FILL_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueue a command to fill an image object with a specified color.
|
|
|
+ * \param fillColor is the color to use to fill the image.
|
|
|
+ * This is a four component RGBA unsigned integer color value if
|
|
|
+ * the image channel data type is an unnormalized unsigned integer
|
|
|
+ * type.
|
|
|
+ */
|
|
|
+ cl_int enqueueFillImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_uint4 fillColor,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueFillImage(
|
|
|
+ object_,
|
|
|
+ image(),
|
|
|
+ static_cast<void*>(&fillColor),
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_FILL_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+ cl_int enqueueCopyImageToBuffer(
|
|
|
+ const Image& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type dst_offset,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueCopyImageToBuffer(
|
|
|
+ object_,
|
|
|
+ src(),
|
|
|
+ dst(),
|
|
|
+ src_origin.data(),
|
|
|
+ region.data(),
|
|
|
+ dst_offset,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_COPY_IMAGE_TO_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueCopyBufferToImage(
|
|
|
+ const Buffer& src,
|
|
|
+ const Image& dst,
|
|
|
+ size_type src_offset,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueCopyBufferToImage(
|
|
|
+ object_,
|
|
|
+ src(),
|
|
|
+ dst(),
|
|
|
+ src_offset,
|
|
|
+ dst_origin.data(),
|
|
|
+ region.data(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_COPY_BUFFER_TO_IMAGE_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ void* enqueueMapBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL,
|
|
|
+ cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int error;
|
|
|
+ void * result = ::clEnqueueMapBuffer(
|
|
|
+ object_, buffer(), blocking, flags, offset, size,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ if (event != NULL && error == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return result;
|
|
|
+ }
|
|
|
+
|
|
|
+ void* enqueueMapImage(
|
|
|
+ const Image& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type * row_pitch,
|
|
|
+ size_type * slice_pitch,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL,
|
|
|
+ cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int error;
|
|
|
+ void * result = ::clEnqueueMapImage(
|
|
|
+ object_, buffer(), blocking, flags,
|
|
|
+ origin.data(),
|
|
|
+ region.data(),
|
|
|
+ row_pitch, slice_pitch,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __ENQUEUE_MAP_IMAGE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ if (event != NULL && error == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+ return result;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host to update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a raw SVM pointer.
|
|
|
+ */
|
|
|
+ template<typename T>
|
|
|
+ cl_int enqueueMapSVM(
|
|
|
+ T* ptr,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(::clEnqueueSVMMap(
|
|
|
+ object_, blocking, flags, static_cast<void*>(ptr), size,
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host to update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a cl::pointer instance.
|
|
|
+ */
|
|
|
+ template<typename T, class D>
|
|
|
+ cl_int enqueueMapSVM(
|
|
|
+ cl::pointer<T, D> &ptr,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(::clEnqueueSVMMap(
|
|
|
+ object_, blocking, flags, static_cast<void*>(ptr.get()), size,
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host to update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a cl::vector instance.
|
|
|
+ */
|
|
|
+ template<typename T, class Alloc>
|
|
|
+ cl_int enqueueMapSVM(
|
|
|
+ cl::vector<T, Alloc> &container,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(::clEnqueueSVMMap(
|
|
|
+ object_, blocking, flags, static_cast<void*>(container.data()), container.size(),
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+ cl_int enqueueUnmapMemObject(
|
|
|
+ const Memory& memory,
|
|
|
+ void* mapped_ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueUnmapMemObject(
|
|
|
+ object_, memory(), mapped_ptr,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will release a coarse-grained SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a raw SVM pointer.
|
|
|
+ */
|
|
|
+ template<typename T>
|
|
|
+ cl_int enqueueUnmapSVM(
|
|
|
+ T* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueSVMUnmap(
|
|
|
+ object_, static_cast<void*>(ptr),
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will release a coarse-grained SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a cl::pointer instance.
|
|
|
+ */
|
|
|
+ template<typename T, class D>
|
|
|
+ cl_int enqueueUnmapSVM(
|
|
|
+ cl::pointer<T, D> &ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueSVMUnmap(
|
|
|
+ object_, static_cast<void*>(ptr.get()),
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will release a coarse-grained SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a cl::vector instance.
|
|
|
+ */
|
|
|
+ template<typename T, class Alloc>
|
|
|
+ cl_int enqueueUnmapSVM(
|
|
|
+ cl::vector<T, Alloc> &container,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueSVMUnmap(
|
|
|
+ object_, static_cast<void*>(container.data()),
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ /**
|
|
|
+ * Enqueues a marker command which waits for either a list of events to complete,
|
|
|
+ * or all previously enqueued commands to complete.
|
|
|
+ *
|
|
|
+ * Enqueues a marker command which waits for either a list of events to complete,
|
|
|
+ * or if the list is empty it waits for all commands previously enqueued in command_queue
|
|
|
+ * to complete before it completes. This command returns an event which can be waited on,
|
|
|
+ * i.e. this event can be waited on to insure that all events either in the event_wait_list
|
|
|
+ * or all previously enqueued commands, queued before this command to command_queue,
|
|
|
+ * have completed.
|
|
|
+ */
|
|
|
+ cl_int enqueueMarkerWithWaitList(
|
|
|
+ const vector<Event> *events = 0,
|
|
|
+ Event *event = 0) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueMarkerWithWaitList(
|
|
|
+ object_,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MARKER_WAIT_LIST_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * A synchronization point that enqueues a barrier operation.
|
|
|
+ *
|
|
|
+ * Enqueues a barrier command which waits for either a list of events to complete,
|
|
|
+ * or if the list is empty it waits for all commands previously enqueued in command_queue
|
|
|
+ * to complete before it completes. This command blocks command execution, that is, any
|
|
|
+ * following commands enqueued after it do not execute until it completes. This command
|
|
|
+ * returns an event which can be waited on, i.e. this event can be waited on to insure that
|
|
|
+ * all events either in the event_wait_list or all previously enqueued commands, queued
|
|
|
+ * before this command to command_queue, have completed.
|
|
|
+ */
|
|
|
+ cl_int enqueueBarrierWithWaitList(
|
|
|
+ const vector<Event> *events = 0,
|
|
|
+ Event *event = 0) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueBarrierWithWaitList(
|
|
|
+ object_,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_BARRIER_WAIT_LIST_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command to indicate with which device a set of memory objects
|
|
|
+ * should be associated.
|
|
|
+ */
|
|
|
+ cl_int enqueueMigrateMemObjects(
|
|
|
+ const vector<Memory> &memObjects,
|
|
|
+ cl_mem_migration_flags flags,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL
|
|
|
+ ) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+
|
|
|
+ vector<cl_mem> localMemObjects(memObjects.size());
|
|
|
+
|
|
|
+ for( int i = 0; i < (int)memObjects.size(); ++i ) {
|
|
|
+ localMemObjects[i] = memObjects[i]();
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueMigrateMemObjects(
|
|
|
+ object_,
|
|
|
+ (cl_uint)memObjects.size(),
|
|
|
+ localMemObjects.data(),
|
|
|
+ flags,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate ranges within a set of
|
|
|
+ * SVM allocations with a device.
|
|
|
+ * @param sizes - The length from each pointer to migrate.
|
|
|
+ */
|
|
|
+ template<typename T>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<T*> &svmRawPointers,
|
|
|
+ const cl::vector<size_type> &sizes,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(::clEnqueueSVMMigrateMem(
|
|
|
+ object_,
|
|
|
+ svmRawPointers.size(), static_cast<void**>(svmRawPointers.data()),
|
|
|
+ sizes.data(), // array of sizes not passed
|
|
|
+ flags,
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MIGRATE_SVM_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate a set of SVM allocations with
|
|
|
+ * a device.
|
|
|
+ */
|
|
|
+ template<typename T>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<T*> &svmRawPointers,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ return enqueueMigrateSVM(svmRawPointers, cl::vector<size_type>(svmRawPointers.size()), flags, events, event);
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate ranges within a set of
|
|
|
+ * SVM allocations with a device.
|
|
|
+ * @param sizes - The length from each pointer to migrate.
|
|
|
+ */
|
|
|
+ template<typename T, class D>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<cl::pointer<T, D>> &svmPointers,
|
|
|
+ const cl::vector<size_type> &sizes,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl::vector<void*> svmRawPointers;
|
|
|
+ svmRawPointers.reserve(svmPointers.size());
|
|
|
+ for (auto p : svmPointers) {
|
|
|
+ svmRawPointers.push_back(static_cast<void*>(p.get()));
|
|
|
+ }
|
|
|
+
|
|
|
+ return enqueueMigrateSVM(svmRawPointers, sizes, flags, events, event);
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate a set of SVM allocations with
|
|
|
+ * a device.
|
|
|
+ */
|
|
|
+ template<typename T, class D>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<cl::pointer<T, D>> &svmPointers,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ return enqueueMigrateSVM(svmPointers, cl::vector<size_type>(svmPointers.size()), flags, events, event);
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate ranges within a set of
|
|
|
+ * SVM allocations with a device.
|
|
|
+ * @param sizes - The length from the beginning of each container to migrate.
|
|
|
+ */
|
|
|
+ template<typename T, class Alloc>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<cl::vector<T, Alloc>> &svmContainers,
|
|
|
+ const cl::vector<size_type> &sizes,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl::vector<void*> svmRawPointers;
|
|
|
+ svmRawPointers.reserve(svmContainers.size());
|
|
|
+ for (auto p : svmContainers) {
|
|
|
+ svmRawPointers.push_back(static_cast<void*>(p.data()));
|
|
|
+ }
|
|
|
+
|
|
|
+ return enqueueMigrateSVM(svmRawPointers, sizes, flags, events, event);
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueues a command that will allow the host associate a set of SVM allocations with
|
|
|
+ * a device.
|
|
|
+ */
|
|
|
+ template<typename T, class Alloc>
|
|
|
+ cl_int enqueueMigrateSVM(
|
|
|
+ const cl::vector<cl::vector<T, Alloc>> &svmContainers,
|
|
|
+ cl_mem_migration_flags flags = 0,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ return enqueueMigrateSVM(svmContainers, cl::vector<size_type>(svmContainers.size()), flags, events, event);
|
|
|
+ }
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+
|
|
|
+ cl_int enqueueNDRangeKernel(
|
|
|
+ const Kernel& kernel,
|
|
|
+ const NDRange& offset,
|
|
|
+ const NDRange& global,
|
|
|
+ const NDRange& local = NullRange,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueNDRangeKernel(
|
|
|
+ object_, kernel(), (cl_uint) global.dimensions(),
|
|
|
+ offset.dimensions() != 0 ? (const size_type*) offset : NULL,
|
|
|
+ (const size_type*) global,
|
|
|
+ local.dimensions() != 0 ? (const size_type*) local : NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_NDRANGE_KERNEL_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_2_APIS)
|
|
|
+ CL_EXT_PREFIX__VERSION_1_2_DEPRECATED cl_int enqueueTask(
|
|
|
+ const Kernel& kernel,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const CL_EXT_SUFFIX__VERSION_1_2_DEPRECATED
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueTask(
|
|
|
+ object_, kernel(),
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_TASK_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif // #if defined(CL_USE_DEPRECATED_OPENCL_1_2_APIS)
|
|
|
+
|
|
|
+ cl_int enqueueNativeKernel(
|
|
|
+ void (CL_CALLBACK *userFptr)(void *),
|
|
|
+ std::pair<void*, size_type> args,
|
|
|
+ const vector<Memory>* mem_objects = NULL,
|
|
|
+ const vector<const void*>* mem_locs = NULL,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ size_type elements = 0;
|
|
|
+ if (mem_objects != NULL) {
|
|
|
+ elements = mem_objects->size();
|
|
|
+ }
|
|
|
+ vector<cl_mem> mems(elements);
|
|
|
+ for (unsigned int i = 0; i < elements; i++) {
|
|
|
+ mems[i] = ((*mem_objects)[i])();
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueNativeKernel(
|
|
|
+ object_, userFptr, args.first, args.second,
|
|
|
+ (mem_objects != NULL) ? (cl_uint) mem_objects->size() : 0,
|
|
|
+ mems.data(),
|
|
|
+ (mem_locs != NULL && mem_locs->size() > 0) ? (const void **) &mem_locs->front() : NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_NATIVE_KERNEL);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+/**
|
|
|
+ * Deprecated APIs for 1.2
|
|
|
+ */
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+ CL_EXT_PREFIX__VERSION_1_1_DEPRECATED
|
|
|
+ cl_int enqueueMarker(Event* event = NULL) const CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueMarker(
|
|
|
+ object_,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_MARKER_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ CL_EXT_PREFIX__VERSION_1_1_DEPRECATED
|
|
|
+ cl_int enqueueWaitForEvents(const vector<Event>& events) const CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clEnqueueWaitForEvents(
|
|
|
+ object_,
|
|
|
+ (cl_uint) events.size(),
|
|
|
+ events.size() > 0 ? (const cl_event*) &events.front() : NULL),
|
|
|
+ __ENQUEUE_WAIT_FOR_EVENTS_ERR);
|
|
|
+ }
|
|
|
+#endif // defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+
|
|
|
+ cl_int enqueueAcquireGLObjects(
|
|
|
+ const vector<Memory>* mem_objects = NULL,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueAcquireGLObjects(
|
|
|
+ object_,
|
|
|
+ (mem_objects != NULL) ? (cl_uint) mem_objects->size() : 0,
|
|
|
+ (mem_objects != NULL && mem_objects->size() > 0) ? (const cl_mem *) &mem_objects->front(): NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_ACQUIRE_GL_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueReleaseGLObjects(
|
|
|
+ const vector<Memory>* mem_objects = NULL,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueReleaseGLObjects(
|
|
|
+ object_,
|
|
|
+ (mem_objects != NULL) ? (cl_uint) mem_objects->size() : 0,
|
|
|
+ (mem_objects != NULL && mem_objects->size() > 0) ? (const cl_mem *) &mem_objects->front(): NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_RELEASE_GL_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+#if defined (CL_HPP_USE_DX_INTEROP)
|
|
|
+typedef CL_API_ENTRY cl_int (CL_API_CALL *PFN_clEnqueueAcquireD3D10ObjectsKHR)(
|
|
|
+ cl_command_queue command_queue, cl_uint num_objects,
|
|
|
+ const cl_mem* mem_objects, cl_uint num_events_in_wait_list,
|
|
|
+ const cl_event* event_wait_list, cl_event* event);
|
|
|
+typedef CL_API_ENTRY cl_int (CL_API_CALL *PFN_clEnqueueReleaseD3D10ObjectsKHR)(
|
|
|
+ cl_command_queue command_queue, cl_uint num_objects,
|
|
|
+ const cl_mem* mem_objects, cl_uint num_events_in_wait_list,
|
|
|
+ const cl_event* event_wait_list, cl_event* event);
|
|
|
+
|
|
|
+ cl_int enqueueAcquireD3D10Objects(
|
|
|
+ const vector<Memory>* mem_objects = NULL,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ static PFN_clEnqueueAcquireD3D10ObjectsKHR pfn_clEnqueueAcquireD3D10ObjectsKHR = NULL;
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ cl_context context = getInfo<CL_QUEUE_CONTEXT>();
|
|
|
+ cl::Device device(getInfo<CL_QUEUE_DEVICE>());
|
|
|
+ cl_platform_id platform = device.getInfo<CL_DEVICE_PLATFORM>();
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_(platform, clEnqueueAcquireD3D10ObjectsKHR);
|
|
|
+#endif
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clEnqueueAcquireD3D10ObjectsKHR);
|
|
|
+#endif
|
|
|
+
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ pfn_clEnqueueAcquireD3D10ObjectsKHR(
|
|
|
+ object_,
|
|
|
+ (mem_objects != NULL) ? (cl_uint) mem_objects->size() : 0,
|
|
|
+ (mem_objects != NULL && mem_objects->size() > 0) ? (const cl_mem *) &mem_objects->front(): NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_ACQUIRE_GL_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int enqueueReleaseD3D10Objects(
|
|
|
+ const vector<Memory>* mem_objects = NULL,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL) const
|
|
|
+ {
|
|
|
+ static PFN_clEnqueueReleaseD3D10ObjectsKHR pfn_clEnqueueReleaseD3D10ObjectsKHR = NULL;
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+ cl_context context = getInfo<CL_QUEUE_CONTEXT>();
|
|
|
+ cl::Device device(getInfo<CL_QUEUE_DEVICE>());
|
|
|
+ cl_platform_id platform = device.getInfo<CL_DEVICE_PLATFORM>();
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_(platform, clEnqueueReleaseD3D10ObjectsKHR);
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 120
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+ CL_HPP_INIT_CL_EXT_FCN_PTR_(clEnqueueReleaseD3D10ObjectsKHR);
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ pfn_clEnqueueReleaseD3D10ObjectsKHR(
|
|
|
+ object_,
|
|
|
+ (mem_objects != NULL) ? (cl_uint) mem_objects->size() : 0,
|
|
|
+ (mem_objects != NULL && mem_objects->size() > 0) ? (const cl_mem *) &mem_objects->front(): NULL,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_RELEASE_GL_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+ }
|
|
|
+#endif
|
|
|
+
|
|
|
+/**
|
|
|
+ * Deprecated APIs for 1.2
|
|
|
+ */
|
|
|
+#if defined(CL_USE_DEPRECATED_OPENCL_1_1_APIS)
|
|
|
+ CL_EXT_PREFIX__VERSION_1_1_DEPRECATED
|
|
|
+ cl_int enqueueBarrier() const CL_EXT_SUFFIX__VERSION_1_1_DEPRECATED
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ ::clEnqueueBarrier(object_),
|
|
|
+ __ENQUEUE_BARRIER_ERR);
|
|
|
+ }
|
|
|
+#endif // CL_USE_DEPRECATED_OPENCL_1_1_APIS
|
|
|
+
|
|
|
+ cl_int flush() const
|
|
|
+ {
|
|
|
+ return detail::errHandler(::clFlush(object_), __FLUSH_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_int finish() const
|
|
|
+ {
|
|
|
+ return detail::errHandler(::clFinish(object_), __FINISH_ERR);
|
|
|
+ }
|
|
|
+}; // CommandQueue
|
|
|
+
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ std::once_flag CommandQueue::default_initialized_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ CommandQueue CommandQueue::default_;
|
|
|
+CL_HPP_DEFINE_STATIC_MEMBER_ cl_int CommandQueue::default_error_ = CL_SUCCESS;
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+enum class DeviceQueueProperties : cl_command_queue_properties
|
|
|
+{
|
|
|
+ None = 0,
|
|
|
+ Profiling = CL_QUEUE_PROFILING_ENABLE,
|
|
|
+};
|
|
|
+
|
|
|
+inline DeviceQueueProperties operator|(DeviceQueueProperties lhs, DeviceQueueProperties rhs)
|
|
|
+{
|
|
|
+ return static_cast<DeviceQueueProperties>(static_cast<cl_command_queue_properties>(lhs) | static_cast<cl_command_queue_properties>(rhs));
|
|
|
+}
|
|
|
+
|
|
|
+/*! \class DeviceCommandQueue
|
|
|
+ * \brief DeviceCommandQueue interface for device cl_command_queues.
|
|
|
+ */
|
|
|
+class DeviceCommandQueue : public detail::Wrapper<cl_command_queue>
|
|
|
+{
|
|
|
+public:
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Trivial empty constructor to create a null queue.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue() { }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Default construct device command queue on default context and device
|
|
|
+ */
|
|
|
+ DeviceCommandQueue(DeviceQueueProperties properties, cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl::Context context = cl::Context::getDefault();
|
|
|
+ cl::Device device = cl::Device::getDefault();
|
|
|
+
|
|
|
+ cl_command_queue_properties mergedProperties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | static_cast<cl_command_queue_properties>(properties);
|
|
|
+
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, mergedProperties, 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Create a device command queue for a specified device in the passed context.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ const Device& device,
|
|
|
+ DeviceQueueProperties properties = DeviceQueueProperties::None,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_command_queue_properties mergedProperties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | static_cast<cl_command_queue_properties>(properties);
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, mergedProperties, 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Create a device command queue for a specified device in the passed context.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue(
|
|
|
+ const Context& context,
|
|
|
+ const Device& device,
|
|
|
+ cl_uint queueSize,
|
|
|
+ DeviceQueueProperties properties = DeviceQueueProperties::None,
|
|
|
+ cl_int* err = NULL)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_command_queue_properties mergedProperties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | static_cast<cl_command_queue_properties>(properties);
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, mergedProperties,
|
|
|
+ CL_QUEUE_SIZE, queueSize,
|
|
|
+ 0 };
|
|
|
+ object_ = ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Constructor from cl_command_queue - takes ownership.
|
|
|
+ *
|
|
|
+ * \param retainObject will cause the constructor to retain its cl object.
|
|
|
+ * Defaults to false to maintain compatibility with
|
|
|
+ * earlier versions.
|
|
|
+ */
|
|
|
+ explicit DeviceCommandQueue(const cl_command_queue& commandQueue, bool retainObject = false) :
|
|
|
+ detail::Wrapper<cl_type>(commandQueue, retainObject) { }
|
|
|
+
|
|
|
+ DeviceCommandQueue& operator = (const cl_command_queue& rhs)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(rhs);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Copy constructor to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue(const DeviceCommandQueue& queue) : detail::Wrapper<cl_type>(queue) {}
|
|
|
+
|
|
|
+ /*! \brief Copy assignment to forward copy to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue& operator = (const DeviceCommandQueue &queue)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(queue);
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*! \brief Move constructor to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue(DeviceCommandQueue&& queue) CL_HPP_NOEXCEPT_ : detail::Wrapper<cl_type>(std::move(queue)) {}
|
|
|
+
|
|
|
+ /*! \brief Move assignment to forward move to the superclass correctly.
|
|
|
+ * Required for MSVC.
|
|
|
+ */
|
|
|
+ DeviceCommandQueue& operator = (DeviceCommandQueue &&queue)
|
|
|
+ {
|
|
|
+ detail::Wrapper<cl_type>::operator=(std::move(queue));
|
|
|
+ return *this;
|
|
|
+ }
|
|
|
+
|
|
|
+ template <typename T>
|
|
|
+ cl_int getInfo(cl_command_queue_info name, T* param) const
|
|
|
+ {
|
|
|
+ return detail::errHandler(
|
|
|
+ detail::getInfo(
|
|
|
+ &::clGetCommandQueueInfo, object_, name, param),
|
|
|
+ __GET_COMMAND_QUEUE_INFO_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ template <cl_int name> typename
|
|
|
+ detail::param_traits<detail::cl_command_queue_info, name>::param_type
|
|
|
+ getInfo(cl_int* err = NULL) const
|
|
|
+ {
|
|
|
+ typename detail::param_traits<
|
|
|
+ detail::cl_command_queue_info, name>::param_type param;
|
|
|
+ cl_int result = getInfo(name, ¶m);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = result;
|
|
|
+ }
|
|
|
+ return param;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Create a new default device command queue for the default device,
|
|
|
+ * in the default context and of the default size.
|
|
|
+ * If there is already a default queue for the specified device this
|
|
|
+ * function will return the pre-existing queue.
|
|
|
+ */
|
|
|
+ static DeviceCommandQueue makeDefault(
|
|
|
+ cl_int *err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ cl::Context context = cl::Context::getDefault();
|
|
|
+ cl::Device device = cl::Device::getDefault();
|
|
|
+
|
|
|
+ cl_command_queue_properties properties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | CL_QUEUE_ON_DEVICE_DEFAULT;
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties,
|
|
|
+ 0 };
|
|
|
+ DeviceCommandQueue deviceQueue(
|
|
|
+ ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error));
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return deviceQueue;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Create a new default device command queue for the specified device
|
|
|
+ * and of the default size.
|
|
|
+ * If there is already a default queue for the specified device this
|
|
|
+ * function will return the pre-existing queue.
|
|
|
+ */
|
|
|
+ static DeviceCommandQueue makeDefault(
|
|
|
+ const Context &context, const Device &device, cl_int *err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_command_queue_properties properties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | CL_QUEUE_ON_DEVICE_DEFAULT;
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties,
|
|
|
+ 0 };
|
|
|
+ DeviceCommandQueue deviceQueue(
|
|
|
+ ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error));
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return deviceQueue;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Create a new default device command queue for the specified device
|
|
|
+ * and of the requested size in bytes.
|
|
|
+ * If there is already a default queue for the specified device this
|
|
|
+ * function will return the pre-existing queue.
|
|
|
+ */
|
|
|
+ static DeviceCommandQueue makeDefault(
|
|
|
+ const Context &context, const Device &device, cl_uint queueSize, cl_int *err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_command_queue_properties properties =
|
|
|
+ CL_QUEUE_OUT_OF_ORDER_EXEC_MODE_ENABLE | CL_QUEUE_ON_DEVICE | CL_QUEUE_ON_DEVICE_DEFAULT;
|
|
|
+ cl_queue_properties queue_properties[] = {
|
|
|
+ CL_QUEUE_PROPERTIES, properties,
|
|
|
+ CL_QUEUE_SIZE, queueSize,
|
|
|
+ 0 };
|
|
|
+ DeviceCommandQueue deviceQueue(
|
|
|
+ ::clCreateCommandQueueWithProperties(
|
|
|
+ context(), device(), queue_properties, &error));
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return deviceQueue;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+ /*!
|
|
|
+ * Modify the default device command queue to be used for subsequent kernels.
|
|
|
+ * This can update the default command queue for a device repeatedly to account
|
|
|
+ * for kernels that rely on the default.
|
|
|
+ * @return updated default device command queue.
|
|
|
+ */
|
|
|
+ static DeviceCommandQueue updateDefault(const Context &context, const Device &device, const DeviceCommandQueue &default_queue, cl_int *err = nullptr)
|
|
|
+ {
|
|
|
+ cl_int error;
|
|
|
+ error = clSetDefaultDeviceCommandQueue(context.get(), device.get(), default_queue.get());
|
|
|
+
|
|
|
+ detail::errHandler(error, __SET_DEFAULT_DEVICE_COMMAND_QUEUE_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return default_queue;
|
|
|
+ }
|
|
|
+
|
|
|
+ /*!
|
|
|
+ * Return the current default command queue for the specified command queue
|
|
|
+ */
|
|
|
+ static DeviceCommandQueue getDefault(const CommandQueue &queue, cl_int * err = NULL)
|
|
|
+ {
|
|
|
+ return queue.getInfo<CL_QUEUE_DEVICE_DEFAULT>(err);
|
|
|
+ }
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 210
|
|
|
+}; // DeviceCommandQueue
|
|
|
+
|
|
|
+namespace detail
|
|
|
+{
|
|
|
+ // Specialization for device command queue
|
|
|
+ template <>
|
|
|
+ struct KernelArgumentHandler<cl::DeviceCommandQueue, void>
|
|
|
+ {
|
|
|
+ static size_type size(const cl::DeviceCommandQueue&) { return sizeof(cl_command_queue); }
|
|
|
+ static const cl_command_queue* ptr(const cl::DeviceCommandQueue& value) { return &(value()); }
|
|
|
+ };
|
|
|
+} // namespace detail
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+
|
|
|
+template< typename IteratorType >
|
|
|
+Buffer::Buffer(
|
|
|
+ const Context &context,
|
|
|
+ IteratorType startIterator,
|
|
|
+ IteratorType endIterator,
|
|
|
+ bool readOnly,
|
|
|
+ bool useHostPtr,
|
|
|
+ cl_int* err)
|
|
|
+{
|
|
|
+ typedef typename std::iterator_traits<IteratorType>::value_type DataType;
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_mem_flags flags = 0;
|
|
|
+ if( readOnly ) {
|
|
|
+ flags |= CL_MEM_READ_ONLY;
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ flags |= CL_MEM_READ_WRITE;
|
|
|
+ }
|
|
|
+ if( useHostPtr ) {
|
|
|
+ flags |= CL_MEM_USE_HOST_PTR;
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type size = sizeof(DataType)*(endIterator - startIterator);
|
|
|
+
|
|
|
+ if( useHostPtr ) {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, static_cast<DataType*>(&*startIterator), &error);
|
|
|
+ } else {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, 0, &error);
|
|
|
+ }
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ if( !useHostPtr ) {
|
|
|
+ CommandQueue queue(context, 0, &error);
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ error = cl::copy(queue, startIterator, endIterator, *this);
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+template< typename IteratorType >
|
|
|
+Buffer::Buffer(
|
|
|
+ const CommandQueue &queue,
|
|
|
+ IteratorType startIterator,
|
|
|
+ IteratorType endIterator,
|
|
|
+ bool readOnly,
|
|
|
+ bool useHostPtr,
|
|
|
+ cl_int* err)
|
|
|
+{
|
|
|
+ typedef typename std::iterator_traits<IteratorType>::value_type DataType;
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ cl_mem_flags flags = 0;
|
|
|
+ if (readOnly) {
|
|
|
+ flags |= CL_MEM_READ_ONLY;
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ flags |= CL_MEM_READ_WRITE;
|
|
|
+ }
|
|
|
+ if (useHostPtr) {
|
|
|
+ flags |= CL_MEM_USE_HOST_PTR;
|
|
|
+ }
|
|
|
+
|
|
|
+ size_type size = sizeof(DataType)*(endIterator - startIterator);
|
|
|
+
|
|
|
+ Context context = queue.getInfo<CL_QUEUE_CONTEXT>();
|
|
|
+
|
|
|
+ if (useHostPtr) {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, static_cast<DataType*>(&*startIterator), &error);
|
|
|
+ }
|
|
|
+ else {
|
|
|
+ object_ = ::clCreateBuffer(context(), flags, size, 0, &error);
|
|
|
+ }
|
|
|
+
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ if (!useHostPtr) {
|
|
|
+ error = cl::copy(queue, startIterator, endIterator, *this);
|
|
|
+ detail::errHandler(error, __CREATE_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueReadBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueReadBuffer(buffer, blocking, offset, size, ptr, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueWriteBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ const void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueWriteBuffer(buffer, blocking, offset, size, ptr, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+inline void* enqueueMapBuffer(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type offset,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL,
|
|
|
+ cl_int* err = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+
|
|
|
+ void * result = ::clEnqueueMapBuffer(
|
|
|
+ queue(), buffer(), blocking, flags, offset, size,
|
|
|
+ (events != NULL) ? (cl_uint) events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*) &events->front() : NULL,
|
|
|
+ (cl_event*) event,
|
|
|
+ &error);
|
|
|
+
|
|
|
+ detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ if (err != NULL) {
|
|
|
+ *err = error;
|
|
|
+ }
|
|
|
+ return result;
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will allow the host to
|
|
|
+ * update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a raw SVM pointer.
|
|
|
+ */
|
|
|
+template<typename T>
|
|
|
+inline cl_int enqueueMapSVM(
|
|
|
+ T* ptr,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events,
|
|
|
+ Event* event)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueMapSVM(
|
|
|
+ ptr, blocking, flags, size, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will allow the host to
|
|
|
+ * update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a cl::pointer instance.
|
|
|
+ */
|
|
|
+template<typename T, class D>
|
|
|
+inline cl_int enqueueMapSVM(
|
|
|
+ cl::pointer<T, D> ptr,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueMapSVM(
|
|
|
+ ptr, blocking, flags, size, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will allow the host to
|
|
|
+ * update a region of a coarse-grained SVM buffer.
|
|
|
+ * This variant takes a cl::vector instance.
|
|
|
+ */
|
|
|
+template<typename T, class Alloc>
|
|
|
+inline cl_int enqueueMapSVM(
|
|
|
+ cl::vector<T, Alloc> container,
|
|
|
+ cl_bool blocking,
|
|
|
+ cl_map_flags flags,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueMapSVM(
|
|
|
+ container, blocking, flags, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+inline cl_int enqueueUnmapMemObject(
|
|
|
+ const Memory& memory,
|
|
|
+ void* mapped_ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ detail::errHandler(error, __ENQUEUE_MAP_BUFFER_ERR);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ cl_event tmp;
|
|
|
+ cl_int err = detail::errHandler(
|
|
|
+ ::clEnqueueUnmapMemObject(
|
|
|
+ queue(), memory(), mapped_ptr,
|
|
|
+ (events != NULL) ? (cl_uint)events->size() : 0,
|
|
|
+ (events != NULL && events->size() > 0) ? (cl_event*)&events->front() : NULL,
|
|
|
+ (event != NULL) ? &tmp : NULL),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+ if (event != NULL && err == CL_SUCCESS)
|
|
|
+ *event = tmp;
|
|
|
+
|
|
|
+ return err;
|
|
|
+}
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will release a coarse-grained
|
|
|
+ * SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a raw SVM pointer.
|
|
|
+ */
|
|
|
+template<typename T>
|
|
|
+inline cl_int enqueueUnmapSVM(
|
|
|
+ T* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return detail::errHandler(queue.enqueueUnmapSVM(ptr, events, event),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will release a coarse-grained
|
|
|
+ * SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a cl::pointer instance.
|
|
|
+ */
|
|
|
+template<typename T, class D>
|
|
|
+inline cl_int enqueueUnmapSVM(
|
|
|
+ cl::pointer<T, D> &ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return detail::errHandler(queue.enqueueUnmapSVM(ptr, events, event),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Enqueues to the default queue a command that will release a coarse-grained
|
|
|
+ * SVM buffer back to the OpenCL runtime.
|
|
|
+ * This variant takes a cl::vector instance.
|
|
|
+ */
|
|
|
+template<typename T, class Alloc>
|
|
|
+inline cl_int enqueueUnmapSVM(
|
|
|
+ cl::vector<T, Alloc> &container,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return detail::errHandler(error, __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+ }
|
|
|
+
|
|
|
+ return detail::errHandler(queue.enqueueUnmapSVM(container, events, event),
|
|
|
+ __ENQUEUE_UNMAP_MEM_OBJECT_ERR);
|
|
|
+}
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+inline cl_int enqueueCopyBuffer(
|
|
|
+ const Buffer& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ size_type src_offset,
|
|
|
+ size_type dst_offset,
|
|
|
+ size_type size,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueCopyBuffer(src, dst, src_offset, dst_offset, size, events, event);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Blocking copy operation between iterators and a buffer.
|
|
|
+ * Host to Device.
|
|
|
+ * Uses default command queue.
|
|
|
+ */
|
|
|
+template< typename IteratorType >
|
|
|
+inline cl_int copy( IteratorType startIterator, IteratorType endIterator, cl::Buffer &buffer )
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS)
|
|
|
+ return error;
|
|
|
+
|
|
|
+ return cl::copy(queue, startIterator, endIterator, buffer);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Blocking copy operation between iterators and a buffer.
|
|
|
+ * Device to Host.
|
|
|
+ * Uses default command queue.
|
|
|
+ */
|
|
|
+template< typename IteratorType >
|
|
|
+inline cl_int copy( const cl::Buffer &buffer, IteratorType startIterator, IteratorType endIterator )
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+ if (error != CL_SUCCESS)
|
|
|
+ return error;
|
|
|
+
|
|
|
+ return cl::copy(queue, buffer, startIterator, endIterator);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Blocking copy operation between iterators and a buffer.
|
|
|
+ * Host to Device.
|
|
|
+ * Uses specified queue.
|
|
|
+ */
|
|
|
+template< typename IteratorType >
|
|
|
+inline cl_int copy( const CommandQueue &queue, IteratorType startIterator, IteratorType endIterator, cl::Buffer &buffer )
|
|
|
+{
|
|
|
+ typedef typename std::iterator_traits<IteratorType>::value_type DataType;
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ size_type length = endIterator-startIterator;
|
|
|
+ size_type byteLength = length*sizeof(DataType);
|
|
|
+
|
|
|
+ DataType *pointer =
|
|
|
+ static_cast<DataType*>(queue.enqueueMapBuffer(buffer, CL_TRUE, CL_MAP_WRITE, 0, byteLength, 0, 0, &error));
|
|
|
+ // if exceptions enabled, enqueueMapBuffer will throw
|
|
|
+ if( error != CL_SUCCESS ) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+#if defined(_MSC_VER)
|
|
|
+ std::copy(
|
|
|
+ startIterator,
|
|
|
+ endIterator,
|
|
|
+ stdext::checked_array_iterator<DataType*>(
|
|
|
+ pointer, length));
|
|
|
+#else
|
|
|
+ std::copy(startIterator, endIterator, pointer);
|
|
|
+#endif
|
|
|
+ Event endEvent;
|
|
|
+ error = queue.enqueueUnmapMemObject(buffer, pointer, 0, &endEvent);
|
|
|
+ // if exceptions enabled, enqueueUnmapMemObject will throw
|
|
|
+ if( error != CL_SUCCESS ) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+ endEvent.wait();
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+ * Blocking copy operation between iterators and a buffer.
|
|
|
+ * Device to Host.
|
|
|
+ * Uses specified queue.
|
|
|
+ */
|
|
|
+template< typename IteratorType >
|
|
|
+inline cl_int copy( const CommandQueue &queue, const cl::Buffer &buffer, IteratorType startIterator, IteratorType endIterator )
|
|
|
+{
|
|
|
+ typedef typename std::iterator_traits<IteratorType>::value_type DataType;
|
|
|
+ cl_int error;
|
|
|
+
|
|
|
+ size_type length = endIterator-startIterator;
|
|
|
+ size_type byteLength = length*sizeof(DataType);
|
|
|
+
|
|
|
+ DataType *pointer =
|
|
|
+ static_cast<DataType*>(queue.enqueueMapBuffer(buffer, CL_TRUE, CL_MAP_READ, 0, byteLength, 0, 0, &error));
|
|
|
+ // if exceptions enabled, enqueueMapBuffer will throw
|
|
|
+ if( error != CL_SUCCESS ) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+ std::copy(pointer, pointer + length, startIterator);
|
|
|
+ Event endEvent;
|
|
|
+ error = queue.enqueueUnmapMemObject(buffer, pointer, 0, &endEvent);
|
|
|
+ // if exceptions enabled, enqueueUnmapMemObject will throw
|
|
|
+ if( error != CL_SUCCESS ) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+ endEvent.wait();
|
|
|
+ return CL_SUCCESS;
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+/**
|
|
|
+ * Blocking SVM map operation - performs a blocking map underneath.
|
|
|
+ */
|
|
|
+template<typename T, class Alloc>
|
|
|
+inline cl_int mapSVM(cl::vector<T, Alloc> &container)
|
|
|
+{
|
|
|
+ return enqueueMapSVM(container, CL_TRUE, CL_MAP_READ | CL_MAP_WRITE);
|
|
|
+}
|
|
|
+
|
|
|
+/**
|
|
|
+* Blocking SVM map operation - performs a blocking map underneath.
|
|
|
+*/
|
|
|
+template<typename T, class Alloc>
|
|
|
+inline cl_int unmapSVM(cl::vector<T, Alloc> &container)
|
|
|
+{
|
|
|
+ return enqueueUnmapSVM(container);
|
|
|
+}
|
|
|
+
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+inline cl_int enqueueReadBufferRect(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& buffer_offset,
|
|
|
+ const array<size_type, 3>& host_offset,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type buffer_row_pitch,
|
|
|
+ size_type buffer_slice_pitch,
|
|
|
+ size_type host_row_pitch,
|
|
|
+ size_type host_slice_pitch,
|
|
|
+ void *ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueReadBufferRect(
|
|
|
+ buffer,
|
|
|
+ blocking,
|
|
|
+ buffer_offset,
|
|
|
+ host_offset,
|
|
|
+ region,
|
|
|
+ buffer_row_pitch,
|
|
|
+ buffer_slice_pitch,
|
|
|
+ host_row_pitch,
|
|
|
+ host_slice_pitch,
|
|
|
+ ptr,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueWriteBufferRect(
|
|
|
+ const Buffer& buffer,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& buffer_offset,
|
|
|
+ const array<size_type, 3>& host_offset,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type buffer_row_pitch,
|
|
|
+ size_type buffer_slice_pitch,
|
|
|
+ size_type host_row_pitch,
|
|
|
+ size_type host_slice_pitch,
|
|
|
+ const void *ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueWriteBufferRect(
|
|
|
+ buffer,
|
|
|
+ blocking,
|
|
|
+ buffer_offset,
|
|
|
+ host_offset,
|
|
|
+ region,
|
|
|
+ buffer_row_pitch,
|
|
|
+ buffer_slice_pitch,
|
|
|
+ host_row_pitch,
|
|
|
+ host_slice_pitch,
|
|
|
+ ptr,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueCopyBufferRect(
|
|
|
+ const Buffer& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type src_row_pitch,
|
|
|
+ size_type src_slice_pitch,
|
|
|
+ size_type dst_row_pitch,
|
|
|
+ size_type dst_slice_pitch,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueCopyBufferRect(
|
|
|
+ src,
|
|
|
+ dst,
|
|
|
+ src_origin,
|
|
|
+ dst_origin,
|
|
|
+ region,
|
|
|
+ src_row_pitch,
|
|
|
+ src_slice_pitch,
|
|
|
+ dst_row_pitch,
|
|
|
+ dst_slice_pitch,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+#endif // CL_HPP_TARGET_OPENCL_VERSION >= 110
|
|
|
+
|
|
|
+inline cl_int enqueueReadImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type row_pitch,
|
|
|
+ size_type slice_pitch,
|
|
|
+ void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueReadImage(
|
|
|
+ image,
|
|
|
+ blocking,
|
|
|
+ origin,
|
|
|
+ region,
|
|
|
+ row_pitch,
|
|
|
+ slice_pitch,
|
|
|
+ ptr,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueWriteImage(
|
|
|
+ const Image& image,
|
|
|
+ cl_bool blocking,
|
|
|
+ const array<size_type, 3>& origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type row_pitch,
|
|
|
+ size_type slice_pitch,
|
|
|
+ const void* ptr,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueWriteImage(
|
|
|
+ image,
|
|
|
+ blocking,
|
|
|
+ origin,
|
|
|
+ region,
|
|
|
+ row_pitch,
|
|
|
+ slice_pitch,
|
|
|
+ ptr,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueCopyImage(
|
|
|
+ const Image& src,
|
|
|
+ const Image& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueCopyImage(
|
|
|
+ src,
|
|
|
+ dst,
|
|
|
+ src_origin,
|
|
|
+ dst_origin,
|
|
|
+ region,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueCopyImageToBuffer(
|
|
|
+ const Image& src,
|
|
|
+ const Buffer& dst,
|
|
|
+ const array<size_type, 3>& src_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ size_type dst_offset,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueCopyImageToBuffer(
|
|
|
+ src,
|
|
|
+ dst,
|
|
|
+ src_origin,
|
|
|
+ region,
|
|
|
+ dst_offset,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int enqueueCopyBufferToImage(
|
|
|
+ const Buffer& src,
|
|
|
+ const Image& dst,
|
|
|
+ size_type src_offset,
|
|
|
+ const array<size_type, 3>& dst_origin,
|
|
|
+ const array<size_type, 3>& region,
|
|
|
+ const vector<Event>* events = NULL,
|
|
|
+ Event* event = NULL)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.enqueueCopyBufferToImage(
|
|
|
+ src,
|
|
|
+ dst,
|
|
|
+ src_offset,
|
|
|
+ dst_origin,
|
|
|
+ region,
|
|
|
+ events,
|
|
|
+ event);
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+inline cl_int flush(void)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+ return queue.flush();
|
|
|
+}
|
|
|
+
|
|
|
+inline cl_int finish(void)
|
|
|
+{
|
|
|
+ cl_int error;
|
|
|
+ CommandQueue queue = CommandQueue::getDefault(&error);
|
|
|
+
|
|
|
+ if (error != CL_SUCCESS) {
|
|
|
+ return error;
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+ return queue.finish();
|
|
|
+}
|
|
|
+
|
|
|
+class EnqueueArgs
|
|
|
+{
|
|
|
+private:
|
|
|
+ CommandQueue queue_;
|
|
|
+ const NDRange offset_;
|
|
|
+ const NDRange global_;
|
|
|
+ const NDRange local_;
|
|
|
+ vector<Event> events_;
|
|
|
+
|
|
|
+ template<typename... Ts>
|
|
|
+ friend class KernelFunctor;
|
|
|
+
|
|
|
+public:
|
|
|
+ EnqueueArgs(NDRange global) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(Event e, NDRange global) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(Event e, NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(Event e, NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(const vector<Event> &events, NDRange global) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(const vector<Event> &events, NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(const vector<Event> &events, NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(CommandQueue::getDefault()),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, NDRange global) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, Event e, NDRange global) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, Event e, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, Event e, NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local)
|
|
|
+ {
|
|
|
+ events_.push_back(e);
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, const vector<Event> &events, NDRange global) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(NullRange),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, const vector<Event> &events, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(NullRange),
|
|
|
+ global_(global),
|
|
|
+ local_(local),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+
|
|
|
+ EnqueueArgs(CommandQueue &queue, const vector<Event> &events, NDRange offset, NDRange global, NDRange local) :
|
|
|
+ queue_(queue),
|
|
|
+ offset_(offset),
|
|
|
+ global_(global),
|
|
|
+ local_(local),
|
|
|
+ events_(events)
|
|
|
+ {
|
|
|
+
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+
|
|
|
+//----------------------------------------------------------------------------------------------
|
|
|
+
|
|
|
+
|
|
|
+/**
|
|
|
+ * Type safe kernel functor.
|
|
|
+ *
|
|
|
+ */
|
|
|
+template<typename... Ts>
|
|
|
+class KernelFunctor
|
|
|
+{
|
|
|
+private:
|
|
|
+ Kernel kernel_;
|
|
|
+
|
|
|
+ template<int index, typename T0, typename... T1s>
|
|
|
+ void setArgs(T0&& t0, T1s&&... t1s)
|
|
|
+ {
|
|
|
+ kernel_.setArg(index, t0);
|
|
|
+ setArgs<index + 1, T1s...>(std::forward<T1s>(t1s)...);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<int index, typename T0>
|
|
|
+ void setArgs(T0&& t0)
|
|
|
+ {
|
|
|
+ kernel_.setArg(index, t0);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<int index>
|
|
|
+ void setArgs()
|
|
|
+ {
|
|
|
+ }
|
|
|
+
|
|
|
+
|
|
|
+public:
|
|
|
+ KernelFunctor(Kernel kernel) : kernel_(kernel)
|
|
|
+ {}
|
|
|
+
|
|
|
+ KernelFunctor(
|
|
|
+ const Program& program,
|
|
|
+ const string name,
|
|
|
+ cl_int * err = NULL) :
|
|
|
+ kernel_(program, name.c_str(), err)
|
|
|
+ {}
|
|
|
+
|
|
|
+ //! \brief Return type of the functor
|
|
|
+ typedef Event result_type;
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueue kernel.
|
|
|
+ * @param args Launch parameters of the kernel.
|
|
|
+ * @param t0... List of kernel arguments based on the template type of the functor.
|
|
|
+ */
|
|
|
+ Event operator() (
|
|
|
+ const EnqueueArgs& args,
|
|
|
+ Ts... ts)
|
|
|
+ {
|
|
|
+ Event event;
|
|
|
+ setArgs<0>(std::forward<Ts>(ts)...);
|
|
|
+
|
|
|
+ args.queue_.enqueueNDRangeKernel(
|
|
|
+ kernel_,
|
|
|
+ args.offset_,
|
|
|
+ args.global_,
|
|
|
+ args.local_,
|
|
|
+ &args.events_,
|
|
|
+ &event);
|
|
|
+
|
|
|
+ return event;
|
|
|
+ }
|
|
|
+
|
|
|
+ /**
|
|
|
+ * Enqueue kernel with support for error code.
|
|
|
+ * @param args Launch parameters of the kernel.
|
|
|
+ * @param t0... List of kernel arguments based on the template type of the functor.
|
|
|
+ * @param error Out parameter returning the error code from the execution.
|
|
|
+ */
|
|
|
+ Event operator() (
|
|
|
+ const EnqueueArgs& args,
|
|
|
+ Ts... ts,
|
|
|
+ cl_int &error)
|
|
|
+ {
|
|
|
+ Event event;
|
|
|
+ setArgs<0>(std::forward<Ts>(ts)...);
|
|
|
+
|
|
|
+ error = args.queue_.enqueueNDRangeKernel(
|
|
|
+ kernel_,
|
|
|
+ args.offset_,
|
|
|
+ args.global_,
|
|
|
+ args.local_,
|
|
|
+ &args.events_,
|
|
|
+ &event);
|
|
|
+
|
|
|
+ return event;
|
|
|
+ }
|
|
|
+
|
|
|
+#if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+ cl_int setSVMPointers(const vector<void*> &pointerList)
|
|
|
+ {
|
|
|
+ return kernel_.setSVMPointers(pointerList);
|
|
|
+ }
|
|
|
+
|
|
|
+ template<typename T0, typename... T1s>
|
|
|
+ cl_int setSVMPointers(const T0 &t0, T1s &... ts)
|
|
|
+ {
|
|
|
+ return kernel_.setSVMPointers(t0, ts...);
|
|
|
+ }
|
|
|
+#endif // #if CL_HPP_TARGET_OPENCL_VERSION >= 200
|
|
|
+
|
|
|
+ Kernel getKernel()
|
|
|
+ {
|
|
|
+ return kernel_;
|
|
|
+ }
|
|
|
+};
|
|
|
+
|
|
|
+namespace compatibility {
|
|
|
+ /**
|
|
|
+ * Backward compatibility class to ensure that cl.hpp code works with cl2.hpp.
|
|
|
+ * Please use KernelFunctor directly.
|
|
|
+ */
|
|
|
+ template<typename... Ts>
|
|
|
+ struct make_kernel
|
|
|
+ {
|
|
|
+ typedef KernelFunctor<Ts...> FunctorType;
|
|
|
+
|
|
|
+ FunctorType functor_;
|
|
|
+
|
|
|
+ make_kernel(
|
|
|
+ const Program& program,
|
|
|
+ const string name,
|
|
|
+ cl_int * err = NULL) :
|
|
|
+ functor_(FunctorType(program, name, err))
|
|
|
+ {}
|
|
|
+
|
|
|
+ make_kernel(
|
|
|
+ const Kernel kernel) :
|
|
|
+ functor_(FunctorType(kernel))
|
|
|
+ {}
|
|
|
+
|
|
|
+ //! \brief Return type of the functor
|
|
|
+ typedef Event result_type;
|
|
|
+
|
|
|
+ //! \brief Function signature of kernel functor with no event dependency.
|
|
|
+ typedef Event type_(
|
|
|
+ const EnqueueArgs&,
|
|
|
+ Ts...);
|
|
|
+
|
|
|
+ Event operator()(
|
|
|
+ const EnqueueArgs& enqueueArgs,
|
|
|
+ Ts... args)
|
|
|
+ {
|
|
|
+ return functor_(
|
|
|
+ enqueueArgs, args...);
|
|
|
+ }
|
|
|
+ };
|
|
|
+} // namespace compatibility
|
|
|
+
|
|
|
+
|
|
|
+//----------------------------------------------------------------------------------------------------------------------
|
|
|
+
|
|
|
+#undef CL_HPP_ERR_STR_
|
|
|
+#if !defined(CL_HPP_USER_OVERRIDE_ERROR_STRINGS)
|
|
|
+#undef __GET_DEVICE_INFO_ERR
|
|
|
+#undef __GET_PLATFORM_INFO_ERR
|
|
|
+#undef __GET_DEVICE_IDS_ERR
|
|
|
+#undef __GET_PLATFORM_IDS_ERR
|
|
|
+#undef __GET_CONTEXT_INFO_ERR
|
|
|
+#undef __GET_EVENT_INFO_ERR
|
|
|
+#undef __GET_EVENT_PROFILE_INFO_ERR
|
|
|
+#undef __GET_MEM_OBJECT_INFO_ERR
|
|
|
+#undef __GET_IMAGE_INFO_ERR
|
|
|
+#undef __GET_SAMPLER_INFO_ERR
|
|
|
+#undef __GET_KERNEL_INFO_ERR
|
|
|
+#undef __GET_KERNEL_ARG_INFO_ERR
|
|
|
+#undef __GET_KERNEL_SUB_GROUP_INFO_ERR
|
|
|
+#undef __GET_KERNEL_WORK_GROUP_INFO_ERR
|
|
|
+#undef __GET_PROGRAM_INFO_ERR
|
|
|
+#undef __GET_PROGRAM_BUILD_INFO_ERR
|
|
|
+#undef __GET_COMMAND_QUEUE_INFO_ERR
|
|
|
+#undef __CREATE_CONTEXT_ERR
|
|
|
+#undef __CREATE_CONTEXT_FROM_TYPE_ERR
|
|
|
+#undef __GET_SUPPORTED_IMAGE_FORMATS_ERR
|
|
|
+#undef __CREATE_BUFFER_ERR
|
|
|
+#undef __COPY_ERR
|
|
|
+#undef __CREATE_SUBBUFFER_ERR
|
|
|
+#undef __CREATE_GL_BUFFER_ERR
|
|
|
+#undef __CREATE_GL_RENDER_BUFFER_ERR
|
|
|
+#undef __GET_GL_OBJECT_INFO_ERR
|
|
|
+#undef __CREATE_IMAGE_ERR
|
|
|
+#undef __CREATE_GL_TEXTURE_ERR
|
|
|
+#undef __IMAGE_DIMENSION_ERR
|
|
|
+#undef __SET_MEM_OBJECT_DESTRUCTOR_CALLBACK_ERR
|
|
|
+#undef __CREATE_USER_EVENT_ERR
|
|
|
+#undef __SET_USER_EVENT_STATUS_ERR
|
|
|
+#undef __SET_EVENT_CALLBACK_ERR
|
|
|
+#undef __WAIT_FOR_EVENTS_ERR
|
|
|
+#undef __CREATE_KERNEL_ERR
|
|
|
+#undef __SET_KERNEL_ARGS_ERR
|
|
|
+#undef __CREATE_PROGRAM_WITH_SOURCE_ERR
|
|
|
+#undef __CREATE_PROGRAM_WITH_IL_ERR
|
|
|
+#undef __CREATE_PROGRAM_WITH_BINARY_ERR
|
|
|
+#undef __CREATE_PROGRAM_WITH_IL_ERR
|
|
|
+#undef __CREATE_PROGRAM_WITH_BUILT_IN_KERNELS_ERR
|
|
|
+#undef __BUILD_PROGRAM_ERR
|
|
|
+#undef __COMPILE_PROGRAM_ERR
|
|
|
+#undef __LINK_PROGRAM_ERR
|
|
|
+#undef __CREATE_KERNELS_IN_PROGRAM_ERR
|
|
|
+#undef __CREATE_COMMAND_QUEUE_WITH_PROPERTIES_ERR
|
|
|
+#undef __CREATE_SAMPLER_WITH_PROPERTIES_ERR
|
|
|
+#undef __SET_COMMAND_QUEUE_PROPERTY_ERR
|
|
|
+#undef __ENQUEUE_READ_BUFFER_ERR
|
|
|
+#undef __ENQUEUE_READ_BUFFER_RECT_ERR
|
|
|
+#undef __ENQUEUE_WRITE_BUFFER_ERR
|
|
|
+#undef __ENQUEUE_WRITE_BUFFER_RECT_ERR
|
|
|
+#undef __ENQEUE_COPY_BUFFER_ERR
|
|
|
+#undef __ENQEUE_COPY_BUFFER_RECT_ERR
|
|
|
+#undef __ENQUEUE_FILL_BUFFER_ERR
|
|
|
+#undef __ENQUEUE_READ_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_WRITE_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_COPY_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_FILL_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_COPY_IMAGE_TO_BUFFER_ERR
|
|
|
+#undef __ENQUEUE_COPY_BUFFER_TO_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_MAP_BUFFER_ERR
|
|
|
+#undef __ENQUEUE_MAP_IMAGE_ERR
|
|
|
+#undef __ENQUEUE_UNMAP_MEM_OBJECT_ERR
|
|
|
+#undef __ENQUEUE_NDRANGE_KERNEL_ERR
|
|
|
+#undef __ENQUEUE_NATIVE_KERNEL
|
|
|
+#undef __ENQUEUE_MIGRATE_MEM_OBJECTS_ERR
|
|
|
+#undef __ENQUEUE_MIGRATE_SVM_ERR
|
|
|
+#undef __ENQUEUE_ACQUIRE_GL_ERR
|
|
|
+#undef __ENQUEUE_RELEASE_GL_ERR
|
|
|
+#undef __CREATE_PIPE_ERR
|
|
|
+#undef __GET_PIPE_INFO_ERR
|
|
|
+#undef __RETAIN_ERR
|
|
|
+#undef __RELEASE_ERR
|
|
|
+#undef __FLUSH_ERR
|
|
|
+#undef __FINISH_ERR
|
|
|
+#undef __VECTOR_CAPACITY_ERR
|
|
|
+#undef __CREATE_SUB_DEVICES_ERR
|
|
|
+#undef __CREATE_SUB_DEVICES_ERR
|
|
|
+#undef __ENQUEUE_MARKER_ERR
|
|
|
+#undef __ENQUEUE_WAIT_FOR_EVENTS_ERR
|
|
|
+#undef __ENQUEUE_BARRIER_ERR
|
|
|
+#undef __UNLOAD_COMPILER_ERR
|
|
|
+#undef __CREATE_GL_TEXTURE_2D_ERR
|
|
|
+#undef __CREATE_GL_TEXTURE_3D_ERR
|
|
|
+#undef __CREATE_IMAGE2D_ERR
|
|
|
+#undef __CREATE_IMAGE3D_ERR
|
|
|
+#undef __CREATE_COMMAND_QUEUE_ERR
|
|
|
+#undef __ENQUEUE_TASK_ERR
|
|
|
+#undef __CREATE_SAMPLER_ERR
|
|
|
+#undef __ENQUEUE_MARKER_WAIT_LIST_ERR
|
|
|
+#undef __ENQUEUE_BARRIER_WAIT_LIST_ERR
|
|
|
+#undef __CLONE_KERNEL_ERR
|
|
|
+#undef __GET_HOST_TIMER_ERR
|
|
|
+#undef __GET_DEVICE_AND_HOST_TIMER_ERR
|
|
|
+
|
|
|
+#endif //CL_HPP_USER_OVERRIDE_ERROR_STRINGS
|
|
|
+
|
|
|
+// Extensions
|
|
|
+#undef CL_HPP_INIT_CL_EXT_FCN_PTR_
|
|
|
+#undef CL_HPP_INIT_CL_EXT_FCN_PTR_PLATFORM_
|
|
|
+
|
|
|
+#if defined(CL_HPP_USE_CL_DEVICE_FISSION)
|
|
|
+#undef CL_HPP_PARAM_NAME_DEVICE_FISSION_
|
|
|
+#endif // CL_HPP_USE_CL_DEVICE_FISSION
|
|
|
+
|
|
|
+#undef CL_HPP_NOEXCEPT_
|
|
|
+#undef CL_HPP_DEFINE_STATIC_MEMBER_
|
|
|
+
|
|
|
+} // namespace cl
|
|
|
+
|
|
|
+#endif // CL_HPP_
|