The SYCL* Compiler compiles C++-based SYCL source files with code for both CPU and a wide range of compute accelerators. The compiler uses Khronos* OpenCL™ API to offload computations to accelerators.
- Prerequisites
- Build SYCL toolchain
- Use SYCL toolchain
- C++ standard
- Known Issues and Limitations
- CUDA backend limitations
- Find More
git
- https://git-scm.com/downloadscmake
version 3.2 or later - http://www.cmake.org/download/python
- https://www.python.org/downloads/release/python-2716/- C++ compiler
- Linux:
GCC
version 5.1.0 or later (including libstdc++) - https://gcc.gnu.org/install/ - Windows:
Visual Studio
version 15.7 preview 4 or later - https://visualstudio.microsoft.com/downloads/
- Linux:
Throughout this document SYCL_HOME
denotes the path to the local directory
created as SYCL workspace. It might be useful to create an environment variable
with the same name.
Linux
export SYCL_HOME=/export/home/sycl_workspace
mkdir $SYCL_HOME
Windows (64-bit)
Open a developer command prompt using one of two methods:
- Click start menu and search for "x64 Native Tools Command Prompt for VS XXXX", where XXXX is a version of installed Visual Studio.
- Ctrl-R, write "cmd", click enter, then run
"C:\Program Files (x86)\Microsoft Visual Studio\2017\Community\VC\Auxiliary\Build\vcvarsall.bat" x64
set SYCL_HOME=%USERPROFILE%\sycl_workspace
mkdir %SYCL_HOME%
Linux
cd $SYCL_HOME
git clone https://github.com/codeplaysoftware/sycl-for-cuda -b sycl
mkdir $SYCL_HOME/build
cd $SYCL_HOME/build
cmake -DCMAKE_BUILD_TYPE=Release -DLLVM_TARGETS_TO_BUILD="X86" \
-DLLVM_EXTERNAL_PROJECTS="llvm-spirv;sycl" \
-DLLVM_ENABLE_PROJECTS="clang;llvm-spirv;sycl" \
-DLLVM_EXTERNAL_SYCL_SOURCE_DIR=$SYCL_HOME/llvm/sycl \
-DLLVM_EXTERNAL_LLVM_SPIRV_SOURCE_DIR=$SYCL_HOME/llvm/llvm-spirv \
$SYCL_HOME/llvm/llvm
make -j`nproc` sycl-toolchain
Windows (64-bit)
cd %SYCL_HOME%
git clone https://github.com/codeplaysoftware/sycl-for-cuda -b sycl
mkdir %SYCL_HOME%\build
cd %SYCL_HOME%\build
cmake -G "Ninja" -DCMAKE_BUILD_TYPE=Release -DLLVM_TARGETS_TO_BUILD="X86" ^
-DLLVM_EXTERNAL_PROJECTS="llvm-spirv;sycl" ^
-DLLVM_ENABLE_PROJECTS="clang;llvm-spirv;sycl" ^
-DLLVM_EXTERNAL_SYCL_SOURCE_DIR="%SYCL_HOME%\llvm\sycl" ^
-DLLVM_EXTERNAL_LLVM_SPIRV_SOURCE_DIR="%SYCL_HOME%\llvm\llvm-spirv" ^
-DCMAKE_C_COMPILER=cl -DCMAKE_CXX_COMPILER=cl ^
"%SYCL_HOME%\llvm\llvm"
ninja sycl-toolchain
To use ahead-of-time compilation for the Intel® processors, additionally build opencl-aot target:
- add
opencl-aot
to-DLLVM_EXTERNAL_PROJECTS
and-DLLVM_ENABLE_PROJECTS
variables above - add
opencl-aot
tomake
(for Linux) orninja
(for Windows) commands above
For more, see opencl-aot documentation.
TODO: add instructions how to deploy built SYCL toolchain.
There is experimental support for building and linking SYCL runtime with libc++ library instead of libstdc++. To enable it the following CMake options should be used.
Linux
-DSYCL_USE_LIBCXX=ON \
-DSYCL_LIBCXX_INCLUDE_PATH=<path to libc++ headers> \
-DSYCL_LIBCXX_LIBRARY_PATH=<path to libc++ and libc++abi libraries>
There is experimental support for SYCL for CUDA devices, currently in the cuda branch.
To enable support for CUDA devices, the following arguments need to be added to the CMake command when building the SYCL compiler.
-DCUDA_TOOLKIT_ROOT_DIR=/usr/local/cuda/ \
-DLLVM_ENABLE_PROJECTS="clang;llvm-spirv;sycl;libclc"\
-DSYCL_BUILD_PI_CUDA=ON\
-DLLVM_TARGETS_TO_BUILD="X86;NVPTX"\
-DLIBCLC_TARGETS_TO_BUILD="nvptx64--;nvptx64--nvidiacl"
Enabling this flag requires an installation of CUDA 10.1 on the system, refer to NVIDIA CUDA Installation Guide for Linux.
Currently, the only combination tested is Ubuntu 18.04 with CUDA 10.2 using a Titan RTX GPU (SM 71), but it should work on any GPU compatible with SM 50 or above.
To run SYCL applications on OpenCL devices, OpenCL implementation(s) must be present in the system.
Please, refer to the Release Notes for recommended Intel runtime versions.
The GPU
runtime that is needed to run SYCL application on Intel GPU
devices
can be downloaded from the following web pages:
-
Windows: Intel® Download Center
To install Intel CPU
runtime for OpenCL devices the corresponding runtime
asset/archive should be downloaded from
SYCL Compiler and Runtime updates and installed using
the following procedure.
Intel CPU
runtime for OpenCL depends on Threading Building Blocks library
which should be downloaded from Threading Building Blocks (TBB)
GitHub repository and installed following
procedure below.
Intel CPU
runtime for OpenCL devices can be switched into Intel FPGA
Emulation device for OpenCL. The following parameter should be set in cl.cfg
file (available in directory containing CPU runtime for OpenCL) to switch
OpenCL device mode:
CL_CONFIG_DEVICES = fpga-emu
Linux
- Extract the archive. For example, for the archive
oclcpu_rt_<cpu_version>.tar.gz
you would run the following commands
mkdir -p /opt/intel/oclcpuexp_<cpu_version>
cd /opt/intel/oclcpuexp_<cpu_version>
tar -zxvf oclcpu_rt_<cpu_version>.tar.gz
- Create ICD file pointing to the new runtime
echo /opt/intel/oclcpuexp_<cpu_version>/x64/libintelocl.so >
/etc/OpenCL/vendors/intel_expcpu.icd
- Extract TBB libraries. For example, for the archive tbb-<tbb_version>-lin.tgz
mkdir -p /opt/intel/tbb_<tbb_version>
cd /opt/intel/tbb_<tbb_version>
tar -zxvf tbb*lin.tgz
- Copy files from or create symbolic links to TBB libraries in OpenCL RT folder:
ln -s /opt/intel/tbb_<tbb_version>/tbb/lib/intel64/gcc4.8/libtbb.so
/opt/intel/oclcpuexp/x64/libtbb.so
ln -s /opt/intel/tbb_<tbb_version>/tbb/lib/intel64/gcc4.8/libtbbmalloc.so
/opt/intel/oclcpuexp/x64/libtbbmalloc.so
ln -s /opt/intel/tbb_<tbb_version>/tbb/lib/intel64/gcc4.8/libtbb.so.2
/opt/intel/oclcpuexp/x64/libtbb.so.2
ln -s /opt/intel/tbb_<tbb_version>/tbb/lib/intel64/gcc4.8/libtbbmalloc.so.2
/opt/intel/oclcpuexp/x64/libtbbmalloc.so.2
- Configure library paths
echo /opt/intel/oclcpuexp_<cpu_version>/x64 >
/etc/ld.so.conf.d/libintelopenclexp.conf
ldconfig -f /etc/ld.so.conf.d/libintelopenclexp.conf
Windows (64-bit)
-
If you need
GPU
as well, then update/install it first. Do it before installingCPU
runtime asGPU
runtime installer may re-write some important files or settings and make existingCPU
runtime not working properly. -
Extract the archive to some folder. For example, to
c:\oclcpu_rt_<cpu_version>
andc:\tbb_<tbb_version>
. -
Run
Command Prompt
asAdministrator
. To do that clickStart
button, typeCommand Prompt
, click the Right mouse button on it, then clickRun As Administrator
, then clickYes
to confirm. -
In the opened windows run
install.bat
provided with the extracted files to install runtime to the system and setup environment variables. So, if the extracted files are inc:\oclcpu_rt_<cpu_version>\
folder, then type the command:
c:\oclcpu_rt_<cpu_version>\install.bat c:\tbb_<tbb_version>\tbb\bin\intel64\vc14
To verify that built SYCL toolchain is working correctly, run:
Linux
make -j`nproc` check-all
Windows (64-bit)
ninja check-all
If no OpenCL GPU/CPU runtimes are available, the corresponding tests are skipped.
Khronos SYCL conformance test suite (CTS) is intended to validate SYCL implementation conformance to Khronos SYCL specification.
Follow Khronos SYCL-CTS instructions from README file to obtain test sources and instructions how build and execute the tests.
To configure testing of "Intel SYCL" toochain set
SYCL_IMPLEMENTATION=Intel_SYCL
and
Intel_SYCL_ROOT=<path to the SYCL installation>
CMake variables.
Linux
cmake -DIntel_SYCL_ROOT=$SYCL_HOME/deploy -DSYCL_IMPLEMENTATION=Intel_SYCL ...
Windows (64-bit)
cmake -DIntel_SYCL_ROOT=%SYCL_HOME%\deploy -DSYCL_IMPLEMENTATION=Intel_SYCL ...
Building Doxygen documentation is similar to building the product itself. First, the following tools need to be installed:
- doxygen
- graphviz
Then you'll need to add the following options to your CMake configuration command:
-DLLVM_ENABLE_DOXYGEN=ON
After CMake cache is generated, build the documentation with doxygen-sycl
target. It will be put to /path/to/build/tools/sycl/doc/html
directory.
A simple SYCL program consists of following parts:
- Header section
- Allocating buffer for data
- Creating SYCL queue
- Submitting command group to SYCL queue which includes the kernel
- Wait for the queue to complete the work
- Use buffer accessor to retrieve the result on the device and verify the data
- The end
Creating a file simple-sycl-app.cpp
with the following C++ SYCL code in it:
#include <CL/sycl.hpp>
int main() {
// Creating buffer of 4 ints to be used inside the kernel code
cl::sycl::buffer<cl::sycl::cl_int, 1> Buffer(4);
// Creating SYCL queue
cl::sycl::queue Queue;
// Size of index space for kernel
cl::sycl::range<1> NumOfWorkItems{Buffer.get_count()};
// Submitting command group(work) to queue
Queue.submit([&](cl::sycl::handler &cgh) {
// Getting write only access to the buffer on a device
auto Accessor = Buffer.get_access<cl::sycl::access::mode::write>(cgh);
// Executing kernel
cgh.parallel_for<class FillBuffer>(
NumOfWorkItems, [=](cl::sycl::id<1> WIid) {
// Fill buffer with indexes
Accessor[WIid] = (cl::sycl::cl_int)WIid.get(0);
});
});
// Getting read only access to the buffer on the host.
// Implicit barrier waiting for queue to complete the work.
const auto HostAccessor = Buffer.get_access<cl::sycl::access::mode::read>();
// Check the results
bool MismatchFound = false;
for (size_t I = 0; I < Buffer.get_count(); ++I) {
if (HostAccessor[I] != I) {
std::cout << "The result is incorrect for element: " << I
<< " , expected: " << I << " , got: " << HostAccessor[I]
<< std::endl;
MismatchFound = true;
}
}
if (!MismatchFound) {
std::cout << "The results are correct!" << std::endl;
}
return MismatchFound;
}
To build simple-sycl-app put bin
and lib
to PATHs:
Linux
export PATH=$SYCL_HOME/build/bin:$PATH
export LD_LIBRARY_PATH=$SYCL_HOME/build/lib:$LD_LIBRARY_PATH
Windows (64-bit)
set PATH=%SYCL_HOME%\build\bin;%PATH%
set LIB=%SYCL_HOME%\build\lib;%LIB%
and run following command:
clang++ -fsycl simple-sycl-app.cpp -o simple-sycl-app.exe
When building for CUDA, use the CUDA target triple as follows:
clang++ -fsycl -fsycl-targets=nvptx64-nvidia-cuda-sycldevice \
simple-sycl-app.cpp -o simple-sycl-app-cuda.exe
This simple-sycl-app.exe
application doesn't specify SYCL device for
execution, so SYCL runtime will use default_selector
logic to select one
of accelerators available in the system or SYCL host device.
Note: nvptx64-nvidia-cuda-sycldevice
is usable with -fsycl-targets
if clang was built with the cmake option SYCL_BUILD_PI_CUDA=ON
.
Linux & Windows
./simple-sycl-app.exe
The results are correct!
Note:
Currently, when the application has been built with the CUDA target, the CUDA backend
must be selected at runtime using the SYCL_BE
environment variable.
SYCL_BE=PI_CUDA ./simple-sycl-app-cuda.exe
NOTE: SYCL developer can specify SYCL device for execution using device
selectors (e.g. cl::sycl::cpu_selector
, cl::sycl::gpu_selector
,
Intel FPGA selector(s)) as
explained in following section Code the program for a specific
GPU.
To specify OpenCL device SYCL provides the abstract cl::sycl::device_selector
class which the can be used to define how the runtime should select the best
device.
The method cl::sycl::device_selector::operator()
of the SYCL
cl::sycl::device_selector
is an abstract member function which takes a
reference to a SYCL device and returns an integer score. This abstract member
function can be implemented in a derived class to provide a logic for selecting
a SYCL device. SYCL runtime uses the device for with the highest score is
returned. Such object can be passed to cl::sycl::queue
and cl::sycl::device
constructors.
The example below illustrates how to use cl::sycl::device_selector
to create
device and queue objects bound to Intel GPU device:
#include <CL/sycl.hpp>
int main() {
class NEOGPUDeviceSelector : public cl::sycl::device_selector {
public:
int operator()(const cl::sycl::device &Device) const override {
using namespace cl::sycl::info;
const std::string DeviceName = Device.get_info<device::name>();
const std::string DeviceVendor = Device.get_info<device::vendor>();
return Device.is_gpu() && (DeviceName.find("HD Graphics NEO") != std::string::npos);
}
};
NEOGPUDeviceSelector Selector;
try {
cl::sycl::queue Queue(Selector);
cl::sycl::device Device(Selector);
} catch (cl::sycl::invalid_parameter_error &E) {
std::cout << E.what() << std::endl;
}
}
The device selector below selects an NVIDIA device only, and won't execute if there is none.
class CUDASelector : public cl::sycl::device_selector {
public:
int operator()(const cl::sycl::device &Device) const override {
using namespace cl::sycl::info;
const std::string DriverVersion = Device.get_info<device::driver_version>();
if (Device.is_gpu() && (DriverVersion.find("CUDA") != std::string::npos)) {
return 1;
};
return -1;
}
};
- Minimally support C++ standard is c++11 on Linux and c++14 on Windows.
- SYCL device compiler fails if the same kernel was used in different translation units.
- SYCL host device is not fully supported.
- 32-bit host/target is not supported.
- SYCL works only with OpenCL implementations supporting out-of-order queues.
- On Windows linking SYCL applications with
/MTd
flag is known to cause crashes.
- Backend is only supported on Linux
- The only combination tested is Ubuntu 18.04 with CUDA 10.2 using a Titan RTX GPU (SM 71), but it should work on any GPU compatible with SM 50 or above
- The NVIDIA OpenCL headers conflict with the OpenCL headers required for this project and may cause compilation issues on some platforms
SYCL 1.2.1 specification: www.khronos.org/registry/SYCL/specs/sycl-1.2.1.pdf