Nicholas Wilson is a student at Murdoch University, studying for his BEng (Hons)/BSc in Industrial Computer Systems (Hons) and Instrumentation & Control/ Molecular Biology & Genetics and Biomedical Science. He just finished his thesis on low-cost defect detection of solar cells by electroluminescence imaging, which gives him time to work on DCompute and write about it for the D Blog. He plays the piano, ice skates, and has spent 7 years putting D to use on number bashing, automation, and anything else that he could make a computer do for him.
DCompute is a framework and compiler extension to support writing native kernels for OpenCL and CUDA in D to utilise GPUs and other accelerators for computationally intensive code. In development are drivers to automate the interactions between user code and the tedious and error prone compute APIs with the goal of enabling the rapid development of high performance D libraries and applications.
Introduction
After watching John Colvin’s DConf 2016 presentation in May of last year on using D’s metaprogramming to make the OpenCL API marginally less horrible to use, I thought, “This would be so much easier to do if we were able to write kernels in D, rather than doing string manipulations in OpenCL C”. At the time, I was coming up to the end of a rather busy semester and thought that would make a good winter[1] project. After all, LDC, the LLVM D Compiler, has access to LLVM’s SPIR-V and PTX backends, and I thought, “It can’t be too hard, its only glue code”. I slightly underestimated the time it would take, finishing the first stage of DCompute (because naming things is hard), mainlining the changes I made to LDC at the end of February, eight months later — just in time for the close of submissions to DConf, where I gave a talk on the progress I had made.
Apart from familiarising myself with the LDC and DMD front-end codebases, I also had to understand the LLVM SPIR-V and PTX backends that I was trying to target, because they require the use of special metadata (for e.g. denoting a function is a kernel) and address spaces, used to represent __global
& friends in OpenCL C and __global__
& friends in CUDA, and introduce these concepts into LDC.
But once I was familiar with the code and had sorted the above discrepancies, it was mostly smooth sailing translating the OpenCL and CUDA modifiers into compiler-recognised attributes and wrapping the intrinsics into an easy to use and consistent interface.
When it was all working and almost ready to merge into mainline LDC, I hit a bit of a snag with regards to CI: the SPIR-V backend that was being developed by Khronos was based on the quite old LLVM 3.6.1 and, despite my suggestions, did not have any releases. So I forward ported the backend and the conversion utility to the master branch of LLVM and made a release myself. Still in progress on this front are converting magic intrinsics to proper LLVM intrinsics and transitioning to a TableGen-driven approach for the backend in preparation for merging the backend into LLVM Trunk. This should hopefully be done soon™.
Current state of DCompute
With the current state of DCompute we are able to write kernels natively in D and have access to most of its language-defining features like templates & static introspection, UFCS, scope guards, ranges & algorithms and CTFE. Notably missing, for hardware and performance reasons, are those features commonly excluded in kernel languages, like function pointers, virtual functions, dynamic recursion, RTTI, exceptions and the use of the garbage collector. Note that unlike OpenCL C++ we allow kernel functions to be templated and have overloads and default values. Still in development is support for images and pipes.
Example code
To write kernels in D, we need to pass -mdcompute-targets=<targets>
to LDC, where <targets>
is a comma-separated list of the desired targets to build for, e.g. ocl-120,cuda-350
for OpenCL 1.2 and CUDA compute capability 3.5, respectively (yes, we can do them all at once!). We get one file for each target, e.g. kernels_ocl120_64.spv
, when built in 64-bit mode, which contains all of the code for that device.
The vector add
kernel in D is:
@compute(CompileFor.deviceOnly) module example; import ldc.dcompute; import dcompute.std.index; alias gf = GlobalPointer!float; @kernel void vadd(gf a, gf b, gf c) { auto x = GlobalIndex.x; a[x] = b[x]+c[x]; }
Modules marked with the @compute
attribute are compiled for each of the command line targets, @kernel
makes a function a kernel, and GlobalPointer
is the equivalent of the __global
qualifier in OpenCL.
Kernels are not restricted to just functions — lambdas & tamplates also work:
@kernel void map(alias F)(KernelArgs!F args) { F(args); } //In host code AutoBuffer!float x,y,z; // y & z initialised with data q.enqueue!(map!((a,b,c) => a=b+c))(x.length)(x, y, z);
Where KernelArgs
translates host types to device types (e.g. buffers to pointers or, as in this example, AutoBuffers to AutoIndexed Pointers) so that we encapsulate the differences in the host and device types.
The last line is the expected syntax for launching kernels, q.enqueue!kernel(dimensions)(args)
, akin to CUDA’s kernel<<<dimensions,queue>>>(args)
. The libraries for launching kernels are in development.
Unlike CUDA, where all the magic for transforming the above expression into code on the host lies in the compiler, q.enqueue!func(sizes)(args)
will be processed by static introspection of the driver library of DCompute.
The sole reason we can do this in D is that we are able to query the mangled name the compiler will give to a symbol via the symbol’s .mangleof
property. This, in combination with D’s easy to use and powerful templates, means we can significantly reduce the mental overhead associated with using the compute APIs. Also, implementing this in the library will be much simpler, and therefore faster to implement, than putting the same behaviour in the compiler. While this may not seem much for CUDA users, this will be a breath of fresh air to OpenCL users (just look at the OpenCL vector add host code example steps 7-11).
While you cant do that just yet in DCompute, development should start to progress quickly and hopefully become a reality soon.
I would like to thank John Colvin for the initial inspiration, Mike Parker for editing, and the LDC folks, David Nadlinger, Kai Nacke, Martin Kinke, with a special thanks to Johan Engelen, for their help with understanding the LDC codebase and reviewing my work.
If you would like to help develop DCompute (or be kept in the loop), feel free to drop a line at the libmir Gitter. Similarly, any efforts preparing the SPIR-V backend for inclusion into LLVM are also greatly appreciated.
[1] Southern hemisphere.
Frankly, the @kernel attribute and the associated compiler switches should be part of the D standard, and every compiler implementation should provide support for it.
That’d be nice, but the current implementations rely on LLVM’s backends.
GCC does not have a SPIR-V backend and as far as I know there’s no OpenCL or CUDA backend either, so GDC can’t support any GPGPU stuff. The situation is similar for DMD.
So for now, using LDC specific extensions shouldn’t really be a problem.