We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
2 parents 10499af + 2a70264 commit f596d43Copy full SHA for f596d43
kernel_tuner/backends/pycuda.py
@@ -180,6 +180,9 @@ def ready_argument_list(self, arguments):
180
# pycuda does not support bool, convert to uint8 instead
181
elif isinstance(arg, np.bool_):
182
gpu_args.append(arg.astype(np.uint8))
183
+ # pycuda does not support 16-bit formats, view them as uint16
184
+ elif isinstance(arg, np.generic) and str(arg.dtype) in ("float16", "bfloat16"):
185
+ gpu_args.append(arg.view(np.uint16))
186
# if not an array, just pass argument along
187
else:
188
gpu_args.append(arg)
0 commit comments