You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We can specialize the cuda backend case in order to call appropriate instructions for f16 precision operators instead, but I opened this issue to raise awareness for other backends/ to get potentially a general solution that lowers to backend specific instructions.
The text was updated successfully, but these errors were encountered:
@JackAKirk, the code you are pointing to is host implementation for half. It does indeed makes a fallback to float math, because we have no half in standard C++.
However, on device side we should be using _Float16 for math operations and that has been the case for more than 4 years now, see #1089
sycl::half
operators likellvm/sycl/include/sycl/half_type.hpp
Line 155 in b6faefb
We can specialize the cuda backend case in order to call appropriate instructions for f16 precision operators instead, but I opened this issue to raise awareness for other backends/ to get potentially a general solution that lowers to backend specific instructions.
The text was updated successfully, but these errors were encountered: