// We need to use shifts here in order to mantain the sign bit for signed
// integers. The compiler should optimize this to (x & 0x00FFFFFF) for
// unsigned integers.
#define CONVERT_TO_24BIT(x) (((x) << 8) >> 8)
_CLC_OVERLOAD _CLC_DEF __CLC_GENTYPE mul24(__CLC_GENTYPE x, __CLC_GENTYPE y){
return CONVERT_TO_24BIT(x) * CONVERT_TO_24BIT(y);
}
#undef CONVERT_TO_24BIT