// Copyright 2021 Google LLC // // This source code is licensed under the BSD-style license found in the // LICENSE file in the root directory of this source tree. $assert BATCH_TILE >= 1 #include #include #include #include #include void xnn_f32_f16_vcvt_ukernel__scalar_fabsf_x${BATCH_TILE}( size_t batch, const float* input, void* output, const union xnn_f32_f16_cvt_params params[restrict XNN_MIN_ELEMENTS(1)]) { assert(batch != 0); assert(batch % sizeof(float) == 0); assert(input != NULL); assert(output != NULL); const float vscale_to_inf = params->scalar_fabsf.scale_to_inf; const uint32_t vexp_bias = params->scalar_fabsf.exp_bias; const float vscale_to_zero = params->scalar_fabsf.scale_to_zero; const uint32_t vexpw_max = params->scalar_fabsf.expw_max; const uint32_t vbias_min = params->scalar_fabsf.bias_min; const uint16_t vexph_mask = params->scalar_fabsf.exph_mask; const uint16_t vmanth_mask = params->scalar_fabsf.manth_mask; const uint16_t vnanh = params->scalar_fabsf.nanh; uint16_t* o = (uint16_t*) output; $if BATCH_TILE == 1: do { const float vx = *input++; const float vabsx = fabsf(vx); uint32_t vsignw = float_as_uint32(vx); const uint32_t vnonsignw = float_as_uint32(vabsx); float vf = vabsx * vscale_to_inf; uint32_t vbias = vnonsignw + vexp_bias; vsignw ^= vnonsignw; vf *= vscale_to_zero; vbias &= vexpw_max; vbias = math_max_u32(vbias, vbias_min); vf += uint32_as_float(vbias); const uint32_t vbits = float_as_uint32(vf); const uint16_t vexph = (uint16_t) (vbits >> 13) & vexph_mask; const uint16_t vmanth = (uint16_t) vbits & vmanth_mask; const uint16_t vsignh = (uint16_t) (vsignw >> 16); uint16_t vh = vexph + vmanth; if XNN_UNPREDICTABLE(vnonsignw > vexpw_max) { vh = vnanh; } vh |= vsignh; *o++ = vh; batch -= sizeof(float); } while (batch != 0); $else: for (; batch >= ${BATCH_TILE} * sizeof(float); batch -= ${BATCH_TILE} * sizeof(float)) { $for N in range(BATCH_TILE): const float vx${N} = input[${N}]; input += ${BATCH_TILE}; $for N in range(BATCH_TILE): const float vabsx${N} = fabsf(vx${N}); $for N in range(BATCH_TILE): uint32_t vsignw${N} = float_as_uint32(vx${N}); $for N in range(BATCH_TILE): const uint32_t vnonsignw${N} = float_as_uint32(vabsx${N}); $for N in range(BATCH_TILE): float vf${N} = vabsx${N} * vscale_to_inf; $for N in range(BATCH_TILE): uint32_t vbias${N} = vnonsignw${N} + vexp_bias; $for N in range(BATCH_TILE): vsignw${N} ^= vnonsignw${N}; $for N in range(BATCH_TILE): vf${N} *= vscale_to_zero; $for N in range(BATCH_TILE): vbias${N} &= vexpw_max; $for N in range(BATCH_TILE): vbias${N} = math_max_u32(vbias${N}, vbias_min); $for N in range(BATCH_TILE): vf${N} += uint32_as_float(vbias${N}); $for N in range(BATCH_TILE): const uint32_t vbits${N} = float_as_uint32(vf${N}); $for N in range(BATCH_TILE): const uint16_t vexph${N} = (uint16_t) (vbits${N} >> 13) & vexph_mask; $for N in range(BATCH_TILE): const uint16_t vmanth${N} = (uint16_t) vbits${N} & vmanth_mask; $for N in range(BATCH_TILE): const uint16_t vsignh${N} = (uint16_t) (vsignw${N} >> 16); $for N in range(BATCH_TILE): uint16_t vh${N} = vexph${N} + vmanth${N}; $for N in range(BATCH_TILE): if XNN_UNPREDICTABLE(vnonsignw${N} > vexpw_max) { vh${N} = vnanh; } $for N in range(BATCH_TILE): vh${N} |= vsignh${N}; $for N in range(BATCH_TILE): o[${N}] = vh${N}; o += ${BATCH_TILE}; } $if BATCH_TILE == 2: if XNN_UNLIKELY(batch != 0) { const float vx = *input; const float vabsx = fabsf(vx); uint32_t vsignw = float_as_uint32(vx); const uint32_t vnonsignw = float_as_uint32(vabsx); float vf = vabsx * vscale_to_inf; uint32_t vbias = vnonsignw + vexp_bias; vsignw ^= vnonsignw; vf *= vscale_to_zero; vbias &= vexpw_max; vbias = math_max_u32(vbias, vbias_min); vf += uint32_as_float(vbias); const uint32_t vbits = float_as_uint32(vf); const uint16_t vexph = (uint16_t) (vbits >> 13) & vexph_mask; const uint16_t vmanth = (uint16_t) vbits & vmanth_mask; const uint16_t vsignh = (uint16_t) (vsignw >> 16); uint16_t vh = vexph + vmanth; if XNN_UNPREDICTABLE(vnonsignw > vexpw_max) { vh = vnanh; } vh |= vsignh; *o = vh; } $else: if XNN_UNLIKELY(batch != 0) { do { const float vx = *input++; const float vabsx = fabsf(vx); uint32_t vsignw = float_as_uint32(vx); const uint32_t vnonsignw = float_as_uint32(vabsx); float vf = vabsx * vscale_to_inf; uint32_t vbias = vnonsignw + vexp_bias; vsignw ^= vnonsignw; vf *= vscale_to_zero; vbias &= vexpw_max; vbias = math_max_u32(vbias, vbias_min); vf += uint32_as_float(vbias); const uint32_t vbits = float_as_uint32(vf); const uint16_t vexph = (uint16_t) (vbits >> 13) & vexph_mask; const uint16_t vmanth = (uint16_t) vbits & vmanth_mask; const uint16_t vsignh = (uint16_t) (vsignw >> 16); uint16_t vh = vexph + vmanth; if XNN_UNPREDICTABLE(vnonsignw > vexpw_max) { vh = vnanh; } vh |= vsignh; *o++ = vh; batch -= sizeof(float); } while (batch != 0); } }