/* * Copyright (C) 2010-2018 Arm Limited or its affiliates. All rights reserved. * * SPDX-License-Identifier: Apache-2.0 * * Licensed under the Apache License, Version 2.0 (the License); you may * not use this file except in compliance with the License. * You may obtain a copy of the License at * * www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an AS IS BASIS, WITHOUT * WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ /* ---------------------------------------------------------------------- * Project: CMSIS NN Library * Title: arm_fully_connected_mat_q7_vec_q15_opt.c * Description: Mixed Q15-Q7 opt fully-connected layer function * * $Date: 17. January 2018 * $Revision: V.1.0.0 * * Target Processor: Cortex-M cores * * -------------------------------------------------------------------- */ #include "arm_math.h" #include "arm_nnfunctions.h" /** * @ingroup groupNN */ /** * @addtogroup FC * @{ */ /** * @brief Mixed Q15-Q7 opt fully-connected layer function * @param[in] pV pointer to input vector * @param[in] pM pointer to matrix weights * @param[in] dim_vec length of the vector * @param[in] num_of_rows number of rows in weight matrix * @param[in] bias_shift amount of left-shift for bias * @param[in] out_shift amount of right-shift for output * @param[in] bias pointer to bias * @param[in,out] pOut pointer to output vector * @param[in,out] vec_buffer pointer to buffer space for input * @return The function returns ARM_MATH_SUCCESS * * @details * * Buffer size: * * vec_buffer size: 0 * * Q7_Q15 version of the fully connected layer * * Weights are in q7_t and Activations are in q15_t * * Limitation: x4 version requires weight reordering to work * * Here we use only one pointer to read 4 rows in the weight * matrix. So if the original q7_t matrix looks like this: * * | a11 | a12 | a13 | a14 | a15 | a16 | a17 | * * | a21 | a22 | a23 | a24 | a25 | a26 | a27 | * * | a31 | a32 | a33 | a34 | a35 | a36 | a37 | * * | a41 | a42 | a43 | a44 | a45 | a46 | a47 | * * | a51 | a52 | a53 | a54 | a55 | a56 | a57 | * * | a61 | a62 | a63 | a64 | a65 | a66 | a67 | * * We operates on multiple-of-4 rows, so the first four rows becomes * * | a11 | a21 | a12 | a22 | a31 | a41 | a32 | a42 | * * | a13 | a23 | a14 | a24 | a33 | a43 | a34 | a44 | * * | a15 | a25 | a16 | a26 | a35 | a45 | a36 | a46 | * * The column left over will be in-order. * which is: * | a17 | a27 | a37 | a47 | * * For the left-over rows, we do 1x1 computation, so the data remains * as its original order. * * So the stored weight matrix looks like this: * * | a11 | a21 | a12 | a22 | a31 | a41 | * * | a32 | a42 | a13 | a23 | a14 | a24 | * * | a33 | a43 | a34 | a44 | a15 | a25 | * * | a16 | a26 | a35 | a45 | a36 | a46 | * * | a17 | a27 | a37 | a47 | a51 | a52 | * * | a53 | a54 | a55 | a56 | a57 | a61 | * * | a62 | a63 | a64 | a65 | a66 | a67 | * */ arm_status arm_fully_connected_mat_q7_vec_q15_opt(const q15_t * pV, const q7_t * pM, const uint16_t dim_vec, const uint16_t num_of_rows, const uint16_t bias_shift, const uint16_t out_shift, const q7_t * bias, q15_t * pOut, q15_t * vec_buffer) { #if defined (ARM_MATH_DSP) /* Run the following code for Cortex-M4 and Cortex-M7 */ const q7_t *pB = pM; q15_t *pO = pOut; const q7_t *pBias = bias; const q15_t *pA = pV; uint16_t rowCnt = num_of_rows >> 2; while (rowCnt) { q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum2 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum3 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum4 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); uint16_t colCnt = dim_vec >> 1; pA = pV; #ifdef USE_INTRINSIC #ifndef ARM_MATH_BIG_ENDIAN while (colCnt) { q31_t inM11, inM12, inM13, inM14; q31_t inV; inV = *__SIMD32(pA)++; inM11 = *__SIMD32(pB)++; inM12 = __SXTB16(__ROR(inM11, 8)); inM11 = __SXTB16(inM11); sum = __SMLAD(inM11, inV, sum); sum2 = __SMLAD(inM12, inV, sum2); inM13 = *__SIMD32(pB)++; inM14 = __SXTB16(__ROR(inM13, 8)); inM13 = __SXTB16(inM13); sum3 = __SMLAD(inM13, inV, sum3); sum4 = __SMLAD(inM14, inV, sum4); colCnt--; } #else while (colCnt) { q31_t inM11, inM12, inM13, inM14; q31_t inV; inV = *__SIMD32(pA)++; inM11 = *__SIMD32(pB)++; inM12 = __SXTB16(__ROR(inM11, 8)); inM11 = __SXTB16(inM11); sum = __SMLAD(inM12, inV, sum); sum2 = __SMLAD(inM11, inV, sum2); inM13 = *__SIMD32(pB)++; inM14 = __SXTB16(__ROR(inM13, 8)); inM13 = __SXTB16(inM13); sum3 = __SMLAD(inM14, inV, sum3); sum4 = __SMLAD(inM13, inV, sum4); colCnt--; } #endif /* ARM_MATH_BIG_ENDIAN */ #else /* * register needed: * loop counter: colCnt * accumulators: sum, sum2, sum3, sum4 * pointers: pB, pA * weight data: inM11, inM12, inM13, inM14 * activation data: inV */ #ifndef ARM_MATH_BIG_ENDIAN asm volatile ("COL_LOOP_%=:\n" "ldr.w r4, [%[pA]], #4\n" "ldr.w r1, [%[pB]], #8\n" "mov.w r0, r1, ror #8\n" "sxtb16 r0, r0\n" "sxtb16 r1, r1\n" "smlad %[sum], r4, r1, %[sum]\n" "smlad %[sum2], r4, r0, %[sum2]\n" "ldr.w r3, [%[pB], #-4]\n" "mov.w r2, r3, ror #8\n" "sxtb16 r2, r2\n" "sxtb16 r3, r3\n" "smlad %[sum3], r4, r3, %[sum3]\n" "smlad %[sum4], r4, r2, %[sum4]\n" "subs %[colCnt], #1\n" "bne COL_LOOP_%=\n":[sum] "+r"(sum), [sum2] "+r"(sum2),[sum3] "+r"(sum3), [sum4] "+r"(sum4),[pB] "+r"(pB),[pA] "+r"(pA):[colCnt] "r"(colCnt):"r0", "r1", "r2", "r3", "r4"); #else asm volatile ("COL_LOOP_%=:\n" "ldr.w r4, [%[pA]], #4\n" "ldr.w r1, [%[pB]], #8\n" "mov.w r0, r1, ror #8\n" "sxtb16 r0, r0\n" "sxtb16 r1, r1\n" "smlad %[sum], r4, r0, %[sum]\n" "smlad %[sum2], r4, r1, %[sum2]\n" "ldr.w r3, [%[pB], #-4]\n" "mov.w r2, r3, ror #8\n" "sxtb16 r2, r2\n" "sxtb16 r3, r3\n" "smlad %[sum3], r4, r2, %[sum3]\n" "smlad %[sum4], r4, r3, %[sum4]\n" "subs %[colCnt], #1\n" "bne COL_LOOP_%=\n":[sum] "+r"(sum), [sum2] "+r"(sum2),[sum3] "+r"(sum3), [sum4] "+r"(sum4),[pB] "+r"(pB),[pA] "+r"(pA):[colCnt] "r"(colCnt):"r0", "r1", "r2", "r3", "r4"); #endif /* ARM_MATH_BIG_ENDIAN */ #endif /* USE_INTRINSIC */ colCnt = dim_vec & 0x1; while (colCnt) { q15_t inV = *pA++; q7_t inM = *pB++; q7_t inM2 = *pB++; q7_t inM3 = *pB++; q7_t inM4 = *pB++; sum += inV * inM; sum2 += inV * inM2; sum3 += inV * inM3; sum4 += inV * inM4; colCnt--; } /* while over colCnt */ *pO++ = (q15_t) (__SSAT((sum >> out_shift), 16)); *pO++ = (q15_t) (__SSAT((sum2 >> out_shift), 16)); *pO++ = (q15_t) (__SSAT((sum3 >> out_shift), 16)); *pO++ = (q15_t) (__SSAT((sum4 >> out_shift), 16)); /* adjust the pointers and counters */ rowCnt--; } /* left-over part of the rows */ rowCnt = num_of_rows & 0x3; while (rowCnt) { q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); uint16_t colCnt = dim_vec >> 2; pA = pV; while (colCnt) { q31_t inV1, inV2, inM11, inM12; pB = (q7_t *) read_and_pad((void *)pB, &inM11, &inM12); inV1 = *__SIMD32(pA)++; sum = __SMLAD(inV1, inM11, sum); inV2 = *__SIMD32(pA)++; sum = __SMLAD(inV2, inM12, sum); colCnt--; } /* left-over of the vector */ colCnt = dim_vec & 0x3; while (colCnt) { q15_t inV = *pA++; q7_t inM = *pB++; sum += inV * inM; colCnt--; } *pO++ = (q15_t) (__SSAT((sum >> out_shift), 16)); rowCnt--; } #else /* Run the following code as reference implementation for Cortex-M0 and Cortex-M3 */ uint16_t rowCnt = num_of_rows >> 2; const q7_t *pB = pM; const q15_t *pA; q15_t *pO = pOut; const q7_t *pBias = bias; while (rowCnt) { q31_t sum = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum2 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum3 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); q31_t sum4 = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); uint16_t colCnt = dim_vec >> 1; pA = pV; while (colCnt) { q15_t inA1 = *pA++; q15_t inA2 = *pA++; q7_t inB1 = *pB++; q7_t inB3 = *pB++; q7_t inB2 = *pB++; q7_t inB4 = *pB++; sum += inA1 * inB1 + inA2 * inB2; sum2 += inA1 * inB3 + inA2 * inB4; inB1 = *pB++; inB3 = *pB++; inB2 = *pB++; inB4 = *pB++; sum3 += inA1 * inB1 + inA2 * inB2; sum4 += inA1 * inB3 + inA2 * inB4; colCnt--; } colCnt = dim_vec & 0x1; while (colCnt) { q15_t inA = *pA++; q7_t inB = *pB++; sum += inA * inB; inB = *pB++; sum2 += inA * inB; inB = *pB++; sum3 += inA * inB; inB = *pB++; sum4 += inA * inB; colCnt--; } *pO++ = (q15_t) __SSAT((sum >> out_shift), 16); *pO++ = (q15_t) __SSAT((sum2 >> out_shift), 16); *pO++ = (q15_t) __SSAT((sum3 >> out_shift), 16); *pO++ = (q15_t) __SSAT((sum4 >> out_shift), 16); rowCnt--; } rowCnt = num_of_rows & 0x3; while (rowCnt) { int ip_out = ((q31_t)(*pBias++) << bias_shift) + NN_ROUND(out_shift); int j; pA = pV; for (j = 0; j < dim_vec; j++) { q15_t inA = *pA++; q7_t inB = *pB++; ip_out += inA * inB; } *pO++ = (q15_t) __SSAT((ip_out >> out_shift), 16); rowCnt--; } #endif /* ARM_MATH_DSP */ /* Return to ARM_MATH_SUCCESS */ return (ARM_MATH_SUCCESS); } /** * @} end of FC group */