pigweed / third_party / github / STMicroelectronics / cmsis_core / 7dd288b23bf605a3a2fafa81a29d2c96a2fd83ce / . / DSP_Lib / Source / FilteringFunctions / arm_correlate_q15.c

/* ---------------------------------------------------------------------- | |

* Copyright (C) 2010-2014 ARM Limited. All rights reserved. | |

* | |

* $Date: 19. March 2015 | |

* $Revision: V.1.4.5 | |

* | |

* Project: CMSIS DSP Library | |

* Title: arm_correlate_q15.c | |

* | |

* Description: Correlation of Q15 sequences. | |

* | |

* Target Processor: Cortex-M4/Cortex-M3/Cortex-M0 | |

* | |

* Redistribution and use in source and binary forms, with or without | |

* modification, are permitted provided that the following conditions | |

* are met: | |

* - Redistributions of source code must retain the above copyright | |

* notice, this list of conditions and the following disclaimer. | |

* - Redistributions in binary form must reproduce the above copyright | |

* notice, this list of conditions and the following disclaimer in | |

* the documentation and/or other materials provided with the | |

* distribution. | |

* - Neither the name of ARM LIMITED nor the names of its contributors | |

* may be used to endorse or promote products derived from this | |

* software without specific prior written permission. | |

* | |

* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS | |

* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT | |

* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS | |

* FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE | |

* COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, | |

* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, | |

* BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; | |

* LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER | |

* CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT | |

* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN | |

* ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | |

* POSSIBILITY OF SUCH DAMAGE. | |

* -------------------------------------------------------------------- */ | |

#include "arm_math.h" | |

/** | |

* @ingroup groupFilters | |

*/ | |

/** | |

* @addtogroup Corr | |

* @{ | |

*/ | |

/** | |

* @brief Correlation of Q15 sequences. | |

* @param[in] *pSrcA points to the first input sequence. | |

* @param[in] srcALen length of the first input sequence. | |

* @param[in] *pSrcB points to the second input sequence. | |

* @param[in] srcBLen length of the second input sequence. | |

* @param[out] *pDst points to the location where the output result is written. Length 2 * max(srcALen, srcBLen) - 1. | |

* @return none. | |

* | |

* @details | |

* <b>Scaling and Overflow Behavior:</b> | |

* | |

* \par | |

* The function is implemented using a 64-bit internal accumulator. | |

* Both inputs are in 1.15 format and multiplications yield a 2.30 result. | |

* The 2.30 intermediate results are accumulated in a 64-bit accumulator in 34.30 format. | |

* This approach provides 33 guard bits and there is no risk of overflow. | |

* The 34.30 result is then truncated to 34.15 format by discarding the low 15 bits and then saturated to 1.15 format. | |

* | |

* \par | |

* Refer to <code>arm_correlate_fast_q15()</code> for a faster but less precise version of this function for Cortex-M3 and Cortex-M4. | |

* | |

* \par | |

* Refer the function <code>arm_correlate_opt_q15()</code> for a faster implementation of this function using scratch buffers. | |

* | |

*/ | |

void arm_correlate_q15( | |

q15_t * pSrcA, | |

uint32_t srcALen, | |

q15_t * pSrcB, | |

uint32_t srcBLen, | |

q15_t * pDst) | |

{ | |

#if (defined(ARM_MATH_CM4) || defined(ARM_MATH_CM3)) && !defined(UNALIGNED_SUPPORT_DISABLE) | |

/* Run the below code for Cortex-M4 and Cortex-M3 */ | |

q15_t *pIn1; /* inputA pointer */ | |

q15_t *pIn2; /* inputB pointer */ | |

q15_t *pOut = pDst; /* output pointer */ | |

q63_t sum, acc0, acc1, acc2, acc3; /* Accumulators */ | |

q15_t *px; /* Intermediate inputA pointer */ | |

q15_t *py; /* Intermediate inputB pointer */ | |

q15_t *pSrc1; /* Intermediate pointers */ | |

q31_t x0, x1, x2, x3, c0; /* temporary variables for holding input and coefficient values */ | |

uint32_t j, k = 0u, count, blkCnt, outBlockSize, blockSize1, blockSize2, blockSize3; /* loop counter */ | |

int32_t inc = 1; /* Destination address modifier */ | |

/* The algorithm implementation is based on the lengths of the inputs. */ | |

/* srcB is always made to slide across srcA. */ | |

/* So srcBLen is always considered as shorter or equal to srcALen */ | |

/* But CORR(x, y) is reverse of CORR(y, x) */ | |

/* So, when srcBLen > srcALen, output pointer is made to point to the end of the output buffer */ | |

/* and the destination pointer modifier, inc is set to -1 */ | |

/* If srcALen > srcBLen, zero pad has to be done to srcB to make the two inputs of same length */ | |

/* But to improve the performance, | |

* we include zeroes in the output instead of zero padding either of the the inputs*/ | |

/* If srcALen > srcBLen, | |

* (srcALen - srcBLen) zeroes has to included in the starting of the output buffer */ | |

/* If srcALen < srcBLen, | |

* (srcALen - srcBLen) zeroes has to included in the ending of the output buffer */ | |

if(srcALen >= srcBLen) | |

{ | |

/* Initialization of inputA pointer */ | |

pIn1 = (pSrcA); | |

/* Initialization of inputB pointer */ | |

pIn2 = (pSrcB); | |

/* Number of output samples is calculated */ | |

outBlockSize = (2u * srcALen) - 1u; | |

/* When srcALen > srcBLen, zero padding is done to srcB | |

* to make their lengths equal. | |

* Instead, (outBlockSize - (srcALen + srcBLen - 1)) | |

* number of output samples are made zero */ | |

j = outBlockSize - (srcALen + (srcBLen - 1u)); | |

/* Updating the pointer position to non zero value */ | |

pOut += j; | |

} | |

else | |

{ | |

/* Initialization of inputA pointer */ | |

pIn1 = (pSrcB); | |

/* Initialization of inputB pointer */ | |

pIn2 = (pSrcA); | |

/* srcBLen is always considered as shorter or equal to srcALen */ | |

j = srcBLen; | |

srcBLen = srcALen; | |

srcALen = j; | |

/* CORR(x, y) = Reverse order(CORR(y, x)) */ | |

/* Hence set the destination pointer to point to the last output sample */ | |

pOut = pDst + ((srcALen + srcBLen) - 2u); | |

/* Destination address modifier is set to -1 */ | |

inc = -1; | |

} | |

/* The function is internally | |

* divided into three parts according to the number of multiplications that has to be | |

* taken place between inputA samples and inputB samples. In the first part of the | |

* algorithm, the multiplications increase by one for every iteration. | |

* In the second part of the algorithm, srcBLen number of multiplications are done. | |

* In the third part of the algorithm, the multiplications decrease by one | |

* for every iteration.*/ | |

/* The algorithm is implemented in three stages. | |

* The loop counters of each stage is initiated here. */ | |

blockSize1 = srcBLen - 1u; | |

blockSize2 = srcALen - (srcBLen - 1u); | |

blockSize3 = blockSize1; | |

/* -------------------------- | |

* Initializations of stage1 | |

* -------------------------*/ | |

/* sum = x[0] * y[srcBlen - 1] | |

* sum = x[0] * y[srcBlen - 2] + x[1] * y[srcBlen - 1] | |

* .... | |

* sum = x[0] * y[0] + x[1] * y[1] +...+ x[srcBLen - 1] * y[srcBLen - 1] | |

*/ | |

/* In this stage the MAC operations are increased by 1 for every iteration. | |

The count variable holds the number of MAC operations performed */ | |

count = 1u; | |

/* Working pointer of inputA */ | |

px = pIn1; | |

/* Working pointer of inputB */ | |

pSrc1 = pIn2 + (srcBLen - 1u); | |

py = pSrc1; | |

/* ------------------------ | |

* Stage1 process | |

* ----------------------*/ | |

/* The first loop starts here */ | |

while(blockSize1 > 0u) | |

{ | |

/* Accumulator is made zero for every iteration */ | |

sum = 0; | |

/* Apply loop unrolling and compute 4 MACs simultaneously. */ | |

k = count >> 2; | |

/* First part of the processing with loop unrolling. Compute 4 MACs at a time. | |

** a second loop below computes MACs for the remaining 1 to 3 samples. */ | |

while(k > 0u) | |

{ | |

/* x[0] * y[srcBLen - 4] , x[1] * y[srcBLen - 3] */ | |

sum = __SMLALD(*__SIMD32(px)++, *__SIMD32(py)++, sum); | |

/* x[3] * y[srcBLen - 1] , x[2] * y[srcBLen - 2] */ | |

sum = __SMLALD(*__SIMD32(px)++, *__SIMD32(py)++, sum); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* If the count is not a multiple of 4, compute any remaining MACs here. | |

** No loop unrolling is used. */ | |

k = count % 0x4u; | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulates */ | |

/* x[0] * y[srcBLen - 1] */ | |

sum = __SMLALD(*px++, *py++, sum); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* Store the result in the accumulator in the destination buffer. */ | |

*pOut = (q15_t) (__SSAT((sum >> 15), 16)); | |

/* Destination pointer is updated according to the address modifier, inc */ | |

pOut += inc; | |

/* Update the inputA and inputB pointers for next MAC calculation */ | |

py = pSrc1 - count; | |

px = pIn1; | |

/* Increment the MAC count */ | |

count++; | |

/* Decrement the loop counter */ | |

blockSize1--; | |

} | |

/* -------------------------- | |

* Initializations of stage2 | |

* ------------------------*/ | |

/* sum = x[0] * y[0] + x[1] * y[1] +...+ x[srcBLen-1] * y[srcBLen-1] | |

* sum = x[1] * y[0] + x[2] * y[1] +...+ x[srcBLen] * y[srcBLen-1] | |

* .... | |

* sum = x[srcALen-srcBLen-2] * y[0] + x[srcALen-srcBLen-1] * y[1] +...+ x[srcALen-1] * y[srcBLen-1] | |

*/ | |

/* Working pointer of inputA */ | |

px = pIn1; | |

/* Working pointer of inputB */ | |

py = pIn2; | |

/* count is index by which the pointer pIn1 to be incremented */ | |

count = 0u; | |

/* ------------------- | |

* Stage2 process | |

* ------------------*/ | |

/* Stage2 depends on srcBLen as in this stage srcBLen number of MACS are performed. | |

* So, to loop unroll over blockSize2, | |

* srcBLen should be greater than or equal to 4, to loop unroll the srcBLen loop */ | |

if(srcBLen >= 4u) | |

{ | |

/* Loop unroll over blockSize2, by 4 */ | |

blkCnt = blockSize2 >> 2u; | |

while(blkCnt > 0u) | |

{ | |

/* Set all accumulators to zero */ | |

acc0 = 0; | |

acc1 = 0; | |

acc2 = 0; | |

acc3 = 0; | |

/* read x[0], x[1] samples */ | |

x0 = *__SIMD32(px); | |

/* read x[1], x[2] samples */ | |

x1 = _SIMD32_OFFSET(px + 1); | |

px += 2u; | |

/* Apply loop unrolling and compute 4 MACs simultaneously. */ | |

k = srcBLen >> 2u; | |

/* First part of the processing with loop unrolling. Compute 4 MACs at a time. | |

** a second loop below computes MACs for the remaining 1 to 3 samples. */ | |

do | |

{ | |

/* Read the first two inputB samples using SIMD: | |

* y[0] and y[1] */ | |

c0 = *__SIMD32(py)++; | |

/* acc0 += x[0] * y[0] + x[1] * y[1] */ | |

acc0 = __SMLALD(x0, c0, acc0); | |

/* acc1 += x[1] * y[0] + x[2] * y[1] */ | |

acc1 = __SMLALD(x1, c0, acc1); | |

/* Read x[2], x[3] */ | |

x2 = *__SIMD32(px); | |

/* Read x[3], x[4] */ | |

x3 = _SIMD32_OFFSET(px + 1); | |

/* acc2 += x[2] * y[0] + x[3] * y[1] */ | |

acc2 = __SMLALD(x2, c0, acc2); | |

/* acc3 += x[3] * y[0] + x[4] * y[1] */ | |

acc3 = __SMLALD(x3, c0, acc3); | |

/* Read y[2] and y[3] */ | |

c0 = *__SIMD32(py)++; | |

/* acc0 += x[2] * y[2] + x[3] * y[3] */ | |

acc0 = __SMLALD(x2, c0, acc0); | |

/* acc1 += x[3] * y[2] + x[4] * y[3] */ | |

acc1 = __SMLALD(x3, c0, acc1); | |

/* Read x[4], x[5] */ | |

x0 = _SIMD32_OFFSET(px + 2); | |

/* Read x[5], x[6] */ | |

x1 = _SIMD32_OFFSET(px + 3); | |

px += 4u; | |

/* acc2 += x[4] * y[2] + x[5] * y[3] */ | |

acc2 = __SMLALD(x0, c0, acc2); | |

/* acc3 += x[5] * y[2] + x[6] * y[3] */ | |

acc3 = __SMLALD(x1, c0, acc3); | |

} while(--k); | |

/* If the srcBLen is not a multiple of 4, compute any remaining MACs here. | |

** No loop unrolling is used. */ | |

k = srcBLen % 0x4u; | |

if(k == 1u) | |

{ | |

/* Read y[4] */ | |

c0 = *py; | |

#ifdef ARM_MATH_BIG_ENDIAN | |

c0 = c0 << 16u; | |

#else | |

c0 = c0 & 0x0000FFFF; | |

#endif /* #ifdef ARM_MATH_BIG_ENDIAN */ | |

/* Read x[7] */ | |

x3 = *__SIMD32(px); | |

px++; | |

/* Perform the multiply-accumulates */ | |

acc0 = __SMLALD(x0, c0, acc0); | |

acc1 = __SMLALD(x1, c0, acc1); | |

acc2 = __SMLALDX(x1, c0, acc2); | |

acc3 = __SMLALDX(x3, c0, acc3); | |

} | |

if(k == 2u) | |

{ | |

/* Read y[4], y[5] */ | |

c0 = *__SIMD32(py); | |

/* Read x[7], x[8] */ | |

x3 = *__SIMD32(px); | |

/* Read x[9] */ | |

x2 = _SIMD32_OFFSET(px + 1); | |

px += 2u; | |

/* Perform the multiply-accumulates */ | |

acc0 = __SMLALD(x0, c0, acc0); | |

acc1 = __SMLALD(x1, c0, acc1); | |

acc2 = __SMLALD(x3, c0, acc2); | |

acc3 = __SMLALD(x2, c0, acc3); | |

} | |

if(k == 3u) | |

{ | |

/* Read y[4], y[5] */ | |

c0 = *__SIMD32(py)++; | |

/* Read x[7], x[8] */ | |

x3 = *__SIMD32(px); | |

/* Read x[9] */ | |

x2 = _SIMD32_OFFSET(px + 1); | |

/* Perform the multiply-accumulates */ | |

acc0 = __SMLALD(x0, c0, acc0); | |

acc1 = __SMLALD(x1, c0, acc1); | |

acc2 = __SMLALD(x3, c0, acc2); | |

acc3 = __SMLALD(x2, c0, acc3); | |

c0 = (*py); | |

/* Read y[6] */ | |

#ifdef ARM_MATH_BIG_ENDIAN | |

c0 = c0 << 16u; | |

#else | |

c0 = c0 & 0x0000FFFF; | |

#endif /* #ifdef ARM_MATH_BIG_ENDIAN */ | |

/* Read x[10] */ | |

x3 = _SIMD32_OFFSET(px + 2); | |

px += 3u; | |

/* Perform the multiply-accumulates */ | |

acc0 = __SMLALDX(x1, c0, acc0); | |

acc1 = __SMLALD(x2, c0, acc1); | |

acc2 = __SMLALDX(x2, c0, acc2); | |

acc3 = __SMLALDX(x3, c0, acc3); | |

} | |

/* Store the result in the accumulator in the destination buffer. */ | |

*pOut = (q15_t) (__SSAT(acc0 >> 15, 16)); | |

/* Destination pointer is updated according to the address modifier, inc */ | |

pOut += inc; | |

*pOut = (q15_t) (__SSAT(acc1 >> 15, 16)); | |

pOut += inc; | |

*pOut = (q15_t) (__SSAT(acc2 >> 15, 16)); | |

pOut += inc; | |

*pOut = (q15_t) (__SSAT(acc3 >> 15, 16)); | |

pOut += inc; | |

/* Increment the count by 4 as 4 output values are computed */ | |

count += 4u; | |

/* Update the inputA and inputB pointers for next MAC calculation */ | |

px = pIn1 + count; | |

py = pIn2; | |

/* Decrement the loop counter */ | |

blkCnt--; | |

} | |

/* If the blockSize2 is not a multiple of 4, compute any remaining output samples here. | |

** No loop unrolling is used. */ | |

blkCnt = blockSize2 % 0x4u; | |

while(blkCnt > 0u) | |

{ | |

/* Accumulator is made zero for every iteration */ | |

sum = 0; | |

/* Apply loop unrolling and compute 4 MACs simultaneously. */ | |

k = srcBLen >> 2u; | |

/* First part of the processing with loop unrolling. Compute 4 MACs at a time. | |

** a second loop below computes MACs for the remaining 1 to 3 samples. */ | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulates */ | |

sum += ((q63_t) * px++ * *py++); | |

sum += ((q63_t) * px++ * *py++); | |

sum += ((q63_t) * px++ * *py++); | |

sum += ((q63_t) * px++ * *py++); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* If the srcBLen is not a multiple of 4, compute any remaining MACs here. | |

** No loop unrolling is used. */ | |

k = srcBLen % 0x4u; | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulates */ | |

sum += ((q63_t) * px++ * *py++); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* Store the result in the accumulator in the destination buffer. */ | |

*pOut = (q15_t) (__SSAT(sum >> 15, 16)); | |

/* Destination pointer is updated according to the address modifier, inc */ | |

pOut += inc; | |

/* Increment count by 1, as one output value is computed */ | |

count++; | |

/* Update the inputA and inputB pointers for next MAC calculation */ | |

px = pIn1 + count; | |

py = pIn2; | |

/* Decrement the loop counter */ | |

blkCnt--; | |

} | |

} | |

else | |

{ | |

/* If the srcBLen is not a multiple of 4, | |

* the blockSize2 loop cannot be unrolled by 4 */ | |

blkCnt = blockSize2; | |

while(blkCnt > 0u) | |

{ | |

/* Accumulator is made zero for every iteration */ | |

sum = 0; | |

/* Loop over srcBLen */ | |

k = srcBLen; | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulate */ | |

sum += ((q63_t) * px++ * *py++); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* Store the result in the accumulator in the destination buffer. */ | |

*pOut = (q15_t) (__SSAT(sum >> 15, 16)); | |

/* Destination pointer is updated according to the address modifier, inc */ | |

pOut += inc; | |

/* Increment the MAC count */ | |

count++; | |

/* Update the inputA and inputB pointers for next MAC calculation */ | |

px = pIn1 + count; | |

py = pIn2; | |

/* Decrement the loop counter */ | |

blkCnt--; | |

} | |

} | |

/* -------------------------- | |

* Initializations of stage3 | |

* -------------------------*/ | |

/* sum += x[srcALen-srcBLen+1] * y[0] + x[srcALen-srcBLen+2] * y[1] +...+ x[srcALen-1] * y[srcBLen-1] | |

* sum += x[srcALen-srcBLen+2] * y[0] + x[srcALen-srcBLen+3] * y[1] +...+ x[srcALen-1] * y[srcBLen-1] | |

* .... | |

* sum += x[srcALen-2] * y[0] + x[srcALen-1] * y[1] | |

* sum += x[srcALen-1] * y[0] | |

*/ | |

/* In this stage the MAC operations are decreased by 1 for every iteration. | |

The count variable holds the number of MAC operations performed */ | |

count = srcBLen - 1u; | |

/* Working pointer of inputA */ | |

pSrc1 = (pIn1 + srcALen) - (srcBLen - 1u); | |

px = pSrc1; | |

/* Working pointer of inputB */ | |

py = pIn2; | |

/* ------------------- | |

* Stage3 process | |

* ------------------*/ | |

while(blockSize3 > 0u) | |

{ | |

/* Accumulator is made zero for every iteration */ | |

sum = 0; | |

/* Apply loop unrolling and compute 4 MACs simultaneously. */ | |

k = count >> 2u; | |

/* First part of the processing with loop unrolling. Compute 4 MACs at a time. | |

** a second loop below computes MACs for the remaining 1 to 3 samples. */ | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulates */ | |

/* sum += x[srcALen - srcBLen + 4] * y[3] , sum += x[srcALen - srcBLen + 3] * y[2] */ | |

sum = __SMLALD(*__SIMD32(px)++, *__SIMD32(py)++, sum); | |

/* sum += x[srcALen - srcBLen + 2] * y[1] , sum += x[srcALen - srcBLen + 1] * y[0] */ | |

sum = __SMLALD(*__SIMD32(px)++, *__SIMD32(py)++, sum); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* If the count is not a multiple of 4, compute any remaining MACs here. | |

** No loop unrolling is used. */ | |

k = count % 0x4u; | |

while(k > 0u) | |

{ | |

/* Perform the multiply-accumulates */ | |

sum = __SMLALD(*px++, *py++, sum); | |

/* Decrement the loop counter */ | |

k--; | |

} | |

/* Store the result in the accumulator in the destination buffer. */ | |

*pOut = (q15_t) (__SSAT((sum >> 15), 16)); | |

/* Destination pointer is updated according to the address modifier, inc */ | |

pOut += inc; | |

/* Update the inputA and inputB pointers for next MAC calculation */ | |

px = ++pSrc1; | |

py = pIn2; | |

/* Decrement the MAC count */ | |

count--; | |

/* Decrement the loop counter */ | |

blockSize3--; | |

} | |

#else | |

/* Run the below code for Cortex-M0 */ | |

q15_t *pIn1 = pSrcA; /* inputA pointer */ | |

q15_t *pIn2 = pSrcB + (srcBLen - 1u); /* inputB pointer */ | |

q63_t sum; /* Accumulators */ | |

uint32_t i = 0u, j; /* loop counters */ | |

uint32_t inv = 0u; /* Reverse order flag */ | |

uint32_t tot = 0u; /* Length */ | |

/* The algorithm implementation is based on the lengths of the inputs. */ | |

/* srcB is always made to slide across srcA. */ | |

/* So srcBLen is always considered as shorter or equal to srcALen */ | |

/* But CORR(x, y) is reverse of CORR(y, x) */ | |

/* So, when srcBLen > srcALen, output pointer is made to point to the end of the output buffer */ | |

/* and a varaible, inv is set to 1 */ | |

/* If lengths are not equal then zero pad has to be done to make the two | |

* inputs of same length. But to improve the performance, we include zeroes | |

* in the output instead of zero padding either of the the inputs*/ | |

/* If srcALen > srcBLen, (srcALen - srcBLen) zeroes has to included in the | |

* starting of the output buffer */ | |

/* If srcALen < srcBLen, (srcALen - srcBLen) zeroes has to included in the | |

* ending of the output buffer */ | |

/* Once the zero padding is done the remaining of the output is calcualted | |

* using convolution but with the shorter signal time shifted. */ | |

/* Calculate the length of the remaining sequence */ | |

tot = ((srcALen + srcBLen) - 2u); | |

if(srcALen > srcBLen) | |

{ | |

/* Calculating the number of zeros to be padded to the output */ | |

j = srcALen - srcBLen; | |

/* Initialise the pointer after zero padding */ | |

pDst += j; | |

} | |

else if(srcALen < srcBLen) | |

{ | |

/* Initialization to inputB pointer */ | |

pIn1 = pSrcB; | |

/* Initialization to the end of inputA pointer */ | |

pIn2 = pSrcA + (srcALen - 1u); | |

/* Initialisation of the pointer after zero padding */ | |

pDst = pDst + tot; | |

/* Swapping the lengths */ | |

j = srcALen; | |

srcALen = srcBLen; | |

srcBLen = j; | |

/* Setting the reverse flag */ | |

inv = 1; | |

} | |

/* Loop to calculate convolution for output length number of times */ | |

for (i = 0u; i <= tot; i++) | |

{ | |

/* Initialize sum with zero to carry on MAC operations */ | |

sum = 0; | |

/* Loop to perform MAC operations according to convolution equation */ | |

for (j = 0u; j <= i; j++) | |

{ | |

/* Check the array limitations */ | |

if((((i - j) < srcBLen) && (j < srcALen))) | |

{ | |

/* z[i] += x[i-j] * y[j] */ | |

sum += ((q31_t) pIn1[j] * pIn2[-((int32_t) i - j)]); | |

} | |

} | |

/* Store the output in the destination buffer */ | |

if(inv == 1) | |

*pDst-- = (q15_t) __SSAT((sum >> 15u), 16u); | |

else | |

*pDst++ = (q15_t) __SSAT((sum >> 15u), 16u); | |

} | |

#endif /*#if (defined(ARM_MATH_CM4) || defined(ARM_MATH_CM3)) && !defined(UNALIGNED_SUPPORT_DISABLE) */ | |

} | |

/** | |

* @} end of Corr group | |

*/ |