123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339 |
- /* Copyright (C) 2007-2008 Jean-Marc Valin
- * Copyright (C) 2008 Thorvald Natvig
- * Copyright (C) 2011 Texas Instruments
- * author Jyri Sarha
- */
- /**
- @file resample_neon.h
- @brief Resampler functions (NEON version)
- */
- /*
- Redistribution and use in source and binary forms, with or without
- modification, are permitted provided that the following conditions
- are met:
- - Redistributions of source code must retain the above copyright
- notice, this list of conditions and the following disclaimer.
- - Redistributions in binary form must reproduce the above copyright
- notice, this list of conditions and the following disclaimer in the
- documentation and/or other materials provided with the distribution.
- - Neither the name of the Xiph.org Foundation nor the names of its
- contributors may be used to endorse or promote products derived from
- this software without specific prior written permission.
- THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR
- CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
- EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
- PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
- PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
- LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
- NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
- SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
- */
- #ifdef FIXED_POINT
- #if defined(__aarch64__)
- static inline int32_t saturate_32bit_to_16bit(int32_t a) {
- int32_t ret;
- asm ("fmov s0, %w[a]\n"
- "sqxtn h0, s0\n"
- "sxtl v0.4s, v0.4h\n"
- "fmov %w[ret], s0\n"
- : [ret] "=r" (ret)
- : [a] "r" (a)
- : "v0" );
- return ret;
- }
- #elif defined(__thumb2__)
- static inline int32_t saturate_32bit_to_16bit(int32_t a) {
- int32_t ret;
- asm ("ssat %[ret], #16, %[a]"
- : [ret] "=r" (ret)
- : [a] "r" (a)
- : );
- return ret;
- }
- #else
- static inline int32_t saturate_32bit_to_16bit(int32_t a) {
- int32_t ret;
- asm ("vmov.s32 d0[0], %[a]\n"
- "vqmovn.s32 d0, q0\n"
- "vmov.s16 %[ret], d0[0]\n"
- : [ret] "=r" (ret)
- : [a] "r" (a)
- : "q0");
- return ret;
- }
- #endif
- #undef WORD2INT
- #define WORD2INT(x) (saturate_32bit_to_16bit(x))
- #define OVERRIDE_INNER_PRODUCT_SINGLE
- /* Only works when len % 4 == 0 and len >= 4 */
- #if defined(__aarch64__)
- static inline int32_t inner_product_single(const int16_t *a, const int16_t *b, unsigned int len)
- {
- int32_t ret;
- uint32_t remainder = len % 16;
- len = len - remainder;
- asm volatile (" cmp %w[len], #0\n"
- " b.ne 1f\n"
- " ld1 {v16.4h}, [%[b]], #8\n"
- " ld1 {v20.4h}, [%[a]], #8\n"
- " subs %w[remainder], %w[remainder], #4\n"
- " smull v0.4s, v16.4h, v20.4h\n"
- " b.ne 4f\n"
- " b 5f\n"
- "1:"
- " ld1 {v16.4h, v17.4h, v18.4h, v19.4h}, [%[b]], #32\n"
- " ld1 {v20.4h, v21.4h, v22.4h, v23.4h}, [%[a]], #32\n"
- " subs %w[len], %w[len], #16\n"
- " smull v0.4s, v16.4h, v20.4h\n"
- " smlal v0.4s, v17.4h, v21.4h\n"
- " smlal v0.4s, v18.4h, v22.4h\n"
- " smlal v0.4s, v19.4h, v23.4h\n"
- " b.eq 3f\n"
- "2:"
- " ld1 {v16.4h, v17.4h, v18.4h, v19.4h}, [%[b]], #32\n"
- " ld1 {v20.4h, v21.4h, v22.4h, v23.4h}, [%[a]], #32\n"
- " subs %w[len], %w[len], #16\n"
- " smlal v0.4s, v16.4h, v20.4h\n"
- " smlal v0.4s, v17.4h, v21.4h\n"
- " smlal v0.4s, v18.4h, v22.4h\n"
- " smlal v0.4s, v19.4h, v23.4h\n"
- " b.ne 2b\n"
- "3:"
- " cmp %w[remainder], #0\n"
- " b.eq 5f\n"
- "4:"
- " ld1 {v18.4h}, [%[b]], #8\n"
- " ld1 {v22.4h}, [%[a]], #8\n"
- " subs %w[remainder], %w[remainder], #4\n"
- " smlal v0.4s, v18.4h, v22.4h\n"
- " b.ne 4b\n"
- "5:"
- " saddlv d0, v0.4s\n"
- " sqxtn s0, d0\n"
- " sqrshrn h0, s0, #15\n"
- " sxtl v0.4s, v0.4h\n"
- " fmov %w[ret], s0\n"
- : [ret] "=r" (ret), [a] "+r" (a), [b] "+r" (b),
- [len] "+r" (len), [remainder] "+r" (remainder)
- :
- : "cc", "v0",
- "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23");
- return ret;
- }
- #else
- static inline int32_t inner_product_single(const int16_t *a, const int16_t *b, unsigned int len)
- {
- int32_t ret;
- uint32_t remainder = len % 16;
- len = len - remainder;
- asm volatile (" cmp %[len], #0\n"
- " bne 1f\n"
- " vld1.16 {d16}, [%[b]]!\n"
- " vld1.16 {d20}, [%[a]]!\n"
- " subs %[remainder], %[remainder], #4\n"
- " vmull.s16 q0, d16, d20\n"
- " beq 5f\n"
- " b 4f\n"
- "1:"
- " vld1.16 {d16, d17, d18, d19}, [%[b]]!\n"
- " vld1.16 {d20, d21, d22, d23}, [%[a]]!\n"
- " subs %[len], %[len], #16\n"
- " vmull.s16 q0, d16, d20\n"
- " vmlal.s16 q0, d17, d21\n"
- " vmlal.s16 q0, d18, d22\n"
- " vmlal.s16 q0, d19, d23\n"
- " beq 3f\n"
- "2:"
- " vld1.16 {d16, d17, d18, d19}, [%[b]]!\n"
- " vld1.16 {d20, d21, d22, d23}, [%[a]]!\n"
- " subs %[len], %[len], #16\n"
- " vmlal.s16 q0, d16, d20\n"
- " vmlal.s16 q0, d17, d21\n"
- " vmlal.s16 q0, d18, d22\n"
- " vmlal.s16 q0, d19, d23\n"
- " bne 2b\n"
- "3:"
- " cmp %[remainder], #0\n"
- " beq 5f\n"
- "4:"
- " vld1.16 {d16}, [%[b]]!\n"
- " vld1.16 {d20}, [%[a]]!\n"
- " subs %[remainder], %[remainder], #4\n"
- " vmlal.s16 q0, d16, d20\n"
- " bne 4b\n"
- "5:"
- " vaddl.s32 q0, d0, d1\n"
- " vadd.s64 d0, d0, d1\n"
- " vqmovn.s64 d0, q0\n"
- " vqrshrn.s32 d0, q0, #15\n"
- " vmov.s16 %[ret], d0[0]\n"
- : [ret] "=r" (ret), [a] "+r" (a), [b] "+r" (b),
- [len] "+r" (len), [remainder] "+r" (remainder)
- :
- : "cc", "q0",
- "d16", "d17", "d18", "d19", "d20", "d21", "d22", "d23");
- return ret;
- }
- #endif // !defined(__aarch64__)
- #elif defined(FLOATING_POINT)
- #if defined(__aarch64__)
- static inline int32_t saturate_float_to_16bit(float a) {
- int32_t ret;
- asm ("fcvtas s1, %s[a]\n"
- "sqxtn h1, s1\n"
- "sxtl v1.4s, v1.4h\n"
- "fmov %w[ret], s1\n"
- : [ret] "=r" (ret)
- : [a] "w" (a)
- : "v1");
- return ret;
- }
- #else
- static inline int32_t saturate_float_to_16bit(float a) {
- int32_t ret;
- asm ("vmov.f32 d0[0], %[a]\n"
- "vcvt.s32.f32 d0, d0, #15\n"
- "vqrshrn.s32 d0, q0, #15\n"
- "vmov.s16 %[ret], d0[0]\n"
- : [ret] "=r" (ret)
- : [a] "r" (a)
- : "q0");
- return ret;
- }
- #endif
- #undef WORD2INT
- #define WORD2INT(x) (saturate_float_to_16bit(x))
- #define OVERRIDE_INNER_PRODUCT_SINGLE
- /* Only works when len % 4 == 0 and len >= 4 */
- #if defined(__aarch64__)
- static inline float inner_product_single(const float *a, const float *b, unsigned int len)
- {
- float ret;
- uint32_t remainder = len % 16;
- len = len - remainder;
- asm volatile (" cmp %w[len], #0\n"
- " b.ne 1f\n"
- " ld1 {v16.4s}, [%[b]], #16\n"
- " ld1 {v20.4s}, [%[a]], #16\n"
- " subs %w[remainder], %w[remainder], #4\n"
- " fmul v1.4s, v16.4s, v20.4s\n"
- " b.ne 4f\n"
- " b 5f\n"
- "1:"
- " ld1 {v16.4s, v17.4s, v18.4s, v19.4s}, [%[b]], #64\n"
- " ld1 {v20.4s, v21.4s, v22.4s, v23.4s}, [%[a]], #64\n"
- " subs %w[len], %w[len], #16\n"
- " fmul v1.4s, v16.4s, v20.4s\n"
- " fmul v2.4s, v17.4s, v21.4s\n"
- " fmul v3.4s, v18.4s, v22.4s\n"
- " fmul v4.4s, v19.4s, v23.4s\n"
- " b.eq 3f\n"
- "2:"
- " ld1 {v16.4s, v17.4s, v18.4s, v19.4s}, [%[b]], #64\n"
- " ld1 {v20.4s, v21.4s, v22.4s, v23.4s}, [%[a]], #64\n"
- " subs %w[len], %w[len], #16\n"
- " fmla v1.4s, v16.4s, v20.4s\n"
- " fmla v2.4s, v17.4s, v21.4s\n"
- " fmla v3.4s, v18.4s, v22.4s\n"
- " fmla v4.4s, v19.4s, v23.4s\n"
- " b.ne 2b\n"
- "3:"
- " fadd v16.4s, v1.4s, v2.4s\n"
- " fadd v17.4s, v3.4s, v4.4s\n"
- " cmp %w[remainder], #0\n"
- " fadd v1.4s, v16.4s, v17.4s\n"
- " b.eq 5f\n"
- "4:"
- " ld1 {v18.4s}, [%[b]], #16\n"
- " ld1 {v22.4s}, [%[a]], #16\n"
- " subs %w[remainder], %w[remainder], #4\n"
- " fmla v1.4s, v18.4s, v22.4s\n"
- " b.ne 4b\n"
- "5:"
- " faddp v1.4s, v1.4s, v1.4s\n"
- " faddp %[ret].4s, v1.4s, v1.4s\n"
- : [ret] "=w" (ret), [a] "+r" (a), [b] "+r" (b),
- [len] "+r" (len), [remainder] "+r" (remainder)
- :
- : "cc", "v1", "v2", "v3", "v4",
- "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23");
- return ret;
- }
- #else
- static inline float inner_product_single(const float *a, const float *b, unsigned int len)
- {
- float ret;
- uint32_t remainder = len % 16;
- len = len - remainder;
- asm volatile (" cmp %[len], #0\n"
- " bne 1f\n"
- " vld1.32 {q4}, [%[b]]!\n"
- " vld1.32 {q8}, [%[a]]!\n"
- " subs %[remainder], %[remainder], #4\n"
- " vmul.f32 q0, q4, q8\n"
- " bne 4f\n"
- " b 5f\n"
- "1:"
- " vld1.32 {q4, q5}, [%[b]]!\n"
- " vld1.32 {q8, q9}, [%[a]]!\n"
- " vld1.32 {q6, q7}, [%[b]]!\n"
- " vld1.32 {q10, q11}, [%[a]]!\n"
- " subs %[len], %[len], #16\n"
- " vmul.f32 q0, q4, q8\n"
- " vmul.f32 q1, q5, q9\n"
- " vmul.f32 q2, q6, q10\n"
- " vmul.f32 q3, q7, q11\n"
- " beq 3f\n"
- "2:"
- " vld1.32 {q4, q5}, [%[b]]!\n"
- " vld1.32 {q8, q9}, [%[a]]!\n"
- " vld1.32 {q6, q7}, [%[b]]!\n"
- " vld1.32 {q10, q11}, [%[a]]!\n"
- " subs %[len], %[len], #16\n"
- " vmla.f32 q0, q4, q8\n"
- " vmla.f32 q1, q5, q9\n"
- " vmla.f32 q2, q6, q10\n"
- " vmla.f32 q3, q7, q11\n"
- " bne 2b\n"
- "3:"
- " vadd.f32 q4, q0, q1\n"
- " vadd.f32 q5, q2, q3\n"
- " cmp %[remainder], #0\n"
- " vadd.f32 q0, q4, q5\n"
- " beq 5f\n"
- "4:"
- " vld1.32 {q6}, [%[b]]!\n"
- " vld1.32 {q10}, [%[a]]!\n"
- " subs %[remainder], %[remainder], #4\n"
- " vmla.f32 q0, q6, q10\n"
- " bne 4b\n"
- "5:"
- " vadd.f32 d0, d0, d1\n"
- " vpadd.f32 d0, d0, d0\n"
- " vmov.f32 %[ret], d0[0]\n"
- : [ret] "=r" (ret), [a] "+r" (a), [b] "+r" (b),
- [len] "+l" (len), [remainder] "+l" (remainder)
- :
- : "cc", "q0", "q1", "q2", "q3",
- "q4", "q5", "q6", "q7", "q8", "q9", "q10", "q11");
- return ret;
- }
- #endif // defined(__aarch64__)
- #endif
|