diff options
Diffstat (limited to 'gst/speexresample/fixed_arm5e.h')
-rw-r--r-- | gst/speexresample/fixed_arm5e.h | 166 |
1 files changed, 166 insertions, 0 deletions
diff --git a/gst/speexresample/fixed_arm5e.h b/gst/speexresample/fixed_arm5e.h new file mode 100644 index 00000000..c995d54b --- /dev/null +++ b/gst/speexresample/fixed_arm5e.h @@ -0,0 +1,166 @@ +/* Copyright (C) 2003 Jean-Marc Valin */ +/** + @file fixed_arm5e.h + @brief ARM-tuned fixed-point operations +*/ +/* + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions + are met: + + - Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + + - Redistributions in binary form must reproduce the above copyright + notice, this list of conditions and the following disclaimer in the + documentation and/or other materials provided with the distribution. + + - Neither the name of the Xiph.org Foundation nor the names of its + contributors may be used to endorse or promote products derived from + this software without specific prior written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR + CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, + EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, + PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR + PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF + LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING + NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS + SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +*/ + +#ifndef FIXED_ARM5E_H +#define FIXED_ARM5E_H + +#undef MULT16_16 +static inline spx_word32_t +MULT16_16 (spx_word16_t x, spx_word16_t y) +{ + int res; +asm ("smulbb %0,%1,%2;\n":"=&r" (res) +: "%r" (x), "r" (y)); + return (res); +} + +#undef MAC16_16 +static inline spx_word32_t +MAC16_16 (spx_word32_t a, spx_word16_t x, spx_word32_t y) +{ + int res; +asm ("smlabb %0,%1,%2,%3;\n":"=&r" (res) +: "%r" (x), "r" (y), "r" (a)); + return (res); +} + +#undef MULT16_32_Q15 +static inline spx_word32_t +MULT16_32_Q15 (spx_word16_t x, spx_word32_t y) +{ + int res; +asm ("smulwb %0,%1,%2;\n":"=&r" (res) +: "%r" (y << 1), "r" (x)); + return (res); +} + +#undef MAC16_32_Q15 +static inline spx_word32_t +MAC16_32_Q15 (spx_word32_t a, spx_word16_t x, spx_word32_t y) +{ + int res; +asm ("smlawb %0,%1,%2,%3;\n":"=&r" (res) +: "%r" (y << 1), "r" (x), "r" (a)); + return (res); +} + +#undef MULT16_32_Q11 +static inline spx_word32_t +MULT16_32_Q11 (spx_word16_t x, spx_word32_t y) +{ + int res; +asm ("smulwb %0,%1,%2;\n":"=&r" (res) +: "%r" (y << 5), "r" (x)); + return (res); +} + +#undef MAC16_32_Q11 +static inline spx_word32_t +MAC16_32_Q11 (spx_word32_t a, spx_word16_t x, spx_word32_t y) +{ + int res; +asm ("smlawb %0,%1,%2,%3;\n":"=&r" (res) +: "%r" (y << 5), "r" (x), "r" (a)); + return (res); +} + +#undef DIV32_16 +static inline short +DIV32_16 (int a, int b) +{ + int res = 0; + int dead1, dead2, dead3, dead4, dead5; + __asm__ __volatile__ ("\teor %5, %0, %1\n" + "\tmovs %4, %0\n" + "\trsbmi %0, %0, #0 \n" + "\tmovs %4, %1\n" + "\trsbmi %1, %1, #0 \n" + "\tmov %4, #1\n" + "\tsubs %3, %0, %1, asl #14 \n" + "\torrpl %2, %2, %4, asl #14 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #13 \n" + "\torrpl %2, %2, %4, asl #13 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #12 \n" + "\torrpl %2, %2, %4, asl #12 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #11 \n" + "\torrpl %2, %2, %4, asl #11 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #10 \n" + "\torrpl %2, %2, %4, asl #10 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #9 \n" + "\torrpl %2, %2, %4, asl #9 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #8 \n" + "\torrpl %2, %2, %4, asl #8 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #7 \n" + "\torrpl %2, %2, %4, asl #7 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #6 \n" + "\torrpl %2, %2, %4, asl #6 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #5 \n" + "\torrpl %2, %2, %4, asl #5 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #4 \n" + "\torrpl %2, %2, %4, asl #4 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #3 \n" + "\torrpl %2, %2, %4, asl #3 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #2 \n" + "\torrpl %2, %2, %4, asl #2 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1, asl #1 \n" + "\torrpl %2, %2, %4, asl #1 \n" + "\tmovpl %0, %3 \n" + "\tsubs %3, %0, %1 \n" + "\torrpl %2, %2, %4 \n" + "\tmovpl %0, %3 \n" + "\tmovs %5, %5, lsr #31 \n" + "\trsbne %2, %2, #0 \n":"=r" (dead1), "=r" (dead2), "=r" (res), + "=r" (dead3), "=r" (dead4), "=r" (dead5) + :"0" (a), "1" (b), "2" (res) + :"memory", "cc"); + return res; +} + + + + +#endif |