;****************************************************************************** ;* x86 optimized Format Conversion Utils ;* Copyright (c) 2008 Loren Merritt ;* Copyright (c) 2012 Justin Ruggles ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" %include "util.asm" SECTION_RODATA 32 pf_s32_inv_scale: times 8 dd 0x30000000 pf_s32_scale: times 8 dd 0x4f000000 pf_s32_clip: times 8 dd 0x4effffff pf_s16_inv_scale: times 4 dd 0x38000000 pf_s16_scale: times 4 dd 0x47000000 pb_shuf_unpack_even: db -1, -1, 0, 1, -1, -1, 2, 3, -1, -1, 8, 9, -1, -1, 10, 11 pb_shuf_unpack_odd: db -1, -1, 4, 5, -1, -1, 6, 7, -1, -1, 12, 13, -1, -1, 14, 15 pb_interleave_words: SHUFFLE_MASK_W 0, 4, 1, 5, 2, 6, 3, 7 pb_deinterleave_words: SHUFFLE_MASK_W 0, 2, 4, 6, 1, 3, 5, 7 pw_zero_even: times 4 dw 0x0000, 0xffff SECTION_TEXT ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_s32(int32_t *dst, const int16_t *src, int len); ;------------------------------------------------------------------------------ INIT_XMM sse2 cglobal conv_s16_to_s32, 3,3,3, dst, src, len lea lenq, [2*lend] lea dstq, [dstq+2*lenq] add srcq, lenq neg lenq .loop: mova m2, [srcq+lenq] pxor m0, m0 pxor m1, m1 punpcklwd m0, m2 punpckhwd m1, m2 mova [dstq+2*lenq ], m0 mova [dstq+2*lenq+mmsize], m1 add lenq, mmsize jl .loop REP_RET ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_flt(float *dst, const int16_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_FLT 0 cglobal conv_s16_to_flt, 3,3,3, dst, src, len lea lenq, [2*lend] add srcq, lenq lea dstq, [dstq + 2*lenq] neg lenq mova m2, [pf_s16_inv_scale] ALIGN 16 .loop: mova m0, [srcq+lenq] S16_TO_S32_SX 0, 1 cvtdq2ps m0, m0 cvtdq2ps m1, m1 mulps m0, m2 mulps m1, m2 mova [dstq+2*lenq ], m0 mova [dstq+2*lenq+mmsize], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_FLT INIT_XMM sse4 CONV_S16_TO_FLT ;------------------------------------------------------------------------------ ; void ff_conv_s32_to_s16(int16_t *dst, const int32_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S32_TO_S16 0 cglobal conv_s32_to_s16, 3,3,4, dst, src, len lea lenq, [2*lend] lea srcq, [srcq+2*lenq] add dstq, lenq neg lenq .loop: mova m0, [srcq+2*lenq ] mova m1, [srcq+2*lenq+ mmsize] mova m2, [srcq+2*lenq+2*mmsize] mova m3, [srcq+2*lenq+3*mmsize] psrad m0, 16 psrad m1, 16 psrad m2, 16 psrad m3, 16 packssdw m0, m1 packssdw m2, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop %if mmsize == 8 emms RET %else REP_RET %endif %endmacro INIT_MMX mmx CONV_S32_TO_S16 INIT_XMM sse2 CONV_S32_TO_S16 ;------------------------------------------------------------------------------ ; void ff_conv_s32_to_flt(float *dst, const int32_t *src, int len); ;------------------------------------------------------------------------------ %macro CONV_S32_TO_FLT 0 cglobal conv_s32_to_flt, 3,3,3, dst, src, len lea lenq, [4*lend] add srcq, lenq add dstq, lenq neg lenq mova m0, [pf_s32_inv_scale] ALIGN 16 .loop: cvtdq2ps m1, [srcq+lenq ] cvtdq2ps m2, [srcq+lenq+mmsize] mulps m1, m1, m0 mulps m2, m2, m0 mova [dstq+lenq ], m1 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S32_TO_FLT %if HAVE_AVX_EXTERNAL INIT_YMM avx CONV_S32_TO_FLT %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s16(int16_t *dst, const float *src, int len); ;------------------------------------------------------------------------------ INIT_XMM sse2 cglobal conv_flt_to_s16, 3,3,5, dst, src, len lea lenq, [2*lend] lea srcq, [srcq+2*lenq] add dstq, lenq neg lenq mova m4, [pf_s16_scale] .loop: mova m0, [srcq+2*lenq ] mova m1, [srcq+2*lenq+1*mmsize] mova m2, [srcq+2*lenq+2*mmsize] mova m3, [srcq+2*lenq+3*mmsize] mulps m0, m4 mulps m1, m4 mulps m2, m4 mulps m3, m4 cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 packssdw m0, m1 packssdw m2, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+mmsize], m2 add lenq, mmsize*2 jl .loop REP_RET ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s32(int32_t *dst, const float *src, int len); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_S32 0 cglobal conv_flt_to_s32, 3,3,6, dst, src, len lea lenq, [lend*4] add srcq, lenq add dstq, lenq neg lenq mova m4, [pf_s32_scale] mova m5, [pf_s32_clip] .loop: mulps m0, m4, [srcq+lenq ] mulps m1, m4, [srcq+lenq+1*mmsize] mulps m2, m4, [srcq+lenq+2*mmsize] mulps m3, m4, [srcq+lenq+3*mmsize] minps m0, m0, m5 minps m1, m1, m5 minps m2, m2, m5 minps m3, m3, m5 cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 mova [dstq+lenq ], m0 mova [dstq+lenq+1*mmsize], m1 mova [dstq+lenq+2*mmsize], m2 mova [dstq+lenq+3*mmsize], m3 add lenq, mmsize*4 jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLT_TO_S32 %if HAVE_AVX_EXTERNAL INIT_YMM avx CONV_FLT_TO_S32 %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16p_to_s16_2ch(int16_t *dst, int16_t *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16P_TO_S16_2CH 0 cglobal conv_s16p_to_s16_2ch, 3,4,5, dst, src0, len, src1 mov src1q, [src0q+gprsize] mov src0q, [src0q ] lea lenq, [2*lend] add src0q, lenq add src1q, lenq lea dstq, [dstq+2*lenq] neg lenq .loop: mova m0, [src0q+lenq ] mova m1, [src1q+lenq ] mova m2, [src0q+lenq+mmsize] mova m3, [src1q+lenq+mmsize] SBUTTERFLY2 wd, 0, 1, 4 SBUTTERFLY2 wd, 2, 3, 4 mova [dstq+2*lenq+0*mmsize], m0 mova [dstq+2*lenq+1*mmsize], m1 mova [dstq+2*lenq+2*mmsize], m2 mova [dstq+2*lenq+3*mmsize], m3 add lenq, 2*mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16P_TO_S16_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16P_TO_S16_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16p_to_s16_6ch(int16_t *dst, int16_t *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ ;------------------------------------------------------------------------------ ; NOTE: In the 6-channel functions, len could be used as an index on x86-64 ; instead of just a counter, which would avoid incrementing the ; pointers, but the extra complexity and amount of code is not worth ; the small gain. On x86-32 there are not enough registers to use len ; as an index without keeping two of the pointers on the stack and ; loading them in each iteration. ;------------------------------------------------------------------------------ %macro CONV_S16P_TO_S16_6CH 0 %if ARCH_X86_64 cglobal conv_s16p_to_s16_6ch, 3,8,7, dst, src0, len, src1, src2, src3, src4, src5 %else cglobal conv_s16p_to_s16_6ch, 2,7,7, dst, src0, src1, src2, src3, src4, src5 %define lend dword r2m %endif mov src1q, [src0q+1*gprsize] mov src2q, [src0q+2*gprsize] mov src3q, [src0q+3*gprsize] mov src4q, [src0q+4*gprsize] mov src5q, [src0q+5*gprsize] mov src0q, [src0q] sub src1q, src0q sub src2q, src0q sub src3q, src0q sub src4q, src0q sub src5q, src0q .loop: %if cpuflag(sse2slow) movq m0, [src0q ] ; m0 = 0, 6, 12, 18, x, x, x, x movq m1, [src0q+src1q] ; m1 = 1, 7, 13, 19, x, x, x, x movq m2, [src0q+src2q] ; m2 = 2, 8, 14, 20, x, x, x, x movq m3, [src0q+src3q] ; m3 = 3, 9, 15, 21, x, x, x, x movq m4, [src0q+src4q] ; m4 = 4, 10, 16, 22, x, x, x, x movq m5, [src0q+src5q] ; m5 = 5, 11, 17, 23, x, x, x, x ; unpack words: punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19 punpcklwd m2, m3 ; m2 = 4, 5, 10, 11, 16, 17, 22, 23 punpcklwd m4, m5 ; m4 = 2, 3, 8, 9, 14, 15, 20, 21 ; blend dwords shufps m1, m0, m2, q2020 ; m1 = 0, 1, 12, 13, 2, 3, 14, 15 shufps m0, m4, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17 shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23 ; shuffle dwords pshufd m0, m0, q1302 ; m0 = 4, 5, 6, 7, 16, 17, 18, 19 pshufd m1, m1, q3120 ; m1 = 0, 1, 2, 3, 12, 13, 14, 15 pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23 movq [dstq+0*mmsize/2], m1 movq [dstq+1*mmsize/2], m0 movq [dstq+2*mmsize/2], m2 movhps [dstq+3*mmsize/2], m1 movhps [dstq+4*mmsize/2], m0 movhps [dstq+5*mmsize/2], m2 add src0q, mmsize/2 add dstq, mmsize*3 sub lend, mmsize/4 %else mova m0, [src0q ] ; m0 = 0, 6, 12, 18, 24, 30, 36, 42 mova m1, [src0q+src1q] ; m1 = 1, 7, 13, 19, 25, 31, 37, 43 mova m2, [src0q+src2q] ; m2 = 2, 8, 14, 20, 26, 32, 38, 44 mova m3, [src0q+src3q] ; m3 = 3, 9, 15, 21, 27, 33, 39, 45 mova m4, [src0q+src4q] ; m4 = 4, 10, 16, 22, 28, 34, 40, 46 mova m5, [src0q+src5q] ; m5 = 5, 11, 17, 23, 29, 35, 41, 47 ; unpack words: SBUTTERFLY2 wd, 0, 1, 6 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19 ; m1 = 24, 25, 30, 31, 36, 37, 42, 43 SBUTTERFLY2 wd, 2, 3, 6 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21 ; m3 = 26, 27, 32, 33, 38, 39, 44, 45 SBUTTERFLY2 wd, 4, 5, 6 ; m4 = 4, 5, 10, 11, 16, 17, 22, 23 ; m5 = 28, 29, 34, 35, 40, 41, 46, 47 ; blend dwords shufps m6, m0, m2, q2020 ; m6 = 0, 1, 12, 13, 2, 3, 14, 15 shufps m0, m4, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17 shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23 SWAP 4,6 ; m4 = 0, 1, 12, 13, 2, 3, 14, 15 shufps m6, m1, m3, q2020 ; m6 = 24, 25, 36, 37, 26, 27, 38, 39 shufps m1, m5, q2031 ; m1 = 30, 31, 42, 43, 28, 29, 40, 41 shufps m3, m5, q3131 ; m3 = 32, 33, 44, 45, 34, 35, 46, 47 SWAP 5,6 ; m5 = 24, 25, 36, 37, 26, 27, 38, 39 ; shuffle dwords pshufd m0, m0, q1302 ; m0 = 4, 5, 6, 7, 16, 17, 18, 19 pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23 pshufd m4, m4, q3120 ; m4 = 0, 1, 2, 3, 12, 13, 14, 15 pshufd m1, m1, q1302 ; m1 = 28, 29, 30, 31, 40, 41, 42, 43 pshufd m3, m3, q3120 ; m3 = 32, 33, 34, 35, 44, 45, 46, 47 pshufd m5, m5, q3120 ; m5 = 24, 25, 26, 27, 36, 37, 38, 39 ; shuffle qwords punpcklqdq m6, m4, m0 ; m6 = 0, 1, 2, 3, 4, 5, 6, 7 punpckhqdq m0, m2 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23 shufps m2, m4, q3210 ; m2 = 8, 9, 10, 11, 12, 13, 14, 15 SWAP 4,6 ; m4 = 0, 1, 2, 3, 4, 5, 6, 7 punpcklqdq m6, m5, m1 ; m6 = 24, 25, 26, 27, 28, 29, 30, 31 punpckhqdq m1, m3 ; m1 = 40, 41, 42, 43, 44, 45, 46, 47 shufps m3, m5, q3210 ; m3 = 32, 33, 34, 35, 36, 37, 38, 39 SWAP 5,6 ; m5 = 24, 25, 26, 27, 28, 29, 30, 31 mova [dstq+0*mmsize], m4 mova [dstq+1*mmsize], m2 mova [dstq+2*mmsize], m0 mova [dstq+3*mmsize], m5 mova [dstq+4*mmsize], m3 mova [dstq+5*mmsize], m1 add src0q, mmsize add dstq, mmsize*6 sub lend, mmsize/2 %endif jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16P_TO_S16_6CH INIT_XMM sse2slow CONV_S16P_TO_S16_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16P_TO_S16_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16p_to_flt_2ch(float *dst, int16_t *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16P_TO_FLT_2CH 0 cglobal conv_s16p_to_flt_2ch, 3,4,6, dst, src0, len, src1 lea lenq, [2*lend] mov src1q, [src0q+gprsize] mov src0q, [src0q ] lea dstq, [dstq+4*lenq] add src0q, lenq add src1q, lenq neg lenq mova m5, [pf_s32_inv_scale] .loop: mova m2, [src0q+lenq] ; m2 = 0, 2, 4, 6, 8, 10, 12, 14 mova m4, [src1q+lenq] ; m4 = 1, 3, 5, 7, 9, 11, 13, 15 SBUTTERFLY2 wd, 2, 4, 3 ; m2 = 0, 1, 2, 3, 4, 5, 6, 7 ; m4 = 8, 9, 10, 11, 12, 13, 14, 15 pxor m3, m3 punpcklwd m0, m3, m2 ; m0 = 0, 1, 2, 3 punpckhwd m1, m3, m2 ; m1 = 4, 5, 6, 7 punpcklwd m2, m3, m4 ; m2 = 8, 9, 10, 11 punpckhwd m3, m4 ; m3 = 12, 13, 14, 15 cvtdq2ps m0, m0 cvtdq2ps m1, m1 cvtdq2ps m2, m2 cvtdq2ps m3, m3 mulps m0, m5 mulps m1, m5 mulps m2, m5 mulps m3, m5 mova [dstq+4*lenq ], m0 mova [dstq+4*lenq+ mmsize], m1 mova [dstq+4*lenq+2*mmsize], m2 mova [dstq+4*lenq+3*mmsize], m3 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16P_TO_FLT_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16P_TO_FLT_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16p_to_flt_6ch(float *dst, int16_t *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16P_TO_FLT_6CH 0 %if ARCH_X86_64 cglobal conv_s16p_to_flt_6ch, 3,8,8, dst, src, len, src1, src2, src3, src4, src5 %else cglobal conv_s16p_to_flt_6ch, 2,7,8, dst, src, src1, src2, src3, src4, src5 %define lend dword r2m %endif mov src1q, [srcq+1*gprsize] mov src2q, [srcq+2*gprsize] mov src3q, [srcq+3*gprsize] mov src4q, [srcq+4*gprsize] mov src5q, [srcq+5*gprsize] mov srcq, [srcq] sub src1q, srcq sub src2q, srcq sub src3q, srcq sub src4q, srcq sub src5q, srcq mova m7, [pf_s32_inv_scale] %if cpuflag(ssse3) %define unpack_even m6 mova m6, [pb_shuf_unpack_even] %if ARCH_X86_64 %define unpack_odd m8 mova m8, [pb_shuf_unpack_odd] %else %define unpack_odd [pb_shuf_unpack_odd] %endif %endif .loop: movq m0, [srcq ] ; m0 = 0, 6, 12, 18, x, x, x, x movq m1, [srcq+src1q] ; m1 = 1, 7, 13, 19, x, x, x, x movq m2, [srcq+src2q] ; m2 = 2, 8, 14, 20, x, x, x, x movq m3, [srcq+src3q] ; m3 = 3, 9, 15, 21, x, x, x, x movq m4, [srcq+src4q] ; m4 = 4, 10, 16, 22, x, x, x, x movq m5, [srcq+src5q] ; m5 = 5, 11, 17, 23, x, x, x, x ; unpack words: punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19 punpcklwd m2, m3 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21 punpcklwd m4, m5 ; m4 = 4, 5, 10, 11, 16, 17, 22, 23 ; blend dwords shufps m1, m4, m0, q3120 ; m1 = 4, 5, 16, 17, 6, 7, 18, 19 shufps m0, m2, q2020 ; m0 = 0, 1, 12, 13, 2, 3, 14, 15 shufps m2, m4, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23 %if cpuflag(ssse3) pshufb m3, m0, unpack_odd ; m3 = 12, 13, 14, 15 pshufb m0, unpack_even ; m0 = 0, 1, 2, 3 pshufb m4, m1, unpack_odd ; m4 = 16, 17, 18, 19 pshufb m1, unpack_even ; m1 = 4, 5, 6, 7 pshufb m5, m2, unpack_odd ; m5 = 20, 21, 22, 23 pshufb m2, unpack_even ; m2 = 8, 9, 10, 11 %else ; shuffle dwords pshufd m0, m0, q3120 ; m0 = 0, 1, 2, 3, 12, 13, 14, 15 pshufd m1, m1, q3120 ; m1 = 4, 5, 6, 7, 16, 17, 18, 19 pshufd m2, m2, q3120 ; m2 = 8, 9, 10, 11, 20, 21, 22, 23 pxor m6, m6 ; convert s16 in m0-m2 to s32 in m0-m5 punpcklwd m3, m6, m0 ; m3 = 0, 1, 2, 3 punpckhwd m4, m6, m0 ; m4 = 12, 13, 14, 15 punpcklwd m0, m6, m1 ; m0 = 4, 5, 6, 7 punpckhwd m5, m6, m1 ; m5 = 16, 17, 18, 19 punpcklwd m1, m6, m2 ; m1 = 8, 9, 10, 11 punpckhwd m6, m2 ; m6 = 20, 21, 22, 23 SWAP 6,2,1,0,3,4,5 ; swap registers 3,0,1,4,5,6 to 0,1,2,3,4,5 %endif cvtdq2ps m0, m0 ; convert s32 to float cvtdq2ps m1, m1 cvtdq2ps m2, m2 cvtdq2ps m3, m3 cvtdq2ps m4, m4 cvtdq2ps m5, m5 mulps m0, m7 ; scale float from s32 range to [-1.0,1.0] mulps m1, m7 mulps m2, m7 mulps m3, m7 mulps m4, m7 mulps m5, m7 mova [dstq ], m0 mova [dstq+ mmsize], m1 mova [dstq+2*mmsize], m2 mova [dstq+3*mmsize], m3 mova [dstq+4*mmsize], m4 mova [dstq+5*mmsize], m5 add srcq, mmsize/2 add dstq, mmsize*6 sub lend, mmsize/4 jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16P_TO_FLT_6CH INIT_XMM ssse3 CONV_S16P_TO_FLT_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16P_TO_FLT_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_fltp_to_s16_2ch(int16_t *dst, float *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLTP_TO_S16_2CH 0 cglobal conv_fltp_to_s16_2ch, 3,4,3, dst, src0, len, src1 lea lenq, [4*lend] mov src1q, [src0q+gprsize] mov src0q, [src0q ] add dstq, lenq add src0q, lenq add src1q, lenq neg lenq mova m2, [pf_s16_scale] %if cpuflag(ssse3) mova m3, [pb_interleave_words] %endif .loop: mulps m0, m2, [src0q+lenq] ; m0 = 0, 2, 4, 6 mulps m1, m2, [src1q+lenq] ; m1 = 1, 3, 5, 7 cvtps2dq m0, m0 cvtps2dq m1, m1 %if cpuflag(ssse3) packssdw m0, m1 ; m0 = 0, 2, 4, 6, 1, 3, 5, 7 pshufb m0, m3 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 %else packssdw m0, m0 ; m0 = 0, 2, 4, 6, x, x, x, x packssdw m1, m1 ; m1 = 1, 3, 5, 7, x, x, x, x punpcklwd m0, m1 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 %endif mova [dstq+lenq], m0 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLTP_TO_S16_2CH INIT_XMM ssse3 CONV_FLTP_TO_S16_2CH ;------------------------------------------------------------------------------ ; void ff_conv_fltp_to_s16_6ch(int16_t *dst, float *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLTP_TO_S16_6CH 0 %if ARCH_X86_64 cglobal conv_fltp_to_s16_6ch, 3,8,7, dst, src, len, src1, src2, src3, src4, src5 %else cglobal conv_fltp_to_s16_6ch, 2,7,7, dst, src, src1, src2, src3, src4, src5 %define lend dword r2m %endif mov src1q, [srcq+1*gprsize] mov src2q, [srcq+2*gprsize] mov src3q, [srcq+3*gprsize] mov src4q, [srcq+4*gprsize] mov src5q, [srcq+5*gprsize] mov srcq, [srcq] sub src1q, srcq sub src2q, srcq sub src3q, srcq sub src4q, srcq sub src5q, srcq movaps xmm6, [pf_s16_scale] .loop: %if cpuflag(sse2) mulps m0, m6, [srcq ] mulps m1, m6, [srcq+src1q] mulps m2, m6, [srcq+src2q] mulps m3, m6, [srcq+src3q] mulps m4, m6, [srcq+src4q] mulps m5, m6, [srcq+src5q] cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 cvtps2dq m4, m4 cvtps2dq m5, m5 packssdw m0, m3 ; m0 = 0, 6, 12, 18, 3, 9, 15, 21 packssdw m1, m4 ; m1 = 1, 7, 13, 19, 4, 10, 16, 22 packssdw m2, m5 ; m2 = 2, 8, 14, 20, 5, 11, 17, 23 ; unpack words: movhlps m3, m0 ; m3 = 3, 9, 15, 21, x, x, x, x punpcklwd m0, m1 ; m0 = 0, 1, 6, 7, 12, 13, 18, 19 punpckhwd m1, m2 ; m1 = 4, 5, 10, 11, 16, 17, 22, 23 punpcklwd m2, m3 ; m2 = 2, 3, 8, 9, 14, 15, 20, 21 ; blend dwords: shufps m3, m0, m2, q2020 ; m3 = 0, 1, 12, 13, 2, 3, 14, 15 shufps m0, m1, q2031 ; m0 = 6, 7, 18, 19, 4, 5, 16, 17 shufps m2, m1, q3131 ; m2 = 8, 9, 20, 21, 10, 11, 22, 23 ; shuffle dwords: shufps m1, m2, m3, q3120 ; m1 = 8, 9, 10, 11, 12, 13, 14, 15 shufps m3, m0, q0220 ; m3 = 0, 1, 2, 3, 4, 5, 6, 7 shufps m0, m2, q3113 ; m0 = 16, 17, 18, 19, 20, 21, 22, 23 mova [dstq+0*mmsize], m3 mova [dstq+1*mmsize], m1 mova [dstq+2*mmsize], m0 %else ; sse movlps xmm0, [srcq ] movlps xmm1, [srcq+src1q] movlps xmm2, [srcq+src2q] movlps xmm3, [srcq+src3q] movlps xmm4, [srcq+src4q] movlps xmm5, [srcq+src5q] mulps xmm0, xmm6 mulps xmm1, xmm6 mulps xmm2, xmm6 mulps xmm3, xmm6 mulps xmm4, xmm6 mulps xmm5, xmm6 cvtps2pi mm0, xmm0 cvtps2pi mm1, xmm1 cvtps2pi mm2, xmm2 cvtps2pi mm3, xmm3 cvtps2pi mm4, xmm4 cvtps2pi mm5, xmm5 packssdw mm0, mm3 ; m0 = 0, 6, 3, 9 packssdw mm1, mm4 ; m1 = 1, 7, 4, 10 packssdw mm2, mm5 ; m2 = 2, 8, 5, 11 ; unpack words pshufw mm3, mm0, q1032 ; m3 = 3, 9, 0, 6 punpcklwd mm0, mm1 ; m0 = 0, 1, 6, 7 punpckhwd mm1, mm2 ; m1 = 4, 5, 10, 11 punpcklwd mm2, mm3 ; m2 = 2, 3, 8, 9 ; unpack dwords pshufw mm3, mm0, q1032 ; m3 = 6, 7, 0, 1 punpckldq mm0, mm2 ; m0 = 0, 1, 2, 3 (final) punpckhdq mm2, mm1 ; m2 = 8, 9, 10, 11 (final) punpckldq mm1, mm3 ; m1 = 4, 5, 6, 7 (final) mova [dstq+0*mmsize], mm0 mova [dstq+1*mmsize], mm1 mova [dstq+2*mmsize], mm2 %endif add srcq, mmsize add dstq, mmsize*3 sub lend, mmsize/4 jg .loop %if mmsize == 8 emms RET %else REP_RET %endif %endmacro INIT_MMX sse CONV_FLTP_TO_S16_6CH INIT_XMM sse2 CONV_FLTP_TO_S16_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLTP_TO_S16_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_fltp_to_flt_2ch(float *dst, float *const *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLTP_TO_FLT_2CH 0 cglobal conv_fltp_to_flt_2ch, 3,4,5, dst, src0, len, src1 mov src1q, [src0q+gprsize] mov src0q, [src0q] lea lenq, [4*lend] add src0q, lenq add src1q, lenq lea dstq, [dstq+2*lenq] neg lenq .loop: mova m0, [src0q+lenq ] mova m1, [src1q+lenq ] mova m2, [src0q+lenq+mmsize] mova m3, [src1q+lenq+mmsize] SBUTTERFLYPS 0, 1, 4 SBUTTERFLYPS 2, 3, 4 mova [dstq+2*lenq+0*mmsize], m0 mova [dstq+2*lenq+1*mmsize], m1 mova [dstq+2*lenq+2*mmsize], m2 mova [dstq+2*lenq+3*mmsize], m3 add lenq, 2*mmsize jl .loop REP_RET %endmacro INIT_XMM sse CONV_FLTP_TO_FLT_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLTP_TO_FLT_2CH %endif ;----------------------------------------------------------------------------- ; void ff_conv_fltp_to_flt_6ch(float *dst, float *const *src, int len, ; int channels); ;----------------------------------------------------------------------------- %macro CONV_FLTP_TO_FLT_6CH 0 cglobal conv_fltp_to_flt_6ch, 2,8,7, dst, src, src1, src2, src3, src4, src5, len %if ARCH_X86_64 mov lend, r2d %else %define lend dword r2m %endif mov src1q, [srcq+1*gprsize] mov src2q, [srcq+2*gprsize] mov src3q, [srcq+3*gprsize] mov src4q, [srcq+4*gprsize] mov src5q, [srcq+5*gprsize] mov srcq, [srcq] sub src1q, srcq sub src2q, srcq sub src3q, srcq sub src4q, srcq sub src5q, srcq .loop: mova m0, [srcq ] mova m1, [srcq+src1q] mova m2, [srcq+src2q] mova m3, [srcq+src3q] mova m4, [srcq+src4q] mova m5, [srcq+src5q] %if cpuflag(sse4) SBUTTERFLYPS 0, 1, 6 SBUTTERFLYPS 2, 3, 6 SBUTTERFLYPS 4, 5, 6 blendps m6, m4, m0, 1100b movlhps m0, m2 movhlps m4, m2 blendps m2, m5, m1, 1100b movlhps m1, m3 movhlps m5, m3 movaps [dstq ], m0 movaps [dstq+16], m6 movaps [dstq+32], m4 movaps [dstq+48], m1 movaps [dstq+64], m2 movaps [dstq+80], m5 %else ; mmx SBUTTERFLY dq, 0, 1, 6 SBUTTERFLY dq, 2, 3, 6 SBUTTERFLY dq, 4, 5, 6 movq [dstq ], m0 movq [dstq+ 8], m2 movq [dstq+16], m4 movq [dstq+24], m1 movq [dstq+32], m3 movq [dstq+40], m5 %endif add srcq, mmsize add dstq, mmsize*6 sub lend, mmsize/4 jg .loop %if mmsize == 8 emms RET %else REP_RET %endif %endmacro INIT_MMX mmx CONV_FLTP_TO_FLT_6CH INIT_XMM sse4 CONV_FLTP_TO_FLT_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLTP_TO_FLT_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_s16p_2ch(int16_t *const *dst, int16_t *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_S16P_2CH 0 cglobal conv_s16_to_s16p_2ch, 3,4,4, dst0, src, len, dst1 lea lenq, [2*lend] mov dst1q, [dst0q+gprsize] mov dst0q, [dst0q ] lea srcq, [srcq+2*lenq] add dst0q, lenq add dst1q, lenq neg lenq %if cpuflag(ssse3) mova m3, [pb_deinterleave_words] %endif .loop: mova m0, [srcq+2*lenq ] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 mova m1, [srcq+2*lenq+mmsize] ; m1 = 8, 9, 10, 11, 12, 13, 14, 15 %if cpuflag(ssse3) pshufb m0, m3 ; m0 = 0, 2, 4, 6, 1, 3, 5, 7 pshufb m1, m3 ; m1 = 8, 10, 12, 14, 9, 11, 13, 15 SBUTTERFLY2 qdq, 0, 1, 2 ; m0 = 0, 2, 4, 6, 8, 10, 12, 14 ; m1 = 1, 3, 5, 7, 9, 11, 13, 15 %else ; sse2 pshuflw m0, m0, q3120 ; m0 = 0, 2, 1, 3, 4, 5, 6, 7 pshufhw m0, m0, q3120 ; m0 = 0, 2, 1, 3, 4, 6, 5, 7 pshuflw m1, m1, q3120 ; m1 = 8, 10, 9, 11, 12, 13, 14, 15 pshufhw m1, m1, q3120 ; m1 = 8, 10, 9, 11, 12, 14, 13, 15 DEINT2_PS 0, 1, 2 ; m0 = 0, 2, 4, 6, 8, 10, 12, 14 ; m1 = 1, 3, 5, 7, 9, 11, 13, 15 %endif mova [dst0q+lenq], m0 mova [dst1q+lenq], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_S16P_2CH INIT_XMM ssse3 CONV_S16_TO_S16P_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16_TO_S16P_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_s16p_6ch(int16_t *const *dst, int16_t *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_S16P_6CH 0 %if ARCH_X86_64 cglobal conv_s16_to_s16p_6ch, 3,8,5, dst, src, len, dst1, dst2, dst3, dst4, dst5 %else cglobal conv_s16_to_s16p_6ch, 2,7,5, dst, src, dst1, dst2, dst3, dst4, dst5 %define lend dword r2m %endif mov dst1q, [dstq+ gprsize] mov dst2q, [dstq+2*gprsize] mov dst3q, [dstq+3*gprsize] mov dst4q, [dstq+4*gprsize] mov dst5q, [dstq+5*gprsize] mov dstq, [dstq ] sub dst1q, dstq sub dst2q, dstq sub dst3q, dstq sub dst4q, dstq sub dst5q, dstq .loop: mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 mova m3, [srcq+1*mmsize] ; m3 = 8, 9, 10, 11, 12, 13, 14, 15 mova m2, [srcq+2*mmsize] ; m2 = 16, 17, 18, 19, 20, 21, 22, 23 PALIGNR m1, m3, m0, 12, m4 ; m1 = 6, 7, 8, 9, 10, 11, x, x shufps m3, m2, q1032 ; m3 = 12, 13, 14, 15, 16, 17, 18, 19 psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x SBUTTERFLY2 wd, 0, 1, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9 ; m1 = 4, 10, 5, 11, x, x, x, x SBUTTERFLY2 wd, 3, 2, 4 ; m3 = 12, 18, 13, 19, 14, 20, 15, 21 ; m2 = 16, 22, 17, 23, x, x, x, x SBUTTERFLY2 dq, 0, 3, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19 ; m3 = 2, 8, 14, 20, 3, 9, 15, 21 punpckldq m1, m2 ; m1 = 4, 10, 16, 22, 5, 11, 17, 23 movq [dstq ], m0 movhps [dstq+dst1q], m0 movq [dstq+dst2q], m3 movhps [dstq+dst3q], m3 movq [dstq+dst4q], m1 movhps [dstq+dst5q], m1 add srcq, mmsize*3 add dstq, mmsize/2 sub lend, mmsize/4 jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_S16P_6CH INIT_XMM ssse3 CONV_S16_TO_S16P_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16_TO_S16P_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_fltp_2ch(float *const *dst, int16_t *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_FLTP_2CH 0 cglobal conv_s16_to_fltp_2ch, 3,4,5, dst0, src, len, dst1 lea lenq, [4*lend] mov dst1q, [dst0q+gprsize] mov dst0q, [dst0q ] add srcq, lenq add dst0q, lenq add dst1q, lenq neg lenq mova m3, [pf_s32_inv_scale] mova m4, [pw_zero_even] .loop: mova m1, [srcq+lenq] pslld m0, m1, 16 pand m1, m4 cvtdq2ps m0, m0 cvtdq2ps m1, m1 mulps m0, m0, m3 mulps m1, m1, m3 mova [dst0q+lenq], m0 mova [dst1q+lenq], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_FLTP_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16_TO_FLTP_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_s16_to_fltp_6ch(float *const *dst, int16_t *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_S16_TO_FLTP_6CH 0 %if ARCH_X86_64 cglobal conv_s16_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5 %else cglobal conv_s16_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5 %define lend dword r2m %endif mov dst1q, [dstq+ gprsize] mov dst2q, [dstq+2*gprsize] mov dst3q, [dstq+3*gprsize] mov dst4q, [dstq+4*gprsize] mov dst5q, [dstq+5*gprsize] mov dstq, [dstq ] sub dst1q, dstq sub dst2q, dstq sub dst3q, dstq sub dst4q, dstq sub dst5q, dstq mova m6, [pf_s16_inv_scale] .loop: mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 mova m3, [srcq+1*mmsize] ; m3 = 8, 9, 10, 11, 12, 13, 14, 15 mova m2, [srcq+2*mmsize] ; m2 = 16, 17, 18, 19, 20, 21, 22, 23 PALIGNR m1, m3, m0, 12, m4 ; m1 = 6, 7, 8, 9, 10, 11, x, x shufps m3, m2, q1032 ; m3 = 12, 13, 14, 15, 16, 17, 18, 19 psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x SBUTTERFLY2 wd, 0, 1, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9 ; m1 = 4, 10, 5, 11, x, x, x, x SBUTTERFLY2 wd, 3, 2, 4 ; m3 = 12, 18, 13, 19, 14, 20, 15, 21 ; m2 = 16, 22, 17, 23, x, x, x, x SBUTTERFLY2 dq, 0, 3, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19 ; m3 = 2, 8, 14, 20, 3, 9, 15, 21 punpckldq m1, m2 ; m1 = 4, 10, 16, 22, 5, 11, 17, 23 S16_TO_S32_SX 0, 2 ; m0 = 0, 6, 12, 18 ; m2 = 1, 7, 13, 19 S16_TO_S32_SX 3, 4 ; m3 = 2, 8, 14, 20 ; m4 = 3, 9, 15, 21 S16_TO_S32_SX 1, 5 ; m1 = 4, 10, 16, 22 ; m5 = 5, 11, 17, 23 SWAP 1,2,3,4 cvtdq2ps m0, m0 cvtdq2ps m1, m1 cvtdq2ps m2, m2 cvtdq2ps m3, m3 cvtdq2ps m4, m4 cvtdq2ps m5, m5 mulps m0, m6 mulps m1, m6 mulps m2, m6 mulps m3, m6 mulps m4, m6 mulps m5, m6 mova [dstq ], m0 mova [dstq+dst1q], m1 mova [dstq+dst2q], m2 mova [dstq+dst3q], m3 mova [dstq+dst4q], m4 mova [dstq+dst5q], m5 add srcq, mmsize*3 add dstq, mmsize sub lend, mmsize/4 jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_S16_TO_FLTP_6CH INIT_XMM ssse3 CONV_S16_TO_FLTP_6CH INIT_XMM sse4 CONV_S16_TO_FLTP_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_S16_TO_FLTP_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s16p_2ch(int16_t *const *dst, float *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_S16P_2CH 0 cglobal conv_flt_to_s16p_2ch, 3,4,6, dst0, src, len, dst1 lea lenq, [2*lend] mov dst1q, [dst0q+gprsize] mov dst0q, [dst0q ] lea srcq, [srcq+4*lenq] add dst0q, lenq add dst1q, lenq neg lenq mova m5, [pf_s16_scale] .loop: mova m0, [srcq+4*lenq ] mova m1, [srcq+4*lenq+ mmsize] mova m2, [srcq+4*lenq+2*mmsize] mova m3, [srcq+4*lenq+3*mmsize] DEINT2_PS 0, 1, 4 DEINT2_PS 2, 3, 4 mulps m0, m0, m5 mulps m1, m1, m5 mulps m2, m2, m5 mulps m3, m3, m5 cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 packssdw m0, m2 packssdw m1, m3 mova [dst0q+lenq], m0 mova [dst1q+lenq], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLT_TO_S16P_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLT_TO_S16P_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_s16p_6ch(int16_t *const *dst, float *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_S16P_6CH 0 %if ARCH_X86_64 cglobal conv_flt_to_s16p_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5 %else cglobal conv_flt_to_s16p_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5 %define lend dword r2m %endif mov dst1q, [dstq+ gprsize] mov dst2q, [dstq+2*gprsize] mov dst3q, [dstq+3*gprsize] mov dst4q, [dstq+4*gprsize] mov dst5q, [dstq+5*gprsize] mov dstq, [dstq ] sub dst1q, dstq sub dst2q, dstq sub dst3q, dstq sub dst4q, dstq sub dst5q, dstq mova m6, [pf_s16_scale] .loop: mulps m0, m6, [srcq+0*mmsize] mulps m3, m6, [srcq+1*mmsize] mulps m1, m6, [srcq+2*mmsize] mulps m4, m6, [srcq+3*mmsize] mulps m2, m6, [srcq+4*mmsize] mulps m5, m6, [srcq+5*mmsize] cvtps2dq m0, m0 cvtps2dq m1, m1 cvtps2dq m2, m2 cvtps2dq m3, m3 cvtps2dq m4, m4 cvtps2dq m5, m5 packssdw m0, m3 ; m0 = 0, 1, 2, 3, 4, 5, 6, 7 packssdw m1, m4 ; m1 = 8, 9, 10, 11, 12, 13, 14, 15 packssdw m2, m5 ; m2 = 16, 17, 18, 19, 20, 21, 22, 23 PALIGNR m3, m1, m0, 12, m4 ; m3 = 6, 7, 8, 9, 10, 11, x, x shufps m1, m2, q1032 ; m1 = 12, 13, 14, 15, 16, 17, 18, 19 psrldq m2, 4 ; m2 = 18, 19, 20, 21, 22, 23, x, x SBUTTERFLY2 wd, 0, 3, 4 ; m0 = 0, 6, 1, 7, 2, 8, 3, 9 ; m3 = 4, 10, 5, 11, x, x, x, x SBUTTERFLY2 wd, 1, 2, 4 ; m1 = 12, 18, 13, 19, 14, 20, 15, 21 ; m2 = 16, 22, 17, 23, x, x, x, x SBUTTERFLY2 dq, 0, 1, 4 ; m0 = 0, 6, 12, 18, 1, 7, 13, 19 ; m1 = 2, 8, 14, 20, 3, 9, 15, 21 punpckldq m3, m2 ; m3 = 4, 10, 16, 22, 5, 11, 17, 23 movq [dstq ], m0 movhps [dstq+dst1q], m0 movq [dstq+dst2q], m1 movhps [dstq+dst3q], m1 movq [dstq+dst4q], m3 movhps [dstq+dst5q], m3 add srcq, mmsize*6 add dstq, mmsize/2 sub lend, mmsize/4 jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLT_TO_S16P_6CH INIT_XMM ssse3 CONV_FLT_TO_S16P_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLT_TO_S16P_6CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_fltp_2ch(float *const *dst, float *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_FLTP_2CH 0 cglobal conv_flt_to_fltp_2ch, 3,4,3, dst0, src, len, dst1 lea lenq, [4*lend] mov dst1q, [dst0q+gprsize] mov dst0q, [dst0q ] lea srcq, [srcq+2*lenq] add dst0q, lenq add dst1q, lenq neg lenq .loop: mova m0, [srcq+2*lenq ] mova m1, [srcq+2*lenq+mmsize] DEINT2_PS 0, 1, 2 mova [dst0q+lenq], m0 mova [dst1q+lenq], m1 add lenq, mmsize jl .loop REP_RET %endmacro INIT_XMM sse CONV_FLT_TO_FLTP_2CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLT_TO_FLTP_2CH %endif ;------------------------------------------------------------------------------ ; void ff_conv_flt_to_fltp_6ch(float *const *dst, float *src, int len, ; int channels); ;------------------------------------------------------------------------------ %macro CONV_FLT_TO_FLTP_6CH 0 %if ARCH_X86_64 cglobal conv_flt_to_fltp_6ch, 3,8,7, dst, src, len, dst1, dst2, dst3, dst4, dst5 %else cglobal conv_flt_to_fltp_6ch, 2,7,7, dst, src, dst1, dst2, dst3, dst4, dst5 %define lend dword r2m %endif mov dst1q, [dstq+ gprsize] mov dst2q, [dstq+2*gprsize] mov dst3q, [dstq+3*gprsize] mov dst4q, [dstq+4*gprsize] mov dst5q, [dstq+5*gprsize] mov dstq, [dstq ] sub dst1q, dstq sub dst2q, dstq sub dst3q, dstq sub dst4q, dstq sub dst5q, dstq .loop: mova m0, [srcq+0*mmsize] ; m0 = 0, 1, 2, 3 mova m1, [srcq+1*mmsize] ; m1 = 4, 5, 6, 7 mova m2, [srcq+2*mmsize] ; m2 = 8, 9, 10, 11 mova m3, [srcq+3*mmsize] ; m3 = 12, 13, 14, 15 mova m4, [srcq+4*mmsize] ; m4 = 16, 17, 18, 19 mova m5, [srcq+5*mmsize] ; m5 = 20, 21, 22, 23 SBUTTERFLY2 dq, 0, 3, 6 ; m0 = 0, 12, 1, 13 ; m3 = 2, 14, 3, 15 SBUTTERFLY2 dq, 1, 4, 6 ; m1 = 4, 16, 5, 17 ; m4 = 6, 18, 7, 19 SBUTTERFLY2 dq, 2, 5, 6 ; m2 = 8, 20, 9, 21 ; m5 = 10, 22, 11, 23 SBUTTERFLY2 dq, 0, 4, 6 ; m0 = 0, 6, 12, 18 ; m4 = 1, 7, 13, 19 SBUTTERFLY2 dq, 3, 2, 6 ; m3 = 2, 8, 14, 20 ; m2 = 3, 9, 15, 21 SBUTTERFLY2 dq, 1, 5, 6 ; m1 = 4, 10, 16, 22 ; m5 = 5, 11, 17, 23 mova [dstq ], m0 mova [dstq+dst1q], m4 mova [dstq+dst2q], m3 mova [dstq+dst3q], m2 mova [dstq+dst4q], m1 mova [dstq+dst5q], m5 add srcq, mmsize*6 add dstq, mmsize sub lend, mmsize/4 jg .loop REP_RET %endmacro INIT_XMM sse2 CONV_FLT_TO_FLTP_6CH %if HAVE_AVX_EXTERNAL INIT_XMM avx CONV_FLT_TO_FLTP_6CH %endif