;****************************************************************************** ;* SIMD-optimized fullpel functions ;* Copyright (c) 2008 Loren Merritt ;* Copyright (c) 2003-2013 Michael Niedermayer ;* Copyright (c) 2013 Daniel Kang ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION .text %macro PAVGB_MMX 4 LOAD %3, %1 por %3, %2 pxor %2, %1 pand %2, %4 psrlq %2, 1 psubb %3, %2 SWAP %2, %3 %endmacro ; void ff_put/avg_pixels(uint8_t *block, const uint8_t *pixels, ; ptrdiff_t line_size, int h) %macro OP_PIXELS 2 %if %2 == mmsize/2 %define LOAD movh %define SAVE movh %define LEN mmsize %else %define LOAD movu %define SAVE mova %define LEN %2 %endif cglobal %1_pixels%2, 4,5,4 movsxdifnidn r2, r2d lea r4, [r2*3] %ifidn %1, avg %if notcpuflag(mmxext) pcmpeqd m6, m6 paddb m6, m6 %endif %endif .loop: %assign %%i 0 %rep LEN/mmsize LOAD m0, [r1 + %%i] LOAD m1, [r1+r2 + %%i] LOAD m2, [r1+r2*2 + %%i] LOAD m3, [r1+r4 + %%i] %ifidn %1, avg %if notcpuflag(mmxext) PAVGB_MMX [r0 + %%i], m0, m4, m6 PAVGB_MMX [r0+r2 + %%i], m1, m5, m6 PAVGB_MMX [r0+r2*2 + %%i], m2, m4, m6 PAVGB_MMX [r0+r4 + %%i], m3, m5, m6 %else pavgb m0, [r0 + %%i] pavgb m1, [r0+r2 + %%i] pavgb m2, [r0+r2*2 + %%i] pavgb m3, [r0+r4 + %%i] %endif %endif SAVE [r0 + %%i], m0 SAVE [r0+r2 + %%i], m1 SAVE [r0+r2*2 + %%i], m2 SAVE [r0+r4 + %%i], m3 %assign %%i %%i+mmsize %endrep sub r3d, 4 lea r1, [r1+r2*4] lea r0, [r0+r2*4] jne .loop RET %endmacro INIT_MMX mmx OP_PIXELS put, 4 OP_PIXELS avg, 4 OP_PIXELS put, 8 OP_PIXELS avg, 8 OP_PIXELS put, 16 OP_PIXELS avg, 16 INIT_MMX mmxext OP_PIXELS avg, 4 OP_PIXELS avg, 8 OP_PIXELS avg, 16 INIT_XMM sse2 OP_PIXELS put, 16 OP_PIXELS avg, 16