/* * Copyright (c) 2009 Mans Rullgard * * This file is part of FFmpeg. * * FFmpeg is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with FFmpeg; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include "libavutil/arm/asm.S" function ff_pix_abs16_armv6, export=1 ldr r0, [sp] push {r4-r9, lr} mov r12, #0 mov lr, #0 ldm r1, {r4-r7} ldr r8, [r2] 1: ldr r9, [r2, #4] pld [r1, r3] usada8 r12, r4, r8, r12 ldr r8, [r2, #8] pld [r2, r3] usada8 lr, r5, r9, lr ldr r9, [r2, #12] usada8 r12, r6, r8, r12 subs r0, r0, #1 usada8 lr, r7, r9, lr beq 2f add r1, r1, r3 ldm r1, {r4-r7} add r2, r2, r3 ldr r8, [r2] b 1b 2: add r0, r12, lr pop {r4-r9, pc} endfunc function ff_pix_abs16_x2_armv6, export=1 ldr r12, [sp] push {r4-r11, lr} mov r0, #0 mov lr, #1 orr lr, lr, lr, lsl #8 orr lr, lr, lr, lsl #16 1: ldr r8, [r2] ldr r9, [r2, #4] lsr r10, r8, #8 ldr r4, [r1] lsr r6, r9, #8 orr r10, r10, r9, lsl #24 ldr r5, [r2, #8] eor r11, r8, r10 uhadd8 r7, r8, r10 orr r6, r6, r5, lsl #24 and r11, r11, lr uadd8 r7, r7, r11 ldr r8, [r1, #4] usada8 r0, r4, r7, r0 eor r7, r9, r6 lsr r10, r5, #8 and r7, r7, lr uhadd8 r4, r9, r6 ldr r6, [r2, #12] uadd8 r4, r4, r7 pld [r1, r3] orr r10, r10, r6, lsl #24 usada8 r0, r8, r4, r0 ldr r4, [r1, #8] eor r11, r5, r10 ldrb r7, [r2, #16] and r11, r11, lr uhadd8 r8, r5, r10 ldr r5, [r1, #12] uadd8 r8, r8, r11 pld [r2, r3] lsr r10, r6, #8 usada8 r0, r4, r8, r0 orr r10, r10, r7, lsl #24 subs r12, r12, #1 eor r11, r6, r10 add r1, r1, r3 uhadd8 r9, r6, r10 and r11, r11, lr uadd8 r9, r9, r11 add r2, r2, r3 usada8 r0, r5, r9, r0 bgt 1b pop {r4-r11, pc} endfunc .macro usad_y2 p0, p1, p2, p3, n0, n1, n2, n3 ldr \n0, [r2] eor \n1, \p0, \n0 uhadd8 \p0, \p0, \n0 and \n1, \n1, lr ldr \n2, [r1] uadd8 \p0, \p0, \n1 ldr \n1, [r2, #4] usada8 r0, \p0, \n2, r0 pld [r1, r3] eor \n3, \p1, \n1 uhadd8 \p1, \p1, \n1 and \n3, \n3, lr ldr \p0, [r1, #4] uadd8 \p1, \p1, \n3 ldr \n2, [r2, #8] usada8 r0, \p1, \p0, r0 pld [r2, r3] eor \p0, \p2, \n2 uhadd8 \p2, \p2, \n2 and \p0, \p0, lr ldr \p1, [r1, #8] uadd8 \p2, \p2, \p0 ldr \n3, [r2, #12] usada8 r0, \p2, \p1, r0 eor \p1, \p3, \n3 uhadd8 \p3, \p3, \n3 and \p1, \p1, lr ldr \p0, [r1, #12] uadd8 \p3, \p3, \p1 add r1, r1, r3 usada8 r0, \p3, \p0, r0 add r2, r2, r3 .endm function ff_pix_abs16_y2_armv6, export=1 pld [r1] pld [r2] ldr r12, [sp] push {r4-r11, lr} mov r0, #0 mov lr, #1 orr lr, lr, lr, lsl #8 orr lr, lr, lr, lsl #16 ldr r4, [r2] ldr r5, [r2, #4] ldr r6, [r2, #8] ldr r7, [r2, #12] add r2, r2, r3 1: usad_y2 r4, r5, r6, r7, r8, r9, r10, r11 subs r12, r12, #2 usad_y2 r8, r9, r10, r11, r4, r5, r6, r7 bgt 1b pop {r4-r11, pc} endfunc function ff_pix_abs8_armv6, export=1 pld [r2, r3] ldr r12, [sp] push {r4-r9, lr} mov r0, #0 mov lr, #0 ldrd_post r4, r5, r1, r3 1: subs r12, r12, #2 ldr r7, [r2, #4] ldr_post r6, r2, r3 ldrd_post r8, r9, r1, r3 usada8 r0, r4, r6, r0 pld [r2, r3] usada8 lr, r5, r7, lr ldr r7, [r2, #4] ldr_post r6, r2, r3 beq 2f ldrd_post r4, r5, r1, r3 usada8 r0, r8, r6, r0 pld [r2, r3] usada8 lr, r9, r7, lr b 1b 2: usada8 r0, r8, r6, r0 usada8 lr, r9, r7, lr add r0, r0, lr pop {r4-r9, pc} endfunc function ff_sse16_armv6, export=1 ldr r12, [sp] push {r4-r9, lr} mov r0, #0 1: ldrd r4, r5, [r1] ldr r8, [r2] uxtb16 lr, r4 uxtb16 r4, r4, ror #8 uxtb16 r9, r8 uxtb16 r8, r8, ror #8 ldr r7, [r2, #4] usub16 lr, lr, r9 usub16 r4, r4, r8 smlad r0, lr, lr, r0 uxtb16 r6, r5 uxtb16 lr, r5, ror #8 uxtb16 r8, r7 uxtb16 r9, r7, ror #8 smlad r0, r4, r4, r0 ldrd r4, r5, [r1, #8] usub16 r6, r6, r8 usub16 r8, lr, r9 ldr r7, [r2, #8] smlad r0, r6, r6, r0 uxtb16 lr, r4 uxtb16 r4, r4, ror #8 uxtb16 r9, r7 uxtb16 r7, r7, ror #8 smlad r0, r8, r8, r0 ldr r8, [r2, #12] usub16 lr, lr, r9 usub16 r4, r4, r7 smlad r0, lr, lr, r0 uxtb16 r6, r5 uxtb16 r5, r5, ror #8 uxtb16 r9, r8 uxtb16 r8, r8, ror #8 smlad r0, r4, r4, r0 usub16 r6, r6, r9 usub16 r5, r5, r8 smlad r0, r6, r6, r0 add r1, r1, r3 add r2, r2, r3 subs r12, r12, #1 smlad r0, r5, r5, r0 bgt 1b pop {r4-r9, pc} endfunc