103793d27a68a7f4d8ff9d6f8b99f76321331803
2 * Copyright (C) 2001-2003 Michael Niedermayer <michaelni@gmx.at>
4 * This file is part of FFmpeg.
6 * FFmpeg is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2.1 of the License, or (at your option) any later version.
11 * FFmpeg is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with FFmpeg; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
21 #include "../swscale_internal.h"
22 #include "libavutil/x86/asm.h"
23 #include "libavutil/x86/cpu.h"
25 #define RET 0xC3 // near return opcode for x86
26 #define PREFETCH "prefetchnta"
29 av_cold
int ff_init_hscaler_mmxext(int dstW
, int xInc
, uint8_t *filterCode
,
30 int16_t *filter
, int32_t *filterPos
,
34 x86_reg imm8OfPShufW1A
;
35 x86_reg imm8OfPShufW2A
;
36 x86_reg fragmentLengthA
;
38 x86_reg imm8OfPShufW1B
;
39 x86_reg imm8OfPShufW2B
;
40 x86_reg fragmentLengthB
;
45 // create an optimized horizontal scaling routine
46 /* This scaler is made of runtime-generated MMXEXT code using specially tuned
47 * pshufw instructions. For every four output pixels, if four input pixels
48 * are enough for the fast bilinear scaling, then a chunk of fragmentB is
49 * used. If five input pixels are needed, then a chunk of fragmentA is used.
58 "movq (%%"REG_d
", %%"REG_a
"), %%mm3 \n\t"
59 "movd (%%"REG_c
", %%"REG_S
"), %%mm0 \n\t"
60 "movd 1(%%"REG_c
", %%"REG_S
"), %%mm1 \n\t"
61 "punpcklbw %%mm7, %%mm1 \n\t"
62 "punpcklbw %%mm7, %%mm0 \n\t"
63 "pshufw $0xFF, %%mm1, %%mm1 \n\t"
65 "pshufw $0xFF, %%mm0, %%mm0 \n\t"
67 "psubw %%mm1, %%mm0 \n\t"
68 "movl 8(%%"REG_b
", %%"REG_a
"), %%esi \n\t"
69 "pmullw %%mm3, %%mm0 \n\t"
70 "psllw $7, %%mm1 \n\t"
71 "paddw %%mm1, %%mm0 \n\t"
73 "movq %%mm0, (%%"REG_D
", %%"REG_a
") \n\t"
75 "add $8, %%"REG_a
" \n\t"
79 "lea " LOCAL_MANGLE(0b
) ", %0 \n\t"
80 "lea " LOCAL_MANGLE(1b
) ", %1 \n\t"
81 "lea " LOCAL_MANGLE(2b
) ", %2 \n\t"
86 "lea " LOCAL_MANGLE(9b
) ", %3 \n\t"
90 : "=r" (fragmentA
), "=r" (imm8OfPShufW1A
), "=r" (imm8OfPShufW2A
),
91 "=r" (fragmentLengthA
)
98 "movq (%%"REG_d
", %%"REG_a
"), %%mm3 \n\t"
99 "movd (%%"REG_c
", %%"REG_S
"), %%mm0 \n\t"
100 "punpcklbw %%mm7, %%mm0 \n\t"
101 "pshufw $0xFF, %%mm0, %%mm1 \n\t"
103 "pshufw $0xFF, %%mm0, %%mm0 \n\t"
105 "psubw %%mm1, %%mm0 \n\t"
106 "movl 8(%%"REG_b
", %%"REG_a
"), %%esi \n\t"
107 "pmullw %%mm3, %%mm0 \n\t"
108 "psllw $7, %%mm1 \n\t"
109 "paddw %%mm1, %%mm0 \n\t"
111 "movq %%mm0, (%%"REG_D
", %%"REG_a
") \n\t"
113 "add $8, %%"REG_a
" \n\t"
117 "lea " LOCAL_MANGLE(0b
) ", %0 \n\t"
118 "lea " LOCAL_MANGLE(1b
) ", %1 \n\t"
119 "lea " LOCAL_MANGLE(2b
) ", %2 \n\t"
124 "lea " LOCAL_MANGLE(9b
) ", %3 \n\t"
128 : "=r" (fragmentB
), "=r" (imm8OfPShufW1B
), "=r" (imm8OfPShufW2B
),
129 "=r" (fragmentLengthB
)
132 xpos
= 0; // lumXInc/2 - 0x8000; // difference between pixel centers
135 for (i
= 0; i
< dstW
/ numSplits
; i
++) {
140 int b
= ((xpos
+ xInc
) >> 16) - xx
;
141 int c
= ((xpos
+ xInc
* 2) >> 16) - xx
;
142 int d
= ((xpos
+ xInc
* 3) >> 16) - xx
;
143 int inc
= (d
+ 1 < 4);
144 uint8_t *fragment
= inc
? fragmentB
: fragmentA
;
145 x86_reg imm8OfPShufW1
= inc
? imm8OfPShufW1B
: imm8OfPShufW1A
;
146 x86_reg imm8OfPShufW2
= inc
? imm8OfPShufW2B
: imm8OfPShufW2A
;
147 x86_reg fragmentLength
= inc
? fragmentLengthB
: fragmentLengthA
;
148 int maxShift
= 3 - (d
+ inc
);
152 filter
[i
] = ((xpos
& 0xFFFF) ^ 0xFFFF) >> 9;
153 filter
[i
+ 1] = (((xpos
+ xInc
) & 0xFFFF) ^ 0xFFFF) >> 9;
154 filter
[i
+ 2] = (((xpos
+ xInc
* 2) & 0xFFFF) ^ 0xFFFF) >> 9;
155 filter
[i
+ 3] = (((xpos
+ xInc
* 3) & 0xFFFF) ^ 0xFFFF) >> 9;
156 filterPos
[i
/ 2] = xx
;
158 memcpy(filterCode
+ fragmentPos
, fragment
, fragmentLength
);
160 filterCode
[fragmentPos
+ imm8OfPShufW1
] = (a
+ inc
) |
164 filterCode
[fragmentPos
+ imm8OfPShufW2
] = a
| (b
<< 2) |
168 if (i
+ 4 - inc
>= dstW
)
169 shift
= maxShift
; // avoid overread
170 else if ((filterPos
[i
/ 2] & 3) <= maxShift
)
171 shift
= filterPos
[i
/ 2] & 3; // align
173 if (shift
&& i
>= shift
) {
174 filterCode
[fragmentPos
+ imm8OfPShufW1
] += 0x55 * shift
;
175 filterCode
[fragmentPos
+ imm8OfPShufW2
] += 0x55 * shift
;
176 filterPos
[i
/ 2] -= shift
;
180 fragmentPos
+= fragmentLength
;
183 filterCode
[fragmentPos
] = RET
;
188 filterPos
[((i
/ 2) + 1) & (~1)] = xpos
>> 16; // needed to jump to the next part
190 return fragmentPos
+ 1;
193 void ff_hyscale_fast_mmxext(SwsContext
*c
, int16_t *dst
,
194 int dstWidth
, const uint8_t *src
,
197 int32_t *filterPos
= c
->hLumFilterPos
;
198 int16_t *filter
= c
->hLumFilter
;
199 void *mmxextFilterCode
= c
->lumMmxextFilterCode
;
210 "mov %%"REG_b
", %5 \n\t"
212 "mov -8(%%rsp), %%"REG_a
" \n\t"
213 "mov %%"REG_a
", %6 \n\t"
217 "mov -8(%%rsp), %%"REG_a
" \n\t"
218 "mov %%"REG_a
", %5 \n\t"
221 "pxor %%mm7, %%mm7 \n\t"
222 "mov %0, %%"REG_c
" \n\t"
223 "mov %1, %%"REG_D
" \n\t"
224 "mov %2, %%"REG_d
" \n\t"
225 "mov %3, %%"REG_b
" \n\t"
226 "xor %%"REG_a
", %%"REG_a
" \n\t" // i
227 PREFETCH
" (%%"REG_c
") \n\t"
228 PREFETCH
" 32(%%"REG_c
") \n\t"
229 PREFETCH
" 64(%%"REG_c
") \n\t"
232 #define CALL_MMXEXT_FILTER_CODE \
233 "movl (%%"REG_b"), %%esi \n\t"\
235 "movl (%%"REG_b", %%"REG_a"), %%esi \n\t"\
236 "add %%"REG_S", %%"REG_c" \n\t"\
237 "add %%"REG_a", %%"REG_D" \n\t"\
238 "xor %%"REG_a", %%"REG_a" \n\t"\
241 #define CALL_MMXEXT_FILTER_CODE \
242 "movl (%%"REG_b"), %%esi \n\t"\
244 "addl (%%"REG_b", %%"REG_a"), %%"REG_c" \n\t"\
245 "add %%"REG_a", %%"REG_D" \n\t"\
246 "xor %%"REG_a", %%"REG_a" \n\t"\
248 #endif /* ARCH_X86_64 */
250 CALL_MMXEXT_FILTER_CODE
251 CALL_MMXEXT_FILTER_CODE
252 CALL_MMXEXT_FILTER_CODE
253 CALL_MMXEXT_FILTER_CODE
254 CALL_MMXEXT_FILTER_CODE
255 CALL_MMXEXT_FILTER_CODE
256 CALL_MMXEXT_FILTER_CODE
257 CALL_MMXEXT_FILTER_CODE
260 "mov %5, %%"REG_b
" \n\t"
262 "mov %6, %%"REG_a
" \n\t"
263 "mov %%"REG_a
", -8(%%rsp) \n\t"
267 "mov %5, %%"REG_a
" \n\t"
268 "mov %%"REG_a
", -8(%%rsp) \n\t"
271 :: "m" (src
), "m" (dst
), "m" (filter
), "m" (filterPos
),
272 "m" (mmxextFilterCode
)
279 : "%"REG_a
, "%"REG_c
, "%"REG_d
, "%"REG_S
, "%"REG_D
285 for (i
=dstWidth
-1; (i
*xInc
)>>16 >=srcW
-1; i
--)
286 dst
[i
] = src
[srcW
-1]*128;
289 void ff_hcscale_fast_mmxext(SwsContext
*c
, int16_t *dst1
, int16_t *dst2
,
290 int dstWidth
, const uint8_t *src1
,
291 const uint8_t *src2
, int srcW
, int xInc
)
293 int32_t *filterPos
= c
->hChrFilterPos
;
294 int16_t *filter
= c
->hChrFilter
;
295 void *mmxextFilterCode
= c
->chrMmxextFilterCode
;
298 DECLARE_ALIGNED(8, uint64_t, ebxsave
);
301 DECLARE_ALIGNED(8, uint64_t, retsave
);
306 "mov %%"REG_b
", %7 \n\t"
308 "mov -8(%%rsp), %%"REG_a
" \n\t"
309 "mov %%"REG_a
", %8 \n\t"
313 "mov -8(%%rsp), %%"REG_a
" \n\t"
314 "mov %%"REG_a
", %7 \n\t"
317 "pxor %%mm7, %%mm7 \n\t"
318 "mov %0, %%"REG_c
" \n\t"
319 "mov %1, %%"REG_D
" \n\t"
320 "mov %2, %%"REG_d
" \n\t"
321 "mov %3, %%"REG_b
" \n\t"
322 "xor %%"REG_a
", %%"REG_a
" \n\t" // i
323 PREFETCH
" (%%"REG_c
") \n\t"
324 PREFETCH
" 32(%%"REG_c
") \n\t"
325 PREFETCH
" 64(%%"REG_c
") \n\t"
327 CALL_MMXEXT_FILTER_CODE
328 CALL_MMXEXT_FILTER_CODE
329 CALL_MMXEXT_FILTER_CODE
330 CALL_MMXEXT_FILTER_CODE
331 "xor %%"REG_a
", %%"REG_a
" \n\t" // i
332 "mov %5, %%"REG_c
" \n\t" // src
333 "mov %6, %%"REG_D
" \n\t" // buf2
334 PREFETCH
" (%%"REG_c
") \n\t"
335 PREFETCH
" 32(%%"REG_c
") \n\t"
336 PREFETCH
" 64(%%"REG_c
") \n\t"
338 CALL_MMXEXT_FILTER_CODE
339 CALL_MMXEXT_FILTER_CODE
340 CALL_MMXEXT_FILTER_CODE
341 CALL_MMXEXT_FILTER_CODE
344 "mov %7, %%"REG_b
" \n\t"
346 "mov %8, %%"REG_a
" \n\t"
347 "mov %%"REG_a
", -8(%%rsp) \n\t"
351 "mov %7, %%"REG_a
" \n\t"
352 "mov %%"REG_a
", -8(%%rsp) \n\t"
355 :: "m" (src1
), "m" (dst1
), "m" (filter
), "m" (filterPos
),
356 "m" (mmxextFilterCode
), "m" (src2
), "m"(dst2
)
363 : "%"REG_a
, "%"REG_c
, "%"REG_d
, "%"REG_S
, "%"REG_D
369 for (i
=dstWidth
-1; (i
*xInc
)>>16 >=srcW
-1; i
--) {
370 dst1
[i
] = src1
[srcW
-1]*128;
371 dst2
[i
] = src2
[srcW
-1]*128;
374 #endif //HAVE_INLINE_ASM