--- libavcodec/x86/Makefile | 3 +- libavcodec/x86/vc1dsp.asm | 585 ----------------------------------- libavcodec/x86/vc1dsp_loopfilter.asm | 317 +++++++++++++++++++ libavcodec/x86/vc1dsp_mc.asm | 292 +++++++++++++++++ 4 files changed, 611 insertions(+), 586 deletions(-) delete mode 100644 libavcodec/x86/vc1dsp.asm create mode 100644 libavcodec/x86/vc1dsp_loopfilter.asm create mode 100644 libavcodec/x86/vc1dsp_mc.asm
diff --git a/libavcodec/x86/Makefile b/libavcodec/x86/Makefile index 629b0ee..839b5bc 100644 --- a/libavcodec/x86/Makefile +++ b/libavcodec/x86/Makefile @@ -120,7 +120,8 @@ YASM-OBJS-$(CONFIG_QPELDSP) += x86/qpeldsp.o \ x86/fpel.o \ x86/qpel.o YASM-OBJS-$(CONFIG_RV34DSP) += x86/rv34dsp.o -YASM-OBJS-$(CONFIG_VC1DSP) += x86/vc1dsp.o +YASM-OBJS-$(CONFIG_VC1DSP) += x86/vc1dsp_loopfilter.o \ + x86/vc1dsp_mc.o YASM-OBJS-$(CONFIG_IDCTDSP) += x86/simple_idct10.o YASM-OBJS-$(CONFIG_VIDEODSP) += x86/videodsp.o YASM-OBJS-$(CONFIG_VP3DSP) += x86/vp3dsp.o diff --git a/libavcodec/x86/vc1dsp.asm b/libavcodec/x86/vc1dsp.asm deleted file mode 100644 index eee42c2..0000000 --- a/libavcodec/x86/vc1dsp.asm +++ /dev/null @@ -1,585 +0,0 @@ -;****************************************************************************** -;* VC1 DSP optimizations -;* Copyright (c) 2007 Christophe GISQUET <christophe.gisq...@free.fr> -;* Copyright (c) 2009 David Conrad -;* -;* This file is part of FFmpeg. -;* -;* FFmpeg is free software; you can redistribute it and/or -;* modify it under the terms of the GNU Lesser General Public -;* License as published by the Free Software Foundation; either -;* version 2.1 of the License, or (at your option) any later version. -;* -;* FFmpeg is distributed in the hope that it will be useful, -;* but WITHOUT ANY WARRANTY; without even the implied warranty of -;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU -;* Lesser General Public License for more details. -;* -;* You should have received a copy of the GNU Lesser General Public -;* License along with FFmpeg; if not, write to the Free Software -;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA -;****************************************************************************** - -%include "libavutil/x86/x86util.asm" - -cextern pw_4 -cextern pw_5 -cextern pw_9 -cextern pw_128 - -section .text - -; dst_low, dst_high (src), zero -; zero-extends one vector from 8 to 16 bits -%macro UNPACK_8TO16 4 - mova m%2, m%3 - punpckh%1 m%3, m%4 - punpckl%1 m%2, m%4 -%endmacro - -%macro STORE_4_WORDS 6 -%if cpuflag(sse4) - pextrw %1, %5, %6+0 - pextrw %2, %5, %6+1 - pextrw %3, %5, %6+2 - pextrw %4, %5, %6+3 -%else - movd %6d, %5 -%if mmsize==16 - psrldq %5, 4 -%else - psrlq %5, 32 -%endif - mov %1, %6w - shr %6, 16 - mov %2, %6w - movd %6d, %5 - mov %3, %6w - shr %6, 16 - mov %4, %6w -%endif -%endmacro - -; in: p1 p0 q0 q1, clobbers p0 -; out: p1 = (2*(p1 - q1) - 5*(p0 - q0) + 4) >> 3 -%macro VC1_LOOP_FILTER_A0 4 - psubw %1, %4 - psubw %2, %3 - paddw %1, %1 - pmullw %2, [pw_5] - psubw %1, %2 - paddw %1, [pw_4] - psraw %1, 3 -%endmacro - -; in: p0 q0 a0 a1 a2 -; m0 m1 m7 m6 m5 -; %1: size -; out: m0=p0' m1=q0' -%macro VC1_FILTER 1 - PABSW m4, m7 - PABSW m3, m6 - PABSW m2, m5 - mova m6, m4 - pminsw m3, m2 - pcmpgtw m6, m3 ; if (a2 < a0 || a1 < a0) - psubw m3, m4 - pmullw m3, [pw_5] ; 5*(a3 - a0) - PABSW m2, m3 - psraw m2, 3 ; abs(d/8) - pxor m7, m3 ; d_sign ^= a0_sign - - pxor m5, m5 - movd m3, r2d -%if %1 > 4 - punpcklbw m3, m3 -%endif - punpcklbw m3, m5 - pcmpgtw m3, m4 ; if (a0 < pq) - pand m6, m3 - - mova m3, m0 - psubw m3, m1 - PABSW m4, m3 - psraw m4, 1 - pxor m3, m7 ; d_sign ^ clip_sign - psraw m3, 15 - pminsw m2, m4 ; min(d, clip) - pcmpgtw m4, m5 - pand m6, m4 ; filt3 (C return value) - -; each set of 4 pixels is not filtered if the 3rd is not -%if mmsize==16 - pshuflw m4, m6, 0xaa -%if %1 > 4 - pshufhw m4, m4, 0xaa -%endif -%else - pshufw m4, m6, 0xaa -%endif - pandn m3, m4 - pand m2, m6 - pand m3, m2 ; d final - - psraw m7, 15 - pxor m3, m7 - psubw m3, m7 - psubw m0, m3 - paddw m1, m3 - packuswb m0, m0 - packuswb m1, m1 -%endmacro - -; 1st param: size of filter -; 2nd param: mov suffix equivalent to the filter size -%macro VC1_V_LOOP_FILTER 2 - pxor m5, m5 - mov%2 m6, [r4] - mov%2 m4, [r4+r1] - mov%2 m7, [r4+2*r1] - mov%2 m0, [r4+r3] - punpcklbw m6, m5 - punpcklbw m4, m5 - punpcklbw m7, m5 - punpcklbw m0, m5 - - VC1_LOOP_FILTER_A0 m6, m4, m7, m0 - mov%2 m1, [r0] - mov%2 m2, [r0+r1] - punpcklbw m1, m5 - punpcklbw m2, m5 - mova m4, m0 - VC1_LOOP_FILTER_A0 m7, m4, m1, m2 - mov%2 m3, [r0+2*r1] - mov%2 m4, [r0+r3] - punpcklbw m3, m5 - punpcklbw m4, m5 - mova m5, m1 - VC1_LOOP_FILTER_A0 m5, m2, m3, m4 - - VC1_FILTER %1 - mov%2 [r4+r3], m0 - mov%2 [r0], m1 -%endmacro - -; 1st param: size of filter -; NOTE: UNPACK_8TO16 this number of 8 bit numbers are in half a register -; 2nd (optional) param: temp register to use for storing words -%macro VC1_H_LOOP_FILTER 1-2 -%if %1 == 4 - movq m0, [r0 -4] - movq m1, [r0+ r1-4] - movq m2, [r0+2*r1-4] - movq m3, [r0+ r3-4] - TRANSPOSE4x4B 0, 1, 2, 3, 4 -%else - movq m0, [r0 -4] - movq m4, [r0+ r1-4] - movq m1, [r0+2*r1-4] - movq m5, [r0+ r3-4] - movq m2, [r4 -4] - movq m6, [r4+ r1-4] - movq m3, [r4+2*r1-4] - movq m7, [r4+ r3-4] - punpcklbw m0, m4 - punpcklbw m1, m5 - punpcklbw m2, m6 - punpcklbw m3, m7 - TRANSPOSE4x4W 0, 1, 2, 3, 4 -%endif - pxor m5, m5 - - UNPACK_8TO16 bw, 6, 0, 5 - UNPACK_8TO16 bw, 7, 1, 5 - VC1_LOOP_FILTER_A0 m6, m0, m7, m1 - UNPACK_8TO16 bw, 4, 2, 5 - mova m0, m1 ; m0 = p0 - VC1_LOOP_FILTER_A0 m7, m1, m4, m2 - UNPACK_8TO16 bw, 1, 3, 5 - mova m5, m4 - VC1_LOOP_FILTER_A0 m5, m2, m1, m3 - SWAP 1, 4 ; m1 = q0 - - VC1_FILTER %1 - punpcklbw m0, m1 -%if %0 > 1 - STORE_4_WORDS [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, %2 -%if %1 > 4 - psrldq m0, 4 - STORE_4_WORDS [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, %2 -%endif -%else - STORE_4_WORDS [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, 0 - STORE_4_WORDS [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, 4 -%endif -%endmacro - - -%macro START_V_FILTER 0 - mov r4, r0 - lea r3, [4*r1] - sub r4, r3 - lea r3, [r1+2*r1] - imul r2, 0x01010101 -%endmacro - -%macro START_H_FILTER 1 - lea r3, [r1+2*r1] -%if %1 > 4 - lea r4, [r0+4*r1] -%endif - imul r2, 0x01010101 -%endmacro - -%macro VC1_LF 0 -cglobal vc1_v_loop_filter_internal - VC1_V_LOOP_FILTER 4, d - ret - -cglobal vc1_h_loop_filter_internal - VC1_H_LOOP_FILTER 4, r4 - ret - -; void ff_vc1_v_loop_filter4_mmxext(uint8_t *src, int stride, int pq) -cglobal vc1_v_loop_filter4, 3,5,0 - START_V_FILTER - call vc1_v_loop_filter_internal - RET - -; void ff_vc1_h_loop_filter4_mmxext(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter4, 3,5,0 - START_H_FILTER 4 - call vc1_h_loop_filter_internal - RET - -; void ff_vc1_v_loop_filter8_mmxext(uint8_t *src, int stride, int pq) -cglobal vc1_v_loop_filter8, 3,5,0 - START_V_FILTER - call vc1_v_loop_filter_internal - add r4, 4 - add r0, 4 - call vc1_v_loop_filter_internal - RET - -; void ff_vc1_h_loop_filter8_mmxext(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter8, 3,5,0 - START_H_FILTER 4 - call vc1_h_loop_filter_internal - lea r0, [r0+4*r1] - call vc1_h_loop_filter_internal - RET -%endmacro - -INIT_MMX mmxext -VC1_LF - -INIT_XMM sse2 -; void ff_vc1_v_loop_filter8_sse2(uint8_t *src, int stride, int pq) -cglobal vc1_v_loop_filter8, 3,5,8 - START_V_FILTER - VC1_V_LOOP_FILTER 8, q - RET - -; void ff_vc1_h_loop_filter8_sse2(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter8, 3,6,8 - START_H_FILTER 8 - VC1_H_LOOP_FILTER 8, r5 - RET - -INIT_MMX ssse3 -; void ff_vc1_v_loop_filter4_ssse3(uint8_t *src, int stride, int pq) -cglobal vc1_v_loop_filter4, 3,5,0 - START_V_FILTER - VC1_V_LOOP_FILTER 4, d - RET - -; void ff_vc1_h_loop_filter4_ssse3(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter4, 3,5,0 - START_H_FILTER 4 - VC1_H_LOOP_FILTER 4, r4 - RET - -INIT_XMM ssse3 -; void ff_vc1_v_loop_filter8_ssse3(uint8_t *src, int stride, int pq) -cglobal vc1_v_loop_filter8, 3,5,8 - START_V_FILTER - VC1_V_LOOP_FILTER 8, q - RET - -; void ff_vc1_h_loop_filter8_ssse3(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter8, 3,6,8 - START_H_FILTER 8 - VC1_H_LOOP_FILTER 8, r5 - RET - -INIT_XMM sse4 -; void ff_vc1_h_loop_filter8_sse4(uint8_t *src, int stride, int pq) -cglobal vc1_h_loop_filter8, 3,5,8 - START_H_FILTER 8 - VC1_H_LOOP_FILTER 8 - RET - -%if HAVE_MMX_INLINE - -; XXX some of these macros are not used right now, but they will in the future -; when more functions are ported. - -%macro OP_PUT 2 ; dst, src -%endmacro - -%macro OP_AVG 2 ; dst, src - pavgb %1, %2 -%endmacro - -%macro NORMALIZE_MMX 1 ; shift - paddw m3, m7 ; +bias-r - paddw m4, m7 ; +bias-r - psraw m3, %1 - psraw m4, %1 -%endmacro - -%macro TRANSFER_DO_PACK 2 ; op, dst - packuswb m3, m4 - %1 m3, [%2] - mova [%2], m3 -%endmacro - -%macro TRANSFER_DONT_PACK 2 ; op, dst - %1 m3, [%2] - %1 m3, [%2 + mmsize] - mova [%2], m3 - mova [mmsize + %2], m4 -%endmacro - -; see MSPEL_FILTER13_CORE for use as UNPACK macro -%macro DO_UNPACK 1 ; reg - punpcklbw %1, m0 -%endmacro -%macro DONT_UNPACK 1 ; reg -%endmacro - -; Compute the rounder 32-r or 8-r and unpacks it to m7 -%macro LOAD_ROUNDER_MMX 1 ; round - movd m7, %1 - punpcklwd m7, m7 - punpckldq m7, m7 -%endmacro - -%macro SHIFT2_LINE 5 ; off, r0, r1, r2, r3 - paddw m%3, m%4 - movh m%2, [srcq + stride_neg2] - pmullw m%3, m6 - punpcklbw m%2, m0 - movh m%5, [srcq + strideq] - psubw m%3, m%2 - punpcklbw m%5, m0 - paddw m%3, m7 - psubw m%3, m%5 - psraw m%3, shift - movu [dstq + %1], m%3 - add srcq, strideq -%endmacro - -INIT_MMX mmx -; void ff_vc1_put_ver_16b_shift2_mmx(int16_t *dst, const uint8_t *src, -; x86_reg stride, int rnd, int64_t shift) -; Sacrificing m6 makes it possible to pipeline loads from src -%if ARCH_X86_32 -cglobal vc1_put_ver_16b_shift2, 3,6,0, dst, src, stride - DECLARE_REG_TMP 3, 4, 5 - %define rnd r3mp - %define shift qword r4m -%else ; X86_64 -cglobal vc1_put_ver_16b_shift2, 4,7,0, dst, src, stride - DECLARE_REG_TMP 4, 5, 6 - %define rnd r3d - ; We need shift either in memory or in a mm reg as it's used in psraw - ; On WIN64, the arg is already on the stack - ; On UNIX64, m5 doesn't seem to be used -%if WIN64 - %define shift r4mp -%else ; UNIX64 - %define shift m5 - mova shift, r4q -%endif ; WIN64 -%endif ; X86_32 -%define stride_neg2 t0q -%define stride_9minus4 t1q -%define i t2q - mov stride_neg2, strideq - neg stride_neg2 - add stride_neg2, stride_neg2 - lea stride_9minus4, [strideq * 9 - 4] - mov i, 3 - LOAD_ROUNDER_MMX rnd - mova m6, [pw_9] - pxor m0, m0 -.loop: - movh m2, [srcq] - add srcq, strideq - movh m3, [srcq] - punpcklbw m2, m0 - punpcklbw m3, m0 - SHIFT2_LINE 0, 1, 2, 3, 4 - SHIFT2_LINE 24, 2, 3, 4, 1 - SHIFT2_LINE 48, 3, 4, 1, 2 - SHIFT2_LINE 72, 4, 1, 2, 3 - SHIFT2_LINE 96, 1, 2, 3, 4 - SHIFT2_LINE 120, 2, 3, 4, 1 - SHIFT2_LINE 144, 3, 4, 1, 2 - SHIFT2_LINE 168, 4, 1, 2, 3 - sub srcq, stride_9minus4 - add dstq, 8 - dec i - jnz .loop - REP_RET -%undef rnd -%undef shift -%undef stride_neg2 -%undef stride_9minus4 -%undef i - -; void ff_vc1_*_hor_16b_shift2_mmx(uint8_t *dst, x86_reg stride, -; const int16_t *src, int rnd); -; Data is already unpacked, so some operations can directly be made from -; memory. -%macro HOR_16B_SHIFT2 2 ; op, opname -cglobal vc1_%2_hor_16b_shift2, 4, 5, 0, dst, stride, src, rnd, h - mov hq, 8 - sub srcq, 2 - sub rndd, (-1+9+9-1) * 1024 ; add -1024 bias - LOAD_ROUNDER_MMX rndq - mova m5, [pw_9] - mova m6, [pw_128] - pxor m0, m0 - -.loop: - mova m1, [srcq + 2 * 0] - mova m2, [srcq + 2 * 0 + mmsize] - mova m3, [srcq + 2 * 1] - mova m4, [srcq + 2 * 1 + mmsize] - paddw m3, [srcq + 2 * 2] - paddw m4, [srcq + 2 * 2 + mmsize] - paddw m1, [srcq + 2 * 3] - paddw m2, [srcq + 2 * 3 + mmsize] - pmullw m3, m5 - pmullw m4, m5 - psubw m3, m1 - psubw m4, m2 - NORMALIZE_MMX 7 - ; remove bias - paddw m3, m6 - paddw m4, m6 - TRANSFER_DO_PACK %1, dstq - add srcq, 24 - add dstq, strideq - dec hq - jnz .loop - - RET -%endmacro - -INIT_MMX mmx -HOR_16B_SHIFT2 OP_PUT, put - -INIT_MMX mmxext -HOR_16B_SHIFT2 OP_AVG, avg -%endif ; HAVE_MMX_INLINE - -%macro INV_TRANS_INIT 0 - movsxdifnidn linesizeq, linesized - movd m0, blockd - SPLATW m0, m0 - pxor m1, m1 - psubw m1, m0 - packuswb m0, m0 - packuswb m1, m1 - - DEFINE_ARGS dest, linesize, linesize3 - lea linesize3q, [linesizeq*3] -%endmacro - -%macro INV_TRANS_PROCESS 1 - mov%1 m2, [destq+linesizeq*0] - mov%1 m3, [destq+linesizeq*1] - mov%1 m4, [destq+linesizeq*2] - mov%1 m5, [destq+linesize3q] - paddusb m2, m0 - paddusb m3, m0 - paddusb m4, m0 - paddusb m5, m0 - psubusb m2, m1 - psubusb m3, m1 - psubusb m4, m1 - psubusb m5, m1 - mov%1 [linesizeq*0+destq], m2 - mov%1 [linesizeq*1+destq], m3 - mov%1 [linesizeq*2+destq], m4 - mov%1 [linesize3q +destq], m5 -%endmacro - -; ff_vc1_inv_trans_?x?_dc_mmxext(uint8_t *dest, int linesize, int16_t *block) -INIT_MMX mmxext -cglobal vc1_inv_trans_4x4_dc, 3,4,0, dest, linesize, block - movsx r3d, WORD [blockq] - mov blockd, r3d ; dc - shl blockd, 4 ; 16 * dc - lea blockd, [blockq+r3+4] ; 17 * dc + 4 - sar blockd, 3 ; >> 3 - mov r3d, blockd ; dc - shl blockd, 4 ; 16 * dc - lea blockd, [blockq+r3+64] ; 17 * dc + 64 - sar blockd, 7 ; >> 7 - - INV_TRANS_INIT - - INV_TRANS_PROCESS h - RET - -INIT_MMX mmxext -cglobal vc1_inv_trans_4x8_dc, 3,4,0, dest, linesize, block - movsx r3d, WORD [blockq] - mov blockd, r3d ; dc - shl blockd, 4 ; 16 * dc - lea blockd, [blockq+r3+4] ; 17 * dc + 4 - sar blockd, 3 ; >> 3 - shl blockd, 2 ; 4 * dc - lea blockd, [blockq*3+64] ; 12 * dc + 64 - sar blockd, 7 ; >> 7 - - INV_TRANS_INIT - - INV_TRANS_PROCESS h - lea destq, [destq+linesizeq*4] - INV_TRANS_PROCESS h - RET - -INIT_MMX mmxext -cglobal vc1_inv_trans_8x4_dc, 3,4,0, dest, linesize, block - movsx blockd, WORD [blockq] ; dc - lea blockd, [blockq*3+1] ; 3 * dc + 1 - sar blockd, 1 ; >> 1 - mov r3d, blockd ; dc - shl blockd, 4 ; 16 * dc - lea blockd, [blockq+r3+64] ; 17 * dc + 64 - sar blockd, 7 ; >> 7 - - INV_TRANS_INIT - - INV_TRANS_PROCESS a - RET - -INIT_MMX mmxext -cglobal vc1_inv_trans_8x8_dc, 3,3,0, dest, linesize, block - movsx blockd, WORD [blockq] ; dc - lea blockd, [blockq*3+1] ; 3 * dc + 1 - sar blockd, 1 ; >> 1 - lea blockd, [blockq*3+16] ; 3 * dc + 16 - sar blockd, 5 ; >> 5 - - INV_TRANS_INIT - - INV_TRANS_PROCESS a - lea destq, [destq+linesizeq*4] - INV_TRANS_PROCESS a - RET diff --git a/libavcodec/x86/vc1dsp_loopfilter.asm b/libavcodec/x86/vc1dsp_loopfilter.asm new file mode 100644 index 0000000..1838f6f --- /dev/null +++ b/libavcodec/x86/vc1dsp_loopfilter.asm @@ -0,0 +1,317 @@ +;****************************************************************************** +;* VC1 loopfilter optimizations +;* Copyright (c) 2009 David Conrad +;* +;* This file is part of FFmpeg. +;* +;* FFmpeg is free software; you can redistribute it and/or +;* modify it under the terms of the GNU Lesser General Public +;* License as published by the Free Software Foundation; either +;* version 2.1 of the License, or (at your option) any later version. +;* +;* FFmpeg is distributed in the hope that it will be useful, +;* but WITHOUT ANY WARRANTY; without even the implied warranty of +;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +;* Lesser General Public License for more details. +;* +;* You should have received a copy of the GNU Lesser General Public +;* License along with FFmpeg; if not, write to the Free Software +;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +;****************************************************************************** + +%include "libavutil/x86/x86util.asm" + +cextern pw_4 +cextern pw_5 + +section .text + +; dst_low, dst_high (src), zero +; zero-extends one vector from 8 to 16 bits +%macro UNPACK_8TO16 4 + mova m%2, m%3 + punpckh%1 m%3, m%4 + punpckl%1 m%2, m%4 +%endmacro + +%macro STORE_4_WORDS 6 +%if cpuflag(sse4) + pextrw %1, %5, %6+0 + pextrw %2, %5, %6+1 + pextrw %3, %5, %6+2 + pextrw %4, %5, %6+3 +%else + movd %6d, %5 +%if mmsize==16 + psrldq %5, 4 +%else + psrlq %5, 32 +%endif + mov %1, %6w + shr %6, 16 + mov %2, %6w + movd %6d, %5 + mov %3, %6w + shr %6, 16 + mov %4, %6w +%endif +%endmacro + +; in: p1 p0 q0 q1, clobbers p0 +; out: p1 = (2*(p1 - q1) - 5*(p0 - q0) + 4) >> 3 +%macro VC1_LOOP_FILTER_A0 4 + psubw %1, %4 + psubw %2, %3 + paddw %1, %1 + pmullw %2, [pw_5] + psubw %1, %2 + paddw %1, [pw_4] + psraw %1, 3 +%endmacro + +; in: p0 q0 a0 a1 a2 +; m0 m1 m7 m6 m5 +; %1: size +; out: m0=p0' m1=q0' +%macro VC1_FILTER 1 + PABSW m4, m7 + PABSW m3, m6 + PABSW m2, m5 + mova m6, m4 + pminsw m3, m2 + pcmpgtw m6, m3 ; if (a2 < a0 || a1 < a0) + psubw m3, m4 + pmullw m3, [pw_5] ; 5*(a3 - a0) + PABSW m2, m3 + psraw m2, 3 ; abs(d/8) + pxor m7, m3 ; d_sign ^= a0_sign + + pxor m5, m5 + movd m3, r2d +%if %1 > 4 + punpcklbw m3, m3 +%endif + punpcklbw m3, m5 + pcmpgtw m3, m4 ; if (a0 < pq) + pand m6, m3 + + mova m3, m0 + psubw m3, m1 + PABSW m4, m3 + psraw m4, 1 + pxor m3, m7 ; d_sign ^ clip_sign + psraw m3, 15 + pminsw m2, m4 ; min(d, clip) + pcmpgtw m4, m5 + pand m6, m4 ; filt3 (C return value) + +; each set of 4 pixels is not filtered if the 3rd is not +%if mmsize==16 + pshuflw m4, m6, 0xaa +%if %1 > 4 + pshufhw m4, m4, 0xaa +%endif +%else + pshufw m4, m6, 0xaa +%endif + pandn m3, m4 + pand m2, m6 + pand m3, m2 ; d final + + psraw m7, 15 + pxor m3, m7 + psubw m3, m7 + psubw m0, m3 + paddw m1, m3 + packuswb m0, m0 + packuswb m1, m1 +%endmacro + +; 1st param: size of filter +; 2nd param: mov suffix equivalent to the filter size +%macro VC1_V_LOOP_FILTER 2 + pxor m5, m5 + mov%2 m6, [r4] + mov%2 m4, [r4+r1] + mov%2 m7, [r4+2*r1] + mov%2 m0, [r4+r3] + punpcklbw m6, m5 + punpcklbw m4, m5 + punpcklbw m7, m5 + punpcklbw m0, m5 + + VC1_LOOP_FILTER_A0 m6, m4, m7, m0 + mov%2 m1, [r0] + mov%2 m2, [r0+r1] + punpcklbw m1, m5 + punpcklbw m2, m5 + mova m4, m0 + VC1_LOOP_FILTER_A0 m7, m4, m1, m2 + mov%2 m3, [r0+2*r1] + mov%2 m4, [r0+r3] + punpcklbw m3, m5 + punpcklbw m4, m5 + mova m5, m1 + VC1_LOOP_FILTER_A0 m5, m2, m3, m4 + + VC1_FILTER %1 + mov%2 [r4+r3], m0 + mov%2 [r0], m1 +%endmacro + +; 1st param: size of filter +; NOTE: UNPACK_8TO16 this number of 8 bit numbers are in half a register +; 2nd (optional) param: temp register to use for storing words +%macro VC1_H_LOOP_FILTER 1-2 +%if %1 == 4 + movq m0, [r0 -4] + movq m1, [r0+ r1-4] + movq m2, [r0+2*r1-4] + movq m3, [r0+ r3-4] + TRANSPOSE4x4B 0, 1, 2, 3, 4 +%else + movq m0, [r0 -4] + movq m4, [r0+ r1-4] + movq m1, [r0+2*r1-4] + movq m5, [r0+ r3-4] + movq m2, [r4 -4] + movq m6, [r4+ r1-4] + movq m3, [r4+2*r1-4] + movq m7, [r4+ r3-4] + punpcklbw m0, m4 + punpcklbw m1, m5 + punpcklbw m2, m6 + punpcklbw m3, m7 + TRANSPOSE4x4W 0, 1, 2, 3, 4 +%endif + pxor m5, m5 + + UNPACK_8TO16 bw, 6, 0, 5 + UNPACK_8TO16 bw, 7, 1, 5 + VC1_LOOP_FILTER_A0 m6, m0, m7, m1 + UNPACK_8TO16 bw, 4, 2, 5 + mova m0, m1 ; m0 = p0 + VC1_LOOP_FILTER_A0 m7, m1, m4, m2 + UNPACK_8TO16 bw, 1, 3, 5 + mova m5, m4 + VC1_LOOP_FILTER_A0 m5, m2, m1, m3 + SWAP 1, 4 ; m1 = q0 + + VC1_FILTER %1 + punpcklbw m0, m1 +%if %0 > 1 + STORE_4_WORDS [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, %2 +%if %1 > 4 + psrldq m0, 4 + STORE_4_WORDS [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, %2 +%endif +%else + STORE_4_WORDS [r0-1], [r0+r1-1], [r0+2*r1-1], [r0+r3-1], m0, 0 + STORE_4_WORDS [r4-1], [r4+r1-1], [r4+2*r1-1], [r4+r3-1], m0, 4 +%endif +%endmacro + + +%macro START_V_FILTER 0 + mov r4, r0 + lea r3, [4*r1] + sub r4, r3 + lea r3, [r1+2*r1] + imul r2, 0x01010101 +%endmacro + +%macro START_H_FILTER 1 + lea r3, [r1+2*r1] +%if %1 > 4 + lea r4, [r0+4*r1] +%endif + imul r2, 0x01010101 +%endmacro + +%macro VC1_LF 0 +cglobal vc1_v_loop_filter_internal + VC1_V_LOOP_FILTER 4, d + ret + +cglobal vc1_h_loop_filter_internal + VC1_H_LOOP_FILTER 4, r4 + ret + +; void ff_vc1_v_loop_filter4_mmxext(uint8_t *src, int stride, int pq) +cglobal vc1_v_loop_filter4, 3,5,0 + START_V_FILTER + call vc1_v_loop_filter_internal + RET + +; void ff_vc1_h_loop_filter4_mmxext(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter4, 3,5,0 + START_H_FILTER 4 + call vc1_h_loop_filter_internal + RET + +; void ff_vc1_v_loop_filter8_mmxext(uint8_t *src, int stride, int pq) +cglobal vc1_v_loop_filter8, 3,5,0 + START_V_FILTER + call vc1_v_loop_filter_internal + add r4, 4 + add r0, 4 + call vc1_v_loop_filter_internal + RET + +; void ff_vc1_h_loop_filter8_mmxext(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter8, 3,5,0 + START_H_FILTER 4 + call vc1_h_loop_filter_internal + lea r0, [r0+4*r1] + call vc1_h_loop_filter_internal + RET +%endmacro + +INIT_MMX mmxext +VC1_LF + +INIT_XMM sse2 +; void ff_vc1_v_loop_filter8_sse2(uint8_t *src, int stride, int pq) +cglobal vc1_v_loop_filter8, 3,5,8 + START_V_FILTER + VC1_V_LOOP_FILTER 8, q + RET + +; void ff_vc1_h_loop_filter8_sse2(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter8, 3,6,8 + START_H_FILTER 8 + VC1_H_LOOP_FILTER 8, r5 + RET + +INIT_MMX ssse3 +; void ff_vc1_v_loop_filter4_ssse3(uint8_t *src, int stride, int pq) +cglobal vc1_v_loop_filter4, 3,5,0 + START_V_FILTER + VC1_V_LOOP_FILTER 4, d + RET + +; void ff_vc1_h_loop_filter4_ssse3(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter4, 3,5,0 + START_H_FILTER 4 + VC1_H_LOOP_FILTER 4, r4 + RET + +INIT_XMM ssse3 +; void ff_vc1_v_loop_filter8_ssse3(uint8_t *src, int stride, int pq) +cglobal vc1_v_loop_filter8, 3,5,8 + START_V_FILTER + VC1_V_LOOP_FILTER 8, q + RET + +; void ff_vc1_h_loop_filter8_ssse3(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter8, 3,6,8 + START_H_FILTER 8 + VC1_H_LOOP_FILTER 8, r5 + RET + +INIT_XMM sse4 +; void ff_vc1_h_loop_filter8_sse4(uint8_t *src, int stride, int pq) +cglobal vc1_h_loop_filter8, 3,5,8 + START_H_FILTER 8 + VC1_H_LOOP_FILTER 8 + RET diff --git a/libavcodec/x86/vc1dsp_mc.asm b/libavcodec/x86/vc1dsp_mc.asm new file mode 100644 index 0000000..7eaf043 --- /dev/null +++ b/libavcodec/x86/vc1dsp_mc.asm @@ -0,0 +1,292 @@ +;****************************************************************************** +;* VC1 motion compensation optimizations +;* Copyright (c) 2007 Christophe GISQUET <christophe.gisq...@free.fr> +;* +;* This file is part of FFmpeg. +;* +;* FFmpeg is free software; you can redistribute it and/or +;* modify it under the terms of the GNU Lesser General Public +;* License as published by the Free Software Foundation; either +;* version 2.1 of the License, or (at your option) any later version. +;* +;* FFmpeg is distributed in the hope that it will be useful, +;* but WITHOUT ANY WARRANTY; without even the implied warranty of +;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU +;* Lesser General Public License for more details. +;* +;* You should have received a copy of the GNU Lesser General Public +;* License along with FFmpeg; if not, write to the Free Software +;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA +;****************************************************************************** + +%include "libavutil/x86/x86util.asm" + +cextern pw_9 +cextern pw_128 + +section .text + +%if HAVE_MMX_INLINE + +; XXX some of these macros are not used right now, but they will in the future +; when more functions are ported. + +%macro OP_PUT 2 ; dst, src +%endmacro + +%macro OP_AVG 2 ; dst, src + pavgb %1, %2 +%endmacro + +%macro NORMALIZE_MMX 1 ; shift + paddw m3, m7 ; +bias-r + paddw m4, m7 ; +bias-r + psraw m3, %1 + psraw m4, %1 +%endmacro + +%macro TRANSFER_DO_PACK 2 ; op, dst + packuswb m3, m4 + %1 m3, [%2] + mova [%2], m3 +%endmacro + +%macro TRANSFER_DONT_PACK 2 ; op, dst + %1 m3, [%2] + %1 m3, [%2 + mmsize] + mova [%2], m3 + mova [mmsize + %2], m4 +%endmacro + +; see MSPEL_FILTER13_CORE for use as UNPACK macro +%macro DO_UNPACK 1 ; reg + punpcklbw %1, m0 +%endmacro +%macro DONT_UNPACK 1 ; reg +%endmacro + +; Compute the rounder 32-r or 8-r and unpacks it to m7 +%macro LOAD_ROUNDER_MMX 1 ; round + movd m7, %1 + punpcklwd m7, m7 + punpckldq m7, m7 +%endmacro + +%macro SHIFT2_LINE 5 ; off, r0, r1, r2, r3 + paddw m%3, m%4 + movh m%2, [srcq + stride_neg2] + pmullw m%3, m6 + punpcklbw m%2, m0 + movh m%5, [srcq + strideq] + psubw m%3, m%2 + punpcklbw m%5, m0 + paddw m%3, m7 + psubw m%3, m%5 + psraw m%3, shift + movu [dstq + %1], m%3 + add srcq, strideq +%endmacro + +INIT_MMX mmx +; void ff_vc1_put_ver_16b_shift2_mmx(int16_t *dst, const uint8_t *src, +; x86_reg stride, int rnd, int64_t shift) +; Sacrificing m6 makes it possible to pipeline loads from src +%if ARCH_X86_32 +cglobal vc1_put_ver_16b_shift2, 3,6,0, dst, src, stride + DECLARE_REG_TMP 3, 4, 5 + %define rnd r3mp + %define shift qword r4m +%else ; X86_64 +cglobal vc1_put_ver_16b_shift2, 4,7,0, dst, src, stride + DECLARE_REG_TMP 4, 5, 6 + %define rnd r3d + ; We need shift either in memory or in a mm reg as it's used in psraw + ; On WIN64, the arg is already on the stack + ; On UNIX64, m5 doesn't seem to be used +%if WIN64 + %define shift r4mp +%else ; UNIX64 + %define shift m5 + mova shift, r4q +%endif ; WIN64 +%endif ; X86_32 +%define stride_neg2 t0q +%define stride_9minus4 t1q +%define i t2q + mov stride_neg2, strideq + neg stride_neg2 + add stride_neg2, stride_neg2 + lea stride_9minus4, [strideq * 9 - 4] + mov i, 3 + LOAD_ROUNDER_MMX rnd + mova m6, [pw_9] + pxor m0, m0 +.loop: + movh m2, [srcq] + add srcq, strideq + movh m3, [srcq] + punpcklbw m2, m0 + punpcklbw m3, m0 + SHIFT2_LINE 0, 1, 2, 3, 4 + SHIFT2_LINE 24, 2, 3, 4, 1 + SHIFT2_LINE 48, 3, 4, 1, 2 + SHIFT2_LINE 72, 4, 1, 2, 3 + SHIFT2_LINE 96, 1, 2, 3, 4 + SHIFT2_LINE 120, 2, 3, 4, 1 + SHIFT2_LINE 144, 3, 4, 1, 2 + SHIFT2_LINE 168, 4, 1, 2, 3 + sub srcq, stride_9minus4 + add dstq, 8 + dec i + jnz .loop + REP_RET +%undef rnd +%undef shift +%undef stride_neg2 +%undef stride_9minus4 +%undef i + +; void ff_vc1_*_hor_16b_shift2_mmx(uint8_t *dst, x86_reg stride, +; const int16_t *src, int rnd); +; Data is already unpacked, so some operations can directly be made from +; memory. +%macro HOR_16B_SHIFT2 2 ; op, opname +cglobal vc1_%2_hor_16b_shift2, 4, 5, 0, dst, stride, src, rnd, h + mov hq, 8 + sub srcq, 2 + sub rndd, (-1+9+9-1) * 1024 ; add -1024 bias + LOAD_ROUNDER_MMX rndq + mova m5, [pw_9] + mova m6, [pw_128] + pxor m0, m0 + +.loop: + mova m1, [srcq + 2 * 0] + mova m2, [srcq + 2 * 0 + mmsize] + mova m3, [srcq + 2 * 1] + mova m4, [srcq + 2 * 1 + mmsize] + paddw m3, [srcq + 2 * 2] + paddw m4, [srcq + 2 * 2 + mmsize] + paddw m1, [srcq + 2 * 3] + paddw m2, [srcq + 2 * 3 + mmsize] + pmullw m3, m5 + pmullw m4, m5 + psubw m3, m1 + psubw m4, m2 + NORMALIZE_MMX 7 + ; remove bias + paddw m3, m6 + paddw m4, m6 + TRANSFER_DO_PACK %1, dstq + add srcq, 24 + add dstq, strideq + dec hq + jnz .loop + + RET +%endmacro + +INIT_MMX mmx +HOR_16B_SHIFT2 OP_PUT, put + +INIT_MMX mmxext +HOR_16B_SHIFT2 OP_AVG, avg +%endif ; HAVE_MMX_INLINE + +%macro INV_TRANS_INIT 0 + movsxdifnidn linesizeq, linesized + movd m0, blockd + SPLATW m0, m0 + pxor m1, m1 + psubw m1, m0 + packuswb m0, m0 + packuswb m1, m1 + + DEFINE_ARGS dest, linesize, linesize3 + lea linesize3q, [linesizeq*3] +%endmacro + +%macro INV_TRANS_PROCESS 1 + mov%1 m2, [destq+linesizeq*0] + mov%1 m3, [destq+linesizeq*1] + mov%1 m4, [destq+linesizeq*2] + mov%1 m5, [destq+linesize3q] + paddusb m2, m0 + paddusb m3, m0 + paddusb m4, m0 + paddusb m5, m0 + psubusb m2, m1 + psubusb m3, m1 + psubusb m4, m1 + psubusb m5, m1 + mov%1 [linesizeq*0+destq], m2 + mov%1 [linesizeq*1+destq], m3 + mov%1 [linesizeq*2+destq], m4 + mov%1 [linesize3q +destq], m5 +%endmacro + +; ff_vc1_inv_trans_?x?_dc_mmxext(uint8_t *dest, int linesize, int16_t *block) +INIT_MMX mmxext +cglobal vc1_inv_trans_4x4_dc, 3,4,0, dest, linesize, block + movsx r3d, WORD [blockq] + mov blockd, r3d ; dc + shl blockd, 4 ; 16 * dc + lea blockd, [blockq+r3+4] ; 17 * dc + 4 + sar blockd, 3 ; >> 3 + mov r3d, blockd ; dc + shl blockd, 4 ; 16 * dc + lea blockd, [blockq+r3+64] ; 17 * dc + 64 + sar blockd, 7 ; >> 7 + + INV_TRANS_INIT + + INV_TRANS_PROCESS h + RET + +INIT_MMX mmxext +cglobal vc1_inv_trans_4x8_dc, 3,4,0, dest, linesize, block + movsx r3d, WORD [blockq] + mov blockd, r3d ; dc + shl blockd, 4 ; 16 * dc + lea blockd, [blockq+r3+4] ; 17 * dc + 4 + sar blockd, 3 ; >> 3 + shl blockd, 2 ; 4 * dc + lea blockd, [blockq*3+64] ; 12 * dc + 64 + sar blockd, 7 ; >> 7 + + INV_TRANS_INIT + + INV_TRANS_PROCESS h + lea destq, [destq+linesizeq*4] + INV_TRANS_PROCESS h + RET + +INIT_MMX mmxext +cglobal vc1_inv_trans_8x4_dc, 3,4,0, dest, linesize, block + movsx blockd, WORD [blockq] ; dc + lea blockd, [blockq*3+1] ; 3 * dc + 1 + sar blockd, 1 ; >> 1 + mov r3d, blockd ; dc + shl blockd, 4 ; 16 * dc + lea blockd, [blockq+r3+64] ; 17 * dc + 64 + sar blockd, 7 ; >> 7 + + INV_TRANS_INIT + + INV_TRANS_PROCESS a + RET + +INIT_MMX mmxext +cglobal vc1_inv_trans_8x8_dc, 3,3,0, dest, linesize, block + movsx blockd, WORD [blockq] ; dc + lea blockd, [blockq*3+1] ; 3 * dc + 1 + sar blockd, 1 ; >> 1 + lea blockd, [blockq*3+16] ; 3 * dc + 16 + sar blockd, 5 ; >> 5 + + INV_TRANS_INIT + + INV_TRANS_PROCESS a + lea destq, [destq+linesizeq*4] + INV_TRANS_PROCESS a + RET -- 1.9.1 _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org http://ffmpeg.org/mailman/listinfo/ffmpeg-devel