;****************************************************************************** ;* Pixel utilities SIMD ;* ;* Copyright (C) 2002-2004 Michael Niedermayer ;* Copyright (C) 2014 Clément Bœsch ;* ;* This file is part of FFmpeg. ;* ;* FFmpeg is free software; you can redistribute it and/or ;* modify it under the terms of the GNU Lesser General Public ;* License as published by the Free Software Foundation; either ;* version 2.1 of the License, or (at your option) any later version. ;* ;* FFmpeg is distributed in the hope that it will be useful, ;* but WITHOUT ANY WARRANTY; without even the implied warranty of ;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ;* Lesser General Public License for more details. ;* ;* You should have received a copy of the GNU Lesser General Public ;* License along with FFmpeg; if not, write to the Free Software ;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA ;****************************************************************************** %include "libavutil/x86/x86util.asm" SECTION .text ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_8x8_mmxext(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- INIT_MMX mmxext cglobal pixelutils_sad_8x8, 4,4,0, src1, stride1, src2, stride2 pxor m2, m2 %rep 4 mova m0, [src1q] mova m1, [src1q + stride1q] psadbw m0, [src2q] psadbw m1, [src2q + stride2q] paddw m2, m0 paddw m2, m1 lea src1q, [src1q + 2*stride1q] lea src2q, [src2q + 2*stride2q] %endrep movd eax, m2 emms RET ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- INIT_XMM sse2 cglobal pixelutils_sad_16x16, 4,4,5, src1, stride1, src2, stride2 movu m4, [src1q] movu m2, [src2q] movu m1, [src1q + stride1q] movu m3, [src2q + stride2q] psadbw m4, m2 psadbw m1, m3 paddw m4, m1 %rep 7 lea src1q, [src1q + 2*stride1q] lea src2q, [src2q + 2*stride2q] movu m0, [src1q] movu m2, [src2q] movu m1, [src1q + stride1q] movu m3, [src2q + stride2q] psadbw m0, m2 psadbw m1, m3 paddw m4, m0 paddw m4, m1 %endrep movhlps m0, m4 paddw m4, m0 movd eax, m4 RET ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_[au]_16x16_sse2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- %macro SAD_XMM_16x16 1 INIT_XMM sse2 cglobal pixelutils_sad_%1_16x16, 4,4,3, src1, stride1, src2, stride2 mov%1 m2, [src2q] psadbw m2, [src1q] mov%1 m1, [src2q + stride2q] psadbw m1, [src1q + stride1q] paddw m2, m1 %rep 7 lea src1q, [src1q + 2*stride1q] lea src2q, [src2q + 2*stride2q] mov%1 m0, [src2q] psadbw m0, [src1q] mov%1 m1, [src2q + stride2q] psadbw m1, [src1q + stride1q] paddw m2, m0 paddw m2, m1 %endrep movhlps m0, m2 paddw m2, m0 movd eax, m2 RET %endmacro SAD_XMM_16x16 a SAD_XMM_16x16 u %macro PROCESS_SAD_32x4_U 0 movu m1, [r2] movu m2, [r2 + 16] movu m3, [r0] movu m4, [r0 + 16] psadbw m1, m3 psadbw m2, m4 paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] movu m1, [r2] movu m2, [r2 + 16] movu m3, [r0] movu m4, [r0 + 16] psadbw m1, m3 psadbw m2, m4 paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] movu m1, [r2] movu m2, [r2 + 16] movu m3, [r0] movu m4, [r0 + 16] psadbw m1, m3 psadbw m2, m4 paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] movu m1, [r2] movu m2, [r2 + 16] movu m3, [r0] movu m4, [r0 + 16] psadbw m1, m3 psadbw m2, m4 paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] %endmacro %macro PROCESS_SAD_32x4 1 mov%1 m1, [r2] mov%1 m2, [r2 + 16] psadbw m1, [r0] psadbw m2, [r0 + 16] paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] mov%1 m1, [r2] mov%1 m2, [r2 + 16] psadbw m1, [r0] psadbw m2, [r0 + 16] paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] mov%1 m1, [r2] mov%1 m2, [r2 + 16] psadbw m1, [r0] psadbw m2, [r0 + 16] paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] mov%1 m1, [r2] mov%1 m2, [r2 + 16] psadbw m1, [r0] psadbw m2, [r0 + 16] paddd m1, m2 paddd m0, m1 lea r2, [r2 + r3] lea r0, [r0 + r1] %endmacro ;----------------------------------------------------------------------------- ; int ff_pixelutils_sad_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;----------------------------------------------------------------------------- INIT_XMM sse2 cglobal pixelutils_sad_32x32, 4,5,5, src1, stride1, src2, stride2 pxor m0, m0 mov r4d, 4 .loop: PROCESS_SAD_32x4_U PROCESS_SAD_32x4_U dec r4d jnz .loop movhlps m1, m0 paddd m0, m1 movd eax, m0 RET ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_[au]_32x32_sse2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- %macro SAD_XMM_32x32 1 INIT_XMM sse2 cglobal pixelutils_sad_%1_32x32, 4,5,3, src1, stride1, src2, stride2 pxor m0, m0 mov r4d, 4 .loop: PROCESS_SAD_32x4 %1 PROCESS_SAD_32x4 %1 dec r4d jnz .loop movhlps m1, m0 paddd m0, m1 movd eax, m0 RET %endmacro SAD_XMM_32x32 a SAD_XMM_32x32 u %if HAVE_AVX2_EXTERNAL ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- INIT_YMM avx2 cglobal pixelutils_sad_32x32, 4,7,5, src1, stride1, src2, stride2 pxor m0, m0 mov r4d, 32/4 lea r5, [stride1q * 3] lea r6, [stride2q * 3] .loop: movu m1, [src1q] ; row 0 of pix0 movu m2, [src2q] ; row 0 of pix1 movu m3, [src1q + stride1q] ; row 1 of pix0 movu m4, [src2q + stride2q] ; row 1 of pix1 psadbw m1, m2 psadbw m3, m4 paddd m0, m1 paddd m0, m3 movu m1, [src1q + 2 * stride1q] ; row 2 of pix0 movu m2, [src2q + 2 * stride2q] ; row 2 of pix1 movu m3, [src1q + r5] ; row 3 of pix0 movu m4, [src2q + r6] ; row 3 of pix1 psadbw m1, m2 psadbw m3, m4 paddd m0, m1 paddd m0, m3 lea src2q, [src2q + 4 * stride2q] lea src1q, [src1q + 4 * stride1q] dec r4d jnz .loop vextracti128 xm1, m0, 1 paddd xm0, xm1 pshufd xm1, xm0, 2 paddd xm0, xm1 movd eax, xm0 RET ;------------------------------------------------------------------------------- ; int ff_pixelutils_sad_[au]_32x32_avx2(const uint8_t *src1, ptrdiff_t stride1, ; const uint8_t *src2, ptrdiff_t stride2); ;------------------------------------------------------------------------------- %macro SAD_AVX2_32x32 1 INIT_YMM avx2 cglobal pixelutils_sad_%1_32x32, 4,7,3, src1, stride1, src2, stride2 pxor m0, m0 mov r4d, 32/4 lea r5, [stride1q * 3] lea r6, [stride2q * 3] .loop: mov%1 m1, [src2q] ; row 0 of pix1 psadbw m1, [src1q] mov%1 m2, [src2q + stride2q] ; row 1 of pix1 psadbw m2, [src1q + stride1q] paddd m0, m1 paddd m0, m2 mov%1 m1, [src2q + 2 * stride2q] ; row 2 of pix1 psadbw m1, [src1q + 2 * stride1q] mov%1 m2, [src2q + r6] ; row 3 of pix1 psadbw m2, [src1q + r5] paddd m0, m1 paddd m0, m2 lea src2q, [src2q + 4 * stride2q] lea src1q, [src1q + 4 * stride1q] dec r4d jnz .loop vextracti128 xm1, m0, 1 paddd xm0, xm1 pshufd xm1, xm0, 2 paddd xm0, xm1 movd eax, xm0 RET %endmacro SAD_AVX2_32x32 a SAD_AVX2_32x32 u %endif