2045 lines
62 KiB
NASM
2045 lines
62 KiB
NASM
;******************************************************************************
|
|
;* VP9 Intra prediction SIMD optimizations
|
|
;*
|
|
;* Copyright (c) 2013 Ronald S. Bultje <rsbultje gmail com>
|
|
;*
|
|
;* Parts based on:
|
|
;* H.264 intra prediction asm optimizations
|
|
;* Copyright (c) 2010 Fiona Glaser
|
|
;* Copyright (c) 2010 Holger Lubitz
|
|
;* Copyright (c) 2010 Loren Merritt
|
|
;* Copyright (c) 2010 Ronald S. Bultje
|
|
;*
|
|
;* This file is part of FFmpeg.
|
|
;*
|
|
;* FFmpeg is free software; you can redistribute it and/or
|
|
;* modify it under the terms of the GNU Lesser General Public
|
|
;* License as published by the Free Software Foundation; either
|
|
;* version 2.1 of the License, or (at your option) any later version.
|
|
;*
|
|
;* FFmpeg is distributed in the hope that it will be useful,
|
|
;* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
;* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
;* Lesser General Public License for more details.
|
|
;*
|
|
;* You should have received a copy of the GNU Lesser General Public
|
|
;* License along with FFmpeg; if not, write to the Free Software
|
|
;* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
;******************************************************************************
|
|
|
|
%include "libavutil/x86/x86util.asm"
|
|
|
|
SECTION_RODATA 32
|
|
|
|
pw_m256: times 16 dw -256
|
|
pw_m255: times 16 dw -255
|
|
pw_4096: times 8 dw 4096
|
|
|
|
pb_4x3_4x2_4x1_4x0: times 4 db 3
|
|
times 4 db 2
|
|
times 4 db 1
|
|
times 4 db 0
|
|
pb_8x1_8x0: times 8 db 1
|
|
times 8 db 0
|
|
pb_8x3_8x2: times 8 db 3
|
|
times 8 db 2
|
|
pb_0to5_2x7: db 0, 1, 2, 3, 4, 5, 7, 7
|
|
times 8 db -1
|
|
pb_0to6_9x7: db 0, 1, 2, 3, 4, 5, 6
|
|
times 9 db 7
|
|
pb_1to6_10x7: db 1, 2, 3, 4, 5, 6
|
|
times 10 db 7
|
|
pb_2to6_3x7:
|
|
pb_2to6_11x7: db 2, 3, 4, 5, 6
|
|
times 11 db 7
|
|
pb_1toE_2xF: db 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 15
|
|
pb_2toE_3xF: db 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 15, 15
|
|
pb_13456_3xm1: db 1, 3, 4, 5, 6
|
|
times 3 db -1
|
|
pb_6012_4xm1: db 6, 0, 1, 2
|
|
times 4 db -1
|
|
pb_6xm1_246_8toE: times 6 db -1
|
|
db 2, 4, 6, 8, 9, 10, 11, 12, 13, 14
|
|
pb_6xm1_BDF_0to6: times 6 db -1
|
|
db 11, 13, 15, 0, 1, 2, 3, 4, 5, 6
|
|
pb_02468ACE_13579BDF: db 0, 2, 4, 6, 8, 10, 12, 14, 1, 3, 5, 7, 9, 11, 13, 15
|
|
|
|
pb_15x0_1xm1: times 15 db 0
|
|
db -1
|
|
pb_0to2_5x3: db 0, 1, 2
|
|
times 5 db 3
|
|
pb_6xm1_2x0: times 6 db -1
|
|
times 2 db 0
|
|
pb_6x0_2xm1: times 6 db 0
|
|
times 2 db -1
|
|
|
|
cextern pb_1
|
|
cextern pb_2
|
|
cextern pb_3
|
|
cextern pb_15
|
|
cextern pw_2
|
|
cextern pw_4
|
|
cextern pw_8
|
|
cextern pw_16
|
|
cextern pw_32
|
|
cextern pw_255
|
|
cextern pw_512
|
|
cextern pw_1024
|
|
cextern pw_2048
|
|
cextern pw_8192
|
|
|
|
SECTION .text
|
|
|
|
; dc_NxN(uint8_t *dst, ptrdiff_t stride, const uint8_t *l, const uint8_t *a)
|
|
|
|
%macro DC_4to8_FUNCS 0
|
|
cglobal vp9_ipred_dc_4x4, 4, 4, 0, dst, stride, l, a
|
|
movd m0, [lq]
|
|
punpckldq m0, [aq]
|
|
pxor m1, m1
|
|
psadbw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_4096]
|
|
pshufb m0, m1
|
|
%else
|
|
paddw m0, [pw_4]
|
|
psraw m0, 3
|
|
punpcklbw m0, m0
|
|
pshufw m0, m0, q0000
|
|
%endif
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*1], m0
|
|
lea dstq, [dstq+strideq*2]
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*1], m0
|
|
RET
|
|
|
|
cglobal vp9_ipred_dc_8x8, 4, 4, 0, dst, stride, l, a
|
|
movq m0, [lq]
|
|
movq m1, [aq]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
psadbw m1, m2
|
|
paddw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_2048]
|
|
pshufb m0, m2
|
|
%else
|
|
paddw m0, [pw_8]
|
|
psraw m0, 4
|
|
punpcklbw m0, m0
|
|
pshufw m0, m0, q0000
|
|
%endif
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
DC_4to8_FUNCS
|
|
INIT_MMX ssse3
|
|
DC_4to8_FUNCS
|
|
|
|
%macro DC_16to32_FUNCS 0
|
|
cglobal vp9_ipred_dc_16x16, 4, 4, 3, dst, stride, l, a
|
|
mova m0, [lq]
|
|
mova m1, [aq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
psadbw m1, m2
|
|
paddw m0, m1
|
|
movhlps m1, m0
|
|
paddw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_1024]
|
|
pshufb m0, m2
|
|
%else
|
|
paddw m0, [pw_16]
|
|
psraw m0, 5
|
|
punpcklbw m0, m0
|
|
pshuflw m0, m0, q0000
|
|
punpcklqdq m0, m0
|
|
%endif
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_dc_32x32, 4, 4, 5, dst, stride, l, a
|
|
mova m0, [lq]
|
|
mova m1, [lq+16]
|
|
mova m2, [aq]
|
|
mova m3, [aq+16]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m4, m4
|
|
psadbw m0, m4
|
|
psadbw m1, m4
|
|
psadbw m2, m4
|
|
psadbw m3, m4
|
|
paddw m0, m1
|
|
paddw m2, m3
|
|
paddw m0, m2
|
|
movhlps m1, m0
|
|
paddw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_512]
|
|
pshufb m0, m4
|
|
%else
|
|
paddw m0, [pw_32]
|
|
psraw m0, 6
|
|
punpcklbw m0, m0
|
|
pshuflw m0, m0, q0000
|
|
punpcklqdq m0, m0
|
|
%endif
|
|
mov cntd, 8
|
|
.loop:
|
|
mova [dstq+strideq*0+ 0], m0
|
|
mova [dstq+strideq*0+16], m0
|
|
mova [dstq+strideq*1+ 0], m0
|
|
mova [dstq+strideq*1+16], m0
|
|
mova [dstq+strideq*2+ 0], m0
|
|
mova [dstq+strideq*2+16], m0
|
|
mova [dstq+stride3q + 0], m0
|
|
mova [dstq+stride3q +16], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DC_16to32_FUNCS
|
|
INIT_XMM ssse3
|
|
DC_16to32_FUNCS
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
cglobal vp9_ipred_dc_32x32, 4, 4, 3, dst, stride, l, a
|
|
mova m0, [lq]
|
|
mova m1, [aq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
psadbw m1, m2
|
|
paddw m0, m1
|
|
vextracti128 xm1, m0, 1
|
|
paddw xm0, xm1
|
|
movhlps xm1, xm0
|
|
paddw xm0, xm1
|
|
pmulhrsw xm0, [pw_512]
|
|
vpbroadcastb m0, xm0
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endif
|
|
|
|
; dc_top/left_NxN(uint8_t *dst, ptrdiff_t stride, const uint8_t *l, const uint8_t *a)
|
|
|
|
%macro DC_1D_4to8_FUNCS 2 ; dir (top or left), arg (a or l)
|
|
cglobal vp9_ipred_dc_%1_4x4, 4, 4, 0, dst, stride, l, a
|
|
movd m0, [%2q]
|
|
pxor m1, m1
|
|
psadbw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_8192]
|
|
pshufb m0, m1
|
|
%else
|
|
paddw m0, [pw_2]
|
|
psraw m0, 2
|
|
punpcklbw m0, m0
|
|
pshufw m0, m0, q0000
|
|
%endif
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*1], m0
|
|
lea dstq, [dstq+strideq*2]
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*1], m0
|
|
RET
|
|
|
|
cglobal vp9_ipred_dc_%1_8x8, 4, 4, 0, dst, stride, l, a
|
|
movq m0, [%2q]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
pxor m1, m1
|
|
psadbw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_4096]
|
|
pshufb m0, m1
|
|
%else
|
|
paddw m0, [pw_4]
|
|
psraw m0, 3
|
|
punpcklbw m0, m0
|
|
pshufw m0, m0, q0000
|
|
%endif
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
DC_1D_4to8_FUNCS top, a
|
|
DC_1D_4to8_FUNCS left, l
|
|
INIT_MMX ssse3
|
|
DC_1D_4to8_FUNCS top, a
|
|
DC_1D_4to8_FUNCS left, l
|
|
|
|
%macro DC_1D_16to32_FUNCS 2; dir (top or left), arg (a or l)
|
|
cglobal vp9_ipred_dc_%1_16x16, 4, 4, 3, dst, stride, l, a
|
|
mova m0, [%2q]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
movhlps m1, m0
|
|
paddw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_2048]
|
|
pshufb m0, m2
|
|
%else
|
|
paddw m0, [pw_8]
|
|
psraw m0, 4
|
|
punpcklbw m0, m0
|
|
pshuflw m0, m0, q0000
|
|
punpcklqdq m0, m0
|
|
%endif
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_dc_%1_32x32, 4, 4, 3, dst, stride, l, a
|
|
mova m0, [%2q]
|
|
mova m1, [%2q+16]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
psadbw m1, m2
|
|
paddw m0, m1
|
|
movhlps m1, m0
|
|
paddw m0, m1
|
|
%if cpuflag(ssse3)
|
|
pmulhrsw m0, [pw_1024]
|
|
pshufb m0, m2
|
|
%else
|
|
paddw m0, [pw_16]
|
|
psraw m0, 5
|
|
punpcklbw m0, m0
|
|
pshuflw m0, m0, q0000
|
|
punpcklqdq m0, m0
|
|
%endif
|
|
mov cntd, 8
|
|
.loop:
|
|
mova [dstq+strideq*0+ 0], m0
|
|
mova [dstq+strideq*0+16], m0
|
|
mova [dstq+strideq*1+ 0], m0
|
|
mova [dstq+strideq*1+16], m0
|
|
mova [dstq+strideq*2+ 0], m0
|
|
mova [dstq+strideq*2+16], m0
|
|
mova [dstq+stride3q + 0], m0
|
|
mova [dstq+stride3q +16], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DC_1D_16to32_FUNCS top, a
|
|
DC_1D_16to32_FUNCS left, l
|
|
INIT_XMM ssse3
|
|
DC_1D_16to32_FUNCS top, a
|
|
DC_1D_16to32_FUNCS left, l
|
|
|
|
%macro DC_1D_AVX2_FUNCS 2 ; dir (top or left), arg (a or l)
|
|
%if HAVE_AVX2_EXTERNAL
|
|
cglobal vp9_ipred_dc_%1_32x32, 4, 4, 3, dst, stride, l, a
|
|
mova m0, [%2q]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
pxor m2, m2
|
|
psadbw m0, m2
|
|
vextracti128 xm1, m0, 1
|
|
paddw xm0, xm1
|
|
movhlps xm1, xm0
|
|
paddw xm0, xm1
|
|
pmulhrsw xm0, [pw_1024]
|
|
vpbroadcastb m0, xm0
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endif
|
|
%endmacro
|
|
|
|
INIT_YMM avx2
|
|
DC_1D_AVX2_FUNCS top, a
|
|
DC_1D_AVX2_FUNCS left, l
|
|
|
|
; v
|
|
|
|
INIT_MMX mmx
|
|
cglobal vp9_ipred_v_8x8, 4, 4, 0, dst, stride, l, a
|
|
movq m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+strideq*2], m0
|
|
movq [dstq+stride3q ], m0
|
|
RET
|
|
|
|
INIT_XMM sse
|
|
cglobal vp9_ipred_v_16x16, 4, 4, 1, dst, stride, l, a
|
|
mova m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
INIT_XMM sse
|
|
cglobal vp9_ipred_v_32x32, 4, 4, 2, dst, stride, l, a
|
|
mova m0, [aq]
|
|
mova m1, [aq+16]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
mov cntd, 8
|
|
.loop:
|
|
mova [dstq+strideq*0+ 0], m0
|
|
mova [dstq+strideq*0+16], m1
|
|
mova [dstq+strideq*1+ 0], m0
|
|
mova [dstq+strideq*1+16], m1
|
|
mova [dstq+strideq*2+ 0], m0
|
|
mova [dstq+strideq*2+16], m1
|
|
mova [dstq+stride3q + 0], m0
|
|
mova [dstq+stride3q +16], m1
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
INIT_YMM avx
|
|
cglobal vp9_ipred_v_32x32, 4, 4, 1, dst, stride, l, a
|
|
mova m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+strideq*2], m0
|
|
mova [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
; h
|
|
|
|
%macro H_XMM_FUNCS 2
|
|
%if notcpuflag(avx)
|
|
cglobal vp9_ipred_h_4x4, 3, 4, 1, dst, stride, l, stride3
|
|
movd m0, [lq]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, [pb_4x3_4x2_4x1_4x0]
|
|
%else
|
|
punpcklbw m0, m0
|
|
pshuflw m0, m0, q0123
|
|
punpcklwd m0, m0
|
|
%endif
|
|
lea stride3q, [strideq*3]
|
|
movd [dstq+strideq*0], m0
|
|
psrldq m0, 4
|
|
movd [dstq+strideq*1], m0
|
|
psrldq m0, 4
|
|
movd [dstq+strideq*2], m0
|
|
psrldq m0, 4
|
|
movd [dstq+stride3q ], m0
|
|
RET
|
|
%endif
|
|
|
|
cglobal vp9_ipred_h_8x8, 3, 5, %1, dst, stride, l, stride3, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m2, [pb_8x1_8x0]
|
|
mova m3, [pb_8x3_8x2]
|
|
%endif
|
|
lea stride3q, [strideq*3]
|
|
mov cntq, 1
|
|
.loop:
|
|
movd m0, [lq+cntq*4]
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, m0, m3
|
|
pshufb m0, m2
|
|
%else
|
|
punpcklbw m0, m0
|
|
punpcklwd m0, m0
|
|
pshufd m1, m0, q2233
|
|
pshufd m0, m0, q0011
|
|
%endif
|
|
movq [dstq+strideq*0], m1
|
|
movhps [dstq+strideq*1], m1
|
|
movq [dstq+strideq*2], m0
|
|
movhps [dstq+stride3q ], m0
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_h_16x16, 3, 5, %2, dst, stride, l, stride3, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m5, [pb_1]
|
|
mova m6, [pb_2]
|
|
mova m7, [pb_3]
|
|
pxor m4, m4
|
|
%endif
|
|
lea stride3q, [strideq*3]
|
|
mov cntq, 3
|
|
.loop:
|
|
movd m3, [lq+cntq*4]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, m3, m7
|
|
pshufb m1, m3, m6
|
|
%else
|
|
punpcklbw m3, m3
|
|
punpcklwd m3, m3
|
|
pshufd m0, m3, q3333
|
|
pshufd m1, m3, q2222
|
|
%endif
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m1
|
|
%if cpuflag(ssse3)
|
|
pshufb m2, m3, m5
|
|
pshufb m3, m4
|
|
%else
|
|
pshufd m2, m3, q1111
|
|
pshufd m3, m3, q0000
|
|
%endif
|
|
mova [dstq+strideq*2], m2
|
|
mova [dstq+stride3q ], m3
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_h_32x32, 3, 5, %2, dst, stride, l, stride3, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m5, [pb_1]
|
|
mova m6, [pb_2]
|
|
mova m7, [pb_3]
|
|
pxor m4, m4
|
|
%endif
|
|
lea stride3q, [strideq*3]
|
|
mov cntq, 7
|
|
.loop:
|
|
movd m3, [lq+cntq*4]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, m3, m7
|
|
pshufb m1, m3, m6
|
|
%else
|
|
punpcklbw m3, m3
|
|
punpcklwd m3, m3
|
|
pshufd m0, m3, q3333
|
|
pshufd m1, m3, q2222
|
|
%endif
|
|
mova [dstq+strideq*0+ 0], m0
|
|
mova [dstq+strideq*0+16], m0
|
|
mova [dstq+strideq*1+ 0], m1
|
|
mova [dstq+strideq*1+16], m1
|
|
%if cpuflag(ssse3)
|
|
pshufb m2, m3, m5
|
|
pshufb m3, m4
|
|
%else
|
|
pshufd m2, m3, q1111
|
|
pshufd m3, m3, q0000
|
|
%endif
|
|
mova [dstq+strideq*2+ 0], m2
|
|
mova [dstq+strideq*2+16], m2
|
|
mova [dstq+stride3q + 0], m3
|
|
mova [dstq+stride3q +16], m3
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
H_XMM_FUNCS 2, 4
|
|
INIT_XMM ssse3
|
|
H_XMM_FUNCS 4, 8
|
|
INIT_XMM avx
|
|
H_XMM_FUNCS 4, 8
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
cglobal vp9_ipred_h_32x32, 3, 5, 8, dst, stride, l, stride3, cnt
|
|
mova m5, [pb_1]
|
|
mova m6, [pb_2]
|
|
mova m7, [pb_3]
|
|
pxor m4, m4
|
|
lea stride3q, [strideq*3]
|
|
mov cntq, 7
|
|
.loop:
|
|
movd xm3, [lq+cntq*4]
|
|
vinserti128 m3, m3, xm3, 1
|
|
pshufb m0, m3, m7
|
|
pshufb m1, m3, m6
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m1
|
|
pshufb m2, m3, m5
|
|
pshufb m3, m4
|
|
mova [dstq+strideq*2], m2
|
|
mova [dstq+stride3q ], m3
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
%endif
|
|
|
|
; tm
|
|
|
|
%macro TM_MMX_FUNCS 0
|
|
cglobal vp9_ipred_tm_4x4, 4, 4, 0, dst, stride, l, a
|
|
pxor m1, m1
|
|
movd m0, [aq]
|
|
pinsrw m2, [aq-1], 0
|
|
punpcklbw m0, m1
|
|
DEFINE_ARGS dst, stride, l, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m3, [pw_m256]
|
|
mova m1, [pw_m255]
|
|
pshufb m2, m3
|
|
%else
|
|
punpcklbw m2, m1
|
|
pshufw m2, m2, q0000
|
|
%endif
|
|
psubw m0, m2
|
|
mov cntq, 1
|
|
.loop:
|
|
pinsrw m2, [lq+cntq*2], 0
|
|
%if cpuflag(ssse3)
|
|
pshufb m4, m2, m1
|
|
pshufb m2, m3
|
|
%else
|
|
punpcklbw m2, m1
|
|
pshufw m4, m2, q1111
|
|
pshufw m2, m2, q0000
|
|
%endif
|
|
paddw m4, m0
|
|
paddw m2, m0
|
|
packuswb m4, m4
|
|
packuswb m2, m2
|
|
movd [dstq+strideq*0], m4
|
|
movd [dstq+strideq*1], m2
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
TM_MMX_FUNCS
|
|
INIT_MMX ssse3
|
|
TM_MMX_FUNCS
|
|
|
|
%macro TM_XMM_FUNCS 0
|
|
cglobal vp9_ipred_tm_8x8, 4, 4, 5, dst, stride, l, a
|
|
pxor m1, m1
|
|
movh m0, [aq]
|
|
pinsrw m2, [aq-1], 0
|
|
punpcklbw m0, m1
|
|
DEFINE_ARGS dst, stride, l, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m3, [pw_m256]
|
|
mova m1, [pw_m255]
|
|
pshufb m2, m3
|
|
%else
|
|
punpcklbw m2, m1
|
|
punpcklwd m2, m2
|
|
pshufd m2, m2, q0000
|
|
%endif
|
|
psubw m0, m2
|
|
mov cntq, 3
|
|
.loop:
|
|
pinsrw m2, [lq+cntq*2], 0
|
|
%if cpuflag(ssse3)
|
|
pshufb m4, m2, m1
|
|
pshufb m2, m3
|
|
%else
|
|
punpcklbw m2, m1
|
|
punpcklwd m2, m2
|
|
pshufd m4, m2, q1111
|
|
pshufd m2, m2, q0000
|
|
%endif
|
|
paddw m4, m0
|
|
paddw m2, m0
|
|
packuswb m4, m2
|
|
movh [dstq+strideq*0], m4
|
|
movhps [dstq+strideq*1], m4
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_tm_16x16, 4, 4, 8, dst, stride, l, a
|
|
pxor m3, m3
|
|
mova m0, [aq]
|
|
pinsrw m2, [aq-1], 0
|
|
punpckhbw m1, m0, m3
|
|
punpcklbw m0, m3
|
|
DEFINE_ARGS dst, stride, l, cnt
|
|
%if cpuflag(ssse3)
|
|
mova m4, [pw_m256]
|
|
mova m3, [pw_m255]
|
|
pshufb m2, m4
|
|
%else
|
|
punpcklbw m2, m3
|
|
punpcklwd m2, m2
|
|
pshufd m2, m2, q0000
|
|
%endif
|
|
psubw m1, m2
|
|
psubw m0, m2
|
|
mov cntq, 7
|
|
.loop:
|
|
pinsrw m7, [lq+cntq*2], 0
|
|
%if cpuflag(ssse3)
|
|
pshufb m5, m7, m3
|
|
pshufb m7, m4
|
|
%else
|
|
punpcklbw m7, m3
|
|
punpcklwd m7, m7
|
|
pshufd m5, m7, q1111
|
|
pshufd m7, m7, q0000
|
|
%endif
|
|
paddw m2, m5, m0
|
|
paddw m5, m1
|
|
paddw m6, m7, m0
|
|
paddw m7, m1
|
|
packuswb m2, m5
|
|
packuswb m6, m7
|
|
mova [dstq+strideq*0], m2
|
|
mova [dstq+strideq*1], m6
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
|
|
%if ARCH_X86_64
|
|
%define mem 0
|
|
%else
|
|
%define mem 64
|
|
%endif
|
|
cglobal vp9_ipred_tm_32x32, 4, 4, 14, mem, dst, stride, l, a
|
|
pxor m5, m5
|
|
pinsrw m4, [aq-1], 0
|
|
mova m0, [aq]
|
|
mova m2, [aq+16]
|
|
DEFINE_ARGS dst, stride, l, cnt
|
|
%if cpuflag(ssse3)
|
|
%if ARCH_X86_64
|
|
mova m12, [pw_m256]
|
|
mova m13, [pw_m255]
|
|
%define pw_m256_reg m12
|
|
%define pw_m255_reg m13
|
|
%else
|
|
%define pw_m256_reg [pw_m256]
|
|
%define pw_m255_reg [pw_m255]
|
|
%endif
|
|
pshufb m4, pw_m256_reg
|
|
%else
|
|
punpcklbw m4, m5
|
|
punpcklwd m4, m4
|
|
pshufd m4, m4, q0000
|
|
%endif
|
|
punpckhbw m1, m0, m5
|
|
punpckhbw m3, m2, m5
|
|
punpcklbw m0, m5
|
|
punpcklbw m2, m5
|
|
psubw m1, m4
|
|
psubw m0, m4
|
|
psubw m3, m4
|
|
psubw m2, m4
|
|
%if ARCH_X86_64
|
|
SWAP 0, 8
|
|
SWAP 1, 9
|
|
SWAP 2, 10
|
|
SWAP 3, 11
|
|
%else
|
|
mova [rsp+0*16], m0
|
|
mova [rsp+1*16], m1
|
|
mova [rsp+2*16], m2
|
|
mova [rsp+3*16], m3
|
|
%endif
|
|
mov cntq, 15
|
|
.loop:
|
|
pinsrw m3, [lq+cntq*2], 0
|
|
%if cpuflag(ssse3)
|
|
pshufb m7, m3, pw_m255_reg
|
|
pshufb m3, pw_m256_reg
|
|
%else
|
|
pxor m7, m7
|
|
punpcklbw m3, m7
|
|
punpcklwd m3, m3
|
|
pshufd m7, m3, q1111
|
|
pshufd m3, m3, q0000
|
|
%endif
|
|
%if ARCH_X86_64
|
|
paddw m4, m7, m8
|
|
paddw m5, m7, m9
|
|
paddw m6, m7, m10
|
|
paddw m7, m11
|
|
paddw m0, m3, m8
|
|
paddw m1, m3, m9
|
|
paddw m2, m3, m10
|
|
paddw m3, m11
|
|
%else
|
|
paddw m4, m7, [rsp+0*16]
|
|
paddw m5, m7, [rsp+1*16]
|
|
paddw m6, m7, [rsp+2*16]
|
|
paddw m7, [rsp+3*16]
|
|
paddw m0, m3, [rsp+0*16]
|
|
paddw m1, m3, [rsp+1*16]
|
|
paddw m2, m3, [rsp+2*16]
|
|
paddw m3, [rsp+3*16]
|
|
%endif
|
|
packuswb m4, m5
|
|
packuswb m6, m7
|
|
packuswb m0, m1
|
|
packuswb m2, m3
|
|
mova [dstq+strideq*0+ 0], m4
|
|
mova [dstq+strideq*0+16], m6
|
|
mova [dstq+strideq*1+ 0], m0
|
|
mova [dstq+strideq*1+16], m2
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
%undef pw_m256_reg
|
|
%undef pw_m255_reg
|
|
%undef mem
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
TM_XMM_FUNCS
|
|
INIT_XMM ssse3
|
|
TM_XMM_FUNCS
|
|
INIT_XMM avx
|
|
TM_XMM_FUNCS
|
|
|
|
%if HAVE_AVX2_EXTERNAL
|
|
INIT_YMM avx2
|
|
cglobal vp9_ipred_tm_32x32, 4, 4, 8, dst, stride, l, a
|
|
pxor m3, m3
|
|
pinsrw xm2, [aq-1], 0
|
|
vinserti128 m2, m2, xm2, 1
|
|
mova m0, [aq]
|
|
DEFINE_ARGS dst, stride, l, cnt
|
|
mova m4, [pw_m256]
|
|
mova m5, [pw_m255]
|
|
pshufb m2, m4
|
|
punpckhbw m1, m0, m3
|
|
punpcklbw m0, m3
|
|
psubw m1, m2
|
|
psubw m0, m2
|
|
mov cntq, 15
|
|
.loop:
|
|
pinsrw xm7, [lq+cntq*2], 0
|
|
vinserti128 m7, m7, xm7, 1
|
|
pshufb m3, m7, m5
|
|
pshufb m7, m4
|
|
paddw m2, m3, m0
|
|
paddw m3, m1
|
|
paddw m6, m7, m0
|
|
paddw m7, m1
|
|
packuswb m2, m3
|
|
packuswb m6, m7
|
|
mova [dstq+strideq*0], m2
|
|
mova [dstq+strideq*1], m6
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntq
|
|
jge .loop
|
|
RET
|
|
%endif
|
|
|
|
; dl
|
|
|
|
%macro LOWPASS 4 ; left [dst], center, right, tmp
|
|
pxor m%4, m%1, m%3
|
|
pand m%4, [pb_1]
|
|
pavgb m%1, m%3
|
|
psubusb m%1, m%4
|
|
pavgb m%1, m%2
|
|
%endmacro
|
|
|
|
%macro DL_MMX_FUNCS 0
|
|
cglobal vp9_ipred_dl_4x4, 4, 4, 0, dst, stride, l, a
|
|
movq m1, [aq]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, m1, [pb_0to5_2x7]
|
|
pshufb m2, m1, [pb_2to6_3x7]
|
|
%else
|
|
punpckhbw m3, m1, m1 ; 44556677
|
|
pand m0, m1, [pb_6xm1_2x0] ; 012345__
|
|
pand m3, [pb_6x0_2xm1] ; ______77
|
|
psrlq m2, m1, 16 ; 234567__
|
|
por m0, m3 ; 01234577
|
|
por m2, m3 ; 23456777
|
|
%endif
|
|
psrlq m1, 8
|
|
LOWPASS 0, 1, 2, 3
|
|
|
|
pshufw m1, m0, q3321
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*2], m1
|
|
psrlq m0, 8
|
|
psrlq m1, 8
|
|
add dstq, strideq
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*2], m1
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
DL_MMX_FUNCS
|
|
INIT_MMX ssse3
|
|
DL_MMX_FUNCS
|
|
|
|
%macro DL_XMM_FUNCS 0
|
|
cglobal vp9_ipred_dl_8x8, 4, 4, 4, dst, stride, stride5, a
|
|
movq m0, [aq]
|
|
lea stride5q, [strideq*5]
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, m0, [pb_1to6_10x7]
|
|
%else
|
|
punpcklbw m1, m0, m0 ; 0011223344556677
|
|
punpckhwd m1, m1 ; 4x4,4x5,4x6,4x7
|
|
%endif
|
|
shufps m0, m1, q3310
|
|
%if notcpuflag(ssse3)
|
|
psrldq m1, m0, 1
|
|
shufps m1, m0, q3210
|
|
%endif
|
|
psrldq m2, m1, 1
|
|
LOWPASS 0, 1, 2, 3
|
|
|
|
pshufd m1, m0, q3321
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*4], m1
|
|
psrldq m0, 1
|
|
psrldq m1, 1
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+stride5q ], m1
|
|
lea dstq, [dstq+strideq*2]
|
|
psrldq m0, 1
|
|
psrldq m1, 1
|
|
movq [dstq+strideq*0], m0
|
|
movq [dstq+strideq*4], m1
|
|
psrldq m0, 1
|
|
psrldq m1, 1
|
|
movq [dstq+strideq*1], m0
|
|
movq [dstq+stride5q ], m1
|
|
RET
|
|
|
|
cglobal vp9_ipred_dl_16x16, 4, 4, 6, dst, stride, l, a
|
|
mova m0, [aq]
|
|
%if cpuflag(ssse3)
|
|
mova m5, [pb_1toE_2xF]
|
|
pshufb m1, m0, m5
|
|
pshufb m2, m1, m5
|
|
pshufb m4, m0, [pb_15]
|
|
%else
|
|
pand m5, m0, [pb_15x0_1xm1] ; _______________F
|
|
psrldq m1, m0, 1 ; 123456789ABCDEF_
|
|
por m1, m5 ; 123456789ABCDEFF
|
|
psrldq m2, m1, 1 ; 23456789ABCDEFF_
|
|
por m2, m5 ; 23456789ABCDEFFF
|
|
pshufhw m4, m1, q3333 ; xxxxxxxxFFFFFFFF
|
|
%endif
|
|
LOWPASS 0, 1, 2, 3
|
|
DEFINE_ARGS dst, stride, cnt, stride9
|
|
lea stride9q, [strideq+strideq*8]
|
|
mov cntd, 4
|
|
|
|
.loop:
|
|
movhlps m4, m0
|
|
mova [dstq+strideq*0], m0
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, m5
|
|
%else
|
|
psrldq m0, 1
|
|
por m0, m5
|
|
%endif
|
|
mova [dstq+strideq*8], m4
|
|
movhlps m4, m0
|
|
mova [dstq+strideq*1], m0
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, m5
|
|
%else
|
|
psrldq m0, 1
|
|
por m0, m5
|
|
%endif
|
|
mova [dstq+stride9q ], m4
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_dl_32x32, 4, 5, 8, dst, stride, cnt, a, dst16
|
|
mova m0, [aq]
|
|
mova m1, [aq+16]
|
|
PALIGNR m2, m1, m0, 1, m4
|
|
PALIGNR m3, m1, m0, 2, m4
|
|
LOWPASS 0, 2, 3, 4
|
|
%if cpuflag(ssse3)
|
|
mova m5, [pb_1toE_2xF]
|
|
pshufb m2, m1, m5
|
|
pshufb m3, m2, m5
|
|
pshufb m6, m1, [pb_15]
|
|
mova m7, m6
|
|
%else
|
|
pand m5, m1, [pb_15x0_1xm1] ; _______________F
|
|
psrldq m2, m1, 1 ; 123456789ABCDEF_
|
|
por m2, m5 ; 123456789ABCDEFF
|
|
psrldq m3, m2, 1 ; 23456789ABCDEFF_
|
|
por m3, m5 ; 23456789ABCDEFFF
|
|
pshufhw m7, m2, q3333 ; xxxxxxxxFFFFFFFF
|
|
pshufd m6, m7, q3333
|
|
%endif
|
|
LOWPASS 1, 2, 3, 4
|
|
lea dst16q, [dstq +strideq*8]
|
|
mov cntd, 8
|
|
lea dst16q, [dst16q+strideq*8]
|
|
.loop:
|
|
movhlps m7, m1
|
|
mova [dstq +strideq*0+ 0], m0
|
|
mova [dstq +strideq*0+16], m1
|
|
movhps [dstq+strideq*8+ 0], m0
|
|
movq [dstq +strideq*8+ 8], m1
|
|
mova [dstq +strideq*8+16], m7
|
|
mova [dst16q+strideq*0+ 0], m1
|
|
mova [dst16q+strideq*0+16], m6
|
|
mova [dst16q+strideq*8+ 0], m7
|
|
mova [dst16q+strideq*8+16], m6
|
|
%if cpuflag(avx)
|
|
vpalignr m0, m1, m0, 1
|
|
pshufb m1, m5
|
|
%elif cpuflag(ssse3)
|
|
palignr m2, m1, m0, 1
|
|
pshufb m1, m5
|
|
mova m0, m2
|
|
%else
|
|
mova m4, m1
|
|
psrldq m0, 1
|
|
pslldq m4, 15
|
|
psrldq m1, 1
|
|
por m0, m4
|
|
por m1, m5
|
|
%endif
|
|
add dstq, strideq
|
|
add dst16q, strideq
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DL_XMM_FUNCS
|
|
INIT_XMM ssse3
|
|
DL_XMM_FUNCS
|
|
INIT_XMM avx
|
|
DL_XMM_FUNCS
|
|
|
|
; dr
|
|
|
|
%macro DR_MMX_FUNCS 0
|
|
cglobal vp9_ipred_dr_4x4, 4, 4, 0, dst, stride, l, a
|
|
movd m0, [lq]
|
|
punpckldq m0, [aq-1]
|
|
movd m1, [aq+3]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
PALIGNR m1, m0, 1, m3
|
|
psrlq m2, m1, 8
|
|
LOWPASS 0, 1, 2, 3
|
|
|
|
movd [dstq+stride3q ], m0
|
|
psrlq m0, 8
|
|
movd [dstq+strideq*2], m0
|
|
psrlq m0, 8
|
|
movd [dstq+strideq*1], m0
|
|
psrlq m0, 8
|
|
movd [dstq+strideq*0], m0
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
DR_MMX_FUNCS
|
|
INIT_MMX ssse3
|
|
DR_MMX_FUNCS
|
|
|
|
%macro DR_XMM_FUNCS 0
|
|
cglobal vp9_ipred_dr_8x8, 4, 4, 4, dst, stride, l, a
|
|
movq m1, [lq]
|
|
movhps m1, [aq-1]
|
|
movd m2, [aq+7]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
pslldq m0, m1, 1
|
|
PALIGNR m2, m1, 1, m3
|
|
LOWPASS 0, 1, 2, 3
|
|
|
|
movhps [dstq+strideq*0], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+strideq*1], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+strideq*2], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+stride3q ], m0
|
|
pslldq m0, 1
|
|
lea dstq, [dstq+strideq*4]
|
|
movhps [dstq+strideq*0], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+strideq*1], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+strideq*2], m0
|
|
pslldq m0, 1
|
|
movhps [dstq+stride3q ], m0
|
|
RET
|
|
|
|
cglobal vp9_ipred_dr_16x16, 4, 4, 6, dst, stride, l, a
|
|
mova m1, [lq]
|
|
movu m2, [aq-1]
|
|
movd m4, [aq+15]
|
|
DEFINE_ARGS dst, stride, stride9, cnt
|
|
lea stride9q, [strideq *3]
|
|
mov cntd, 4
|
|
lea stride9q, [stride9q*3]
|
|
PALIGNR m4, m2, 1, m5
|
|
PALIGNR m3, m2, m1, 15, m5
|
|
LOWPASS 3, 2, 4, 5
|
|
pslldq m0, m1, 1
|
|
PALIGNR m2, m1, 1, m4
|
|
LOWPASS 0, 1, 2, 4
|
|
|
|
.loop:
|
|
mova [dstq+strideq*0 ], m3
|
|
movhps [dstq+strideq*8+0], m0
|
|
movq [dstq+strideq*8+8], m3
|
|
PALIGNR m3, m0, 15, m1
|
|
pslldq m0, 1
|
|
mova [dstq+strideq*1 ], m3
|
|
movhps [dstq+stride9q +0], m0
|
|
movq [dstq+stride9q +8], m3
|
|
PALIGNR m3, m0, 15, m1
|
|
pslldq m0, 1
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_dr_32x32, 4, 4, 8, dst, stride, l, a
|
|
mova m1, [lq]
|
|
mova m2, [lq+16]
|
|
movu m3, [aq-1]
|
|
movu m4, [aq+15]
|
|
movd m5, [aq+31]
|
|
DEFINE_ARGS dst, stride, stride8, cnt
|
|
lea stride8q, [strideq*8]
|
|
PALIGNR m5, m4, 1, m7
|
|
PALIGNR m6, m4, m3, 15, m7
|
|
LOWPASS 5, 4, 6, 7
|
|
PALIGNR m4, m3, 1, m7
|
|
PALIGNR m6, m3, m2, 15, m7
|
|
LOWPASS 4, 3, 6, 7
|
|
PALIGNR m3, m2, 1, m7
|
|
PALIGNR m6, m2, m1, 15, m7
|
|
LOWPASS 3, 2, 6, 7
|
|
PALIGNR m2, m1, 1, m6
|
|
pslldq m0, m1, 1
|
|
LOWPASS 2, 1, 0, 6
|
|
mov cntd, 16
|
|
|
|
; out=m2/m3/m4/m5
|
|
.loop:
|
|
mova [dstq+stride8q*0+ 0], m4
|
|
mova [dstq+stride8q*0+16], m5
|
|
mova [dstq+stride8q*2+ 0], m3
|
|
mova [dstq+stride8q*2+16], m4
|
|
PALIGNR m5, m4, 15, m6
|
|
PALIGNR m4, m3, 15, m6
|
|
PALIGNR m3, m2, 15, m6
|
|
pslldq m2, 1
|
|
add dstq, strideq
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
DR_XMM_FUNCS
|
|
INIT_XMM ssse3
|
|
DR_XMM_FUNCS
|
|
INIT_XMM avx
|
|
DR_XMM_FUNCS
|
|
|
|
; vl
|
|
|
|
INIT_MMX mmxext
|
|
cglobal vp9_ipred_vl_4x4, 4, 4, 0, dst, stride, l, a
|
|
movq m0, [aq]
|
|
psrlq m1, m0, 8
|
|
psrlq m2, m1, 8
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
movd [dstq+strideq*0], m1
|
|
movd [dstq+strideq*1], m2
|
|
lea dstq, [dstq+strideq*2]
|
|
psrlq m1, 8
|
|
psrlq m2, 8
|
|
movd [dstq+strideq*0], m1
|
|
movd [dstq+strideq*1], m2
|
|
RET
|
|
|
|
%macro VL_XMM_FUNCS 0
|
|
cglobal vp9_ipred_vl_8x8, 4, 4, 4, dst, stride, l, a
|
|
movq m0, [aq]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, [pb_0to6_9x7]
|
|
%else
|
|
punpcklbw m1, m0, m0
|
|
punpckhwd m1, m1
|
|
shufps m0, m1, q3310
|
|
%endif
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
psrldq m1, m0, 1
|
|
psrldq m2, m0, 2
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
|
|
movq [dstq+strideq*0], m1
|
|
movq [dstq+strideq*1], m2
|
|
psrldq m1, 1
|
|
psrldq m2, 1
|
|
movq [dstq+strideq*2], m1
|
|
movq [dstq+stride3q ], m2
|
|
lea dstq, [dstq+strideq*4]
|
|
psrldq m1, 1
|
|
psrldq m2, 1
|
|
movq [dstq+strideq*0], m1
|
|
movq [dstq+strideq*1], m2
|
|
psrldq m1, 1
|
|
psrldq m2, 1
|
|
movq [dstq+strideq*2], m1
|
|
movq [dstq+stride3q ], m2
|
|
RET
|
|
|
|
cglobal vp9_ipred_vl_16x16, 4, 4, 5, dst, stride, l, a
|
|
mova m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
%if cpuflag(ssse3)
|
|
mova m4, [pb_1toE_2xF]
|
|
pshufb m1, m0, m4
|
|
pshufb m2, m1, m4
|
|
%else
|
|
pand m4, m0, [pb_15x0_1xm1] ; _______________F
|
|
psrldq m1, m0, 1 ; 123456789ABCDEF_
|
|
por m1, m4 ; 123456789ABCDEFF
|
|
psrldq m2, m1, 1 ; 23456789ABCDEFF_
|
|
por m2, m4 ; 23456789ABCDEFFF
|
|
%endif
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
mov cntd, 4
|
|
.loop:
|
|
mova [dstq+strideq*0], m1
|
|
mova [dstq+strideq*1], m2
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, m4
|
|
pshufb m2, m4
|
|
%else
|
|
psrldq m1, 1
|
|
psrldq m2, 1
|
|
por m1, m4
|
|
por m2, m4
|
|
%endif
|
|
mova [dstq+strideq*2], m1
|
|
mova [dstq+stride3q ], m2
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, m4
|
|
pshufb m2, m4
|
|
%else
|
|
psrldq m1, 1
|
|
psrldq m2, 1
|
|
por m1, m4
|
|
por m2, m4
|
|
%endif
|
|
lea dstq, [dstq+strideq*4]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_vl_32x32, 4, 4, 7, dst, stride, l, a
|
|
mova m0, [aq]
|
|
mova m5, [aq+16]
|
|
DEFINE_ARGS dst, stride, dst16, cnt
|
|
PALIGNR m2, m5, m0, 1, m4
|
|
PALIGNR m3, m5, m0, 2, m4
|
|
lea dst16q, [dstq +strideq*8]
|
|
LOWPASS 3, 2, 0, 6
|
|
pavgb m2, m0
|
|
%if cpuflag(ssse3)
|
|
mova m4, [pb_1toE_2xF]
|
|
pshufb m0, m5, m4
|
|
pshufb m1, m0, m4
|
|
%else
|
|
pand m4, m5, [pb_15x0_1xm1] ; _______________F
|
|
psrldq m0, m5, 1 ; 123456789ABCDEF_
|
|
por m0, m4 ; 123456789ABCDEFF
|
|
psrldq m1, m0, 1 ; 23456789ABCDEFF_
|
|
por m1, m4 ; 23456789ABCDEFFF
|
|
%endif
|
|
lea dst16q, [dst16q+strideq*8]
|
|
LOWPASS 1, 0, 5, 6
|
|
pavgb m0, m5
|
|
%if cpuflag(ssse3)
|
|
pshufb m5, [pb_15]
|
|
%else
|
|
punpckhbw m5, m4, m4
|
|
pshufhw m5, m5, q3333
|
|
punpckhqdq m5, m5
|
|
%endif
|
|
mov cntd, 8
|
|
|
|
.loop:
|
|
%macro %%write 3
|
|
mova [dstq+stride%1+ 0], %2
|
|
mova [dstq+stride%1+16], %3
|
|
movhps [dst16q+stride%1 ], %2
|
|
movu [dst16q+stride%1+ 8], %3
|
|
movq [dst16q+stride%1+24], m5
|
|
%if cpuflag(avx)
|
|
palignr %2, %3, %2, 1
|
|
pshufb %3, m4
|
|
%elif cpuflag(ssse3)
|
|
palignr m6, %3, %2, 1
|
|
pshufb %3, m4
|
|
mova %2, m6
|
|
%else
|
|
pslldq m6, %3, 15
|
|
psrldq %3, 1
|
|
psrldq %2, 1
|
|
por %3, m4
|
|
por %2, m6
|
|
%endif
|
|
%endmacro
|
|
|
|
%%write q*0, m2, m0
|
|
%%write q*1, m3, m1
|
|
lea dstq, [dstq +strideq*2]
|
|
lea dst16q, [dst16q+strideq*2]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
VL_XMM_FUNCS
|
|
INIT_XMM ssse3
|
|
VL_XMM_FUNCS
|
|
INIT_XMM avx
|
|
VL_XMM_FUNCS
|
|
|
|
; vr
|
|
|
|
%macro VR_MMX_FUNCS 0
|
|
cglobal vp9_ipred_vr_4x4, 4, 4, 0, dst, stride, l, a
|
|
movq m1, [aq-1]
|
|
punpckldq m2, [lq]
|
|
movd m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
pavgb m0, m1
|
|
PALIGNR m1, m2, 5, m3
|
|
psrlq m2, m1, 8
|
|
psllq m3, m1, 8
|
|
LOWPASS 2, 1, 3, 4
|
|
|
|
; ABCD <- for the following predictor:
|
|
; EFGH
|
|
; IABC | m0 contains ABCDxxxx
|
|
; JEFG | m2 contains xJIEFGHx
|
|
|
|
%if cpuflag(ssse3)
|
|
punpckldq m0, m2
|
|
pshufb m2, [pb_13456_3xm1]
|
|
movd [dstq+strideq*0], m0
|
|
pshufb m0, [pb_6012_4xm1]
|
|
movd [dstq+stride3q ], m2
|
|
psrlq m2, 8
|
|
movd [dstq+strideq*2], m0
|
|
movd [dstq+strideq*1], m2
|
|
%else
|
|
psllq m1, m2, 40
|
|
psrlq m2, 24
|
|
movd [dstq+strideq*0], m0
|
|
movd [dstq+strideq*1], m2
|
|
PALIGNR m0, m1, 7, m3
|
|
psllq m1, 8
|
|
PALIGNR m2, m1, 7, m3
|
|
movd [dstq+strideq*2], m0
|
|
movd [dstq+stride3q ], m2
|
|
%endif
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
VR_MMX_FUNCS
|
|
INIT_MMX ssse3
|
|
VR_MMX_FUNCS
|
|
|
|
%macro VR_XMM_FUNCS 1 ; n_xmm_regs for 16x16
|
|
cglobal vp9_ipred_vr_8x8, 4, 4, 5, dst, stride, l, a
|
|
movu m1, [aq-1]
|
|
movhps m2, [lq]
|
|
movq m0, [aq]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
pavgb m0, m1
|
|
PALIGNR m1, m2, 9, m3
|
|
pslldq m2, m1, 1
|
|
pslldq m3, m1, 2
|
|
LOWPASS 1, 2, 3, 4
|
|
|
|
; ABCDEFGH <- for the following predictor:
|
|
; IJKLMNOP
|
|
; QABCDEFG | m0 contains ABCDEFGHxxxxxxxx
|
|
; RIJKLMNO | m1 contains xxVUTSRQIJKLMNOP
|
|
; SQABCDEF
|
|
; TRIJKLMN
|
|
; USQABCDE
|
|
; VTRIJKLM
|
|
|
|
%if cpuflag(ssse3)
|
|
punpcklqdq m0, m1 ; ABCDEFGHxxVUTSRQ
|
|
%endif
|
|
movq [dstq+strideq*0], m0
|
|
movhps [dstq+strideq*1], m1
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, [pb_6xm1_BDF_0to6] ; xxxxxxUSQABCDEFG
|
|
pshufb m1, [pb_6xm1_246_8toE] ; xxxxxxVTRIJKLMNO
|
|
%else
|
|
psrlw m2, m1, 8 ; x_U_S_Q_xxxxxxxx
|
|
pand m3, m1, [pw_255] ; x_V_T_R_xxxxxxxx
|
|
packuswb m3, m2 ; xVTRxxxxxUSQxxxx
|
|
pslldq m3, 4 ; xxxxxVTRxxxxxUSQ
|
|
PALIGNR m0, m3, 7, m4 ; xxxxxxUSQABCDEFG
|
|
psrldq m1, 8
|
|
pslldq m3, 8
|
|
PALIGNR m1, m3, 7, m4 ; xxxxxxVTRIJKLMNO
|
|
%endif
|
|
movhps [dstq+strideq*2], m0
|
|
movhps [dstq+stride3q ], m1
|
|
lea dstq, [dstq+strideq*4]
|
|
pslldq m0, 1
|
|
pslldq m1, 1
|
|
movhps [dstq+strideq*0], m0
|
|
movhps [dstq+strideq*1], m1
|
|
pslldq m0, 1
|
|
pslldq m1, 1
|
|
movhps [dstq+strideq*2], m0
|
|
movhps [dstq+stride3q ], m1
|
|
RET
|
|
|
|
cglobal vp9_ipred_vr_16x16, 4, 4, %1, dst, stride, l, a
|
|
mova m0, [aq]
|
|
movu m1, [aq-1]
|
|
mova m2, [lq]
|
|
DEFINE_ARGS dst, stride, stride3, cnt
|
|
lea stride3q, [strideq*3]
|
|
PALIGNR m3, m1, m2, 15, m6
|
|
LOWPASS 3, 1, 0, 4
|
|
pavgb m0, m1
|
|
PALIGNR m1, m2, 1, m6
|
|
pslldq m4, m2, 1
|
|
LOWPASS 1, 2, 4, 5
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, [pb_02468ACE_13579BDF]
|
|
%else
|
|
psrlw m5, m1, 8
|
|
pand m1, [pw_255]
|
|
packuswb m1, m5
|
|
%endif
|
|
mov cntd, 4
|
|
|
|
.loop:
|
|
movlhps m2, m1
|
|
mova [dstq+strideq*0], m0
|
|
mova [dstq+strideq*1], m3
|
|
PALIGNR m4, m0, m1, 15, m6
|
|
PALIGNR m5, m3, m2, 15, m6
|
|
mova [dstq+strideq*2], m4
|
|
mova [dstq+stride3q ], m5
|
|
lea dstq, [dstq+strideq*4]
|
|
PALIGNR m0, m1, 14, m6
|
|
PALIGNR m3, m2, 14, m6
|
|
pslldq m1, 2
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_vr_32x32, 4, 4, 9, dst, stride, l, a
|
|
mova m0, [aq]
|
|
mova m2, [aq+16]
|
|
movu m1, [aq-1]
|
|
PALIGNR m3, m2, m0, 15, m6
|
|
PALIGNR m4, m2, m0, 14, m6
|
|
LOWPASS 4, 3, 2, 5
|
|
pavgb m3, m2
|
|
mova m2, [lq+16]
|
|
PALIGNR m5, m1, m2, 15, m6
|
|
LOWPASS 5, 1, 0, 6
|
|
pavgb m0, m1
|
|
mova m6, [lq]
|
|
%if ARCH_X86_64
|
|
SWAP 0, 8
|
|
%else
|
|
mova [dstq], m0
|
|
%endif
|
|
PALIGNR m1, m2, 1, m0
|
|
PALIGNR m7, m2, m6, 15, m0
|
|
LOWPASS 1, 2, 7, 0
|
|
PALIGNR m2, m6, 1, m0
|
|
pslldq m7, m6, 1
|
|
LOWPASS 2, 6, 7, 0
|
|
%if cpuflag(ssse3)
|
|
pshufb m1, [pb_02468ACE_13579BDF]
|
|
pshufb m2, [pb_02468ACE_13579BDF]
|
|
%else
|
|
psrlw m0, m1, 8
|
|
psrlw m6, m2, 8
|
|
pand m1, [pw_255]
|
|
pand m2, [pw_255]
|
|
packuswb m1, m0
|
|
packuswb m2, m6
|
|
%endif
|
|
DEFINE_ARGS dst, stride, dst16, cnt
|
|
lea dst16q, [dstq +strideq*8]
|
|
lea dst16q, [dst16q+strideq*8]
|
|
SBUTTERFLY qdq, 2, 1, 6
|
|
%if ARCH_X86_64
|
|
SWAP 0, 8
|
|
%else
|
|
mova m0, [dstq]
|
|
%endif
|
|
mov cntd, 8
|
|
|
|
.loop:
|
|
; even lines (0, 2, 4, ...): m1 | m0, m3
|
|
; odd lines (1, 3, 5, ...): m2 | m5, m4
|
|
%macro %%write 4
|
|
mova [dstq+stride%1+ 0], %3
|
|
mova [dstq+stride%1+16], %4
|
|
movhps [dst16q+stride%1 ], %2
|
|
movu [dst16q+stride%1+ 8], %3
|
|
movq [dst16q+stride%1+24], %4
|
|
PALIGNR %4, %3, 15, m6
|
|
PALIGNR %3, %2, 15, m6
|
|
pslldq %2, 1
|
|
%endmacro
|
|
|
|
%%write q*0, m1, m0, m3
|
|
%%write q*1, m2, m5, m4
|
|
lea dstq, [dstq +strideq*2]
|
|
lea dst16q, [dst16q+strideq*2]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
VR_XMM_FUNCS 7
|
|
INIT_XMM ssse3
|
|
VR_XMM_FUNCS 6
|
|
INIT_XMM avx
|
|
VR_XMM_FUNCS 6
|
|
|
|
; hd
|
|
|
|
INIT_MMX mmxext
|
|
cglobal vp9_ipred_hd_4x4, 4, 4, 0, dst, stride, l, a
|
|
movd m0, [lq]
|
|
punpckldq m0, [aq-1]
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
psrlq m1, m0, 8
|
|
psrlq m2, m1, 8
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
|
|
; DHIJ <- for the following predictor:
|
|
; CGDH
|
|
; BFCG | m1 contains ABCDxxxx
|
|
; AEBF | m2 contains EFGHIJxx
|
|
|
|
punpcklbw m1, m2
|
|
punpckhdq m0, m1, m2
|
|
|
|
; m1 contains AEBFCGDH
|
|
; m0 contains CGDHIJxx
|
|
|
|
movd [dstq+stride3q ], m1
|
|
movd [dstq+strideq*1], m0
|
|
psrlq m1, 16
|
|
psrlq m0, 16
|
|
movd [dstq+strideq*2], m1
|
|
movd [dstq+strideq*0], m0
|
|
RET
|
|
|
|
%macro HD_XMM_FUNCS 0
|
|
cglobal vp9_ipred_hd_8x8, 4, 4, 5, dst, stride, l, a
|
|
movq m0, [lq]
|
|
movhps m0, [aq-1]
|
|
DEFINE_ARGS dst, stride, stride3, dst4
|
|
lea stride3q, [strideq*3]
|
|
lea dst4q, [dstq+strideq*4]
|
|
psrldq m1, m0, 1
|
|
psrldq m2, m1, 1
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
|
|
; HPQRSTUV <- for the following predictor
|
|
; GOHPQRST
|
|
; FNGOHPQR | m1 contains ABCDEFGHxxxxxxxx
|
|
; EMFNGOHP | m2 contains IJKLMNOPQRSTUVxx
|
|
; DLEMFNGO
|
|
; CKDLEMFN
|
|
; BJCKDLEM
|
|
; AIBJCKDL
|
|
|
|
punpcklbw m1, m2
|
|
movhlps m2, m2
|
|
|
|
; m1 contains AIBJCKDLEMFNGOHP
|
|
; m2 contains QRSTUVxxxxxxxxxx
|
|
|
|
movhps [dstq +stride3q ], m1
|
|
movq [dst4q+stride3q ], m1
|
|
PALIGNR m3, m2, m1, 2, m4
|
|
movhps [dstq +strideq*2], m3
|
|
movq [dst4q+strideq*2], m3
|
|
PALIGNR m3, m2, m1, 4, m4
|
|
movhps [dstq +strideq*1], m3
|
|
movq [dst4q+strideq*1], m3
|
|
PALIGNR m2, m1, 6, m4
|
|
movhps [dstq +strideq*0], m2
|
|
movq [dst4q+strideq*0], m2
|
|
RET
|
|
|
|
cglobal vp9_ipred_hd_16x16, 4, 6, 7, dst, stride, l, a
|
|
mova m0, [lq]
|
|
movu m3, [aq-1]
|
|
DEFINE_ARGS dst, stride, stride4, dst4, dst8, dst12
|
|
lea stride4q, [strideq*4]
|
|
lea dst4q, [dstq +stride4q]
|
|
lea dst8q, [dst4q+stride4q]
|
|
lea dst12q, [dst8q+stride4q]
|
|
psrldq m4, m3, 1
|
|
psrldq m5, m3, 2
|
|
LOWPASS 5, 4, 3, 6
|
|
PALIGNR m1, m3, m0, 1, m6
|
|
PALIGNR m2, m3, m0, 2, m6
|
|
LOWPASS 2, 1, 0, 6
|
|
pavgb m1, m0
|
|
SBUTTERFLY bw, 1, 2, 6
|
|
|
|
; I PROBABLY INVERTED L0 ad L16 here
|
|
; m1, m2, m5
|
|
.loop:
|
|
sub stride4q, strideq
|
|
movhps [dstq +stride4q +0], m2
|
|
movq [dstq +stride4q +8], m5
|
|
mova [dst4q+stride4q ], m2
|
|
movhps [dst8q+stride4q +0], m1
|
|
movq [dst8q+stride4q +8], m2
|
|
mova [dst12q+stride4q ], m1
|
|
%if cpuflag(avx)
|
|
palignr m1, m2, m1, 2
|
|
palignr m2, m5, m2, 2
|
|
%elif cpuflag(ssse3)
|
|
palignr m3, m2, m1, 2
|
|
palignr m0, m5, m2, 2
|
|
mova m1, m3
|
|
mova m2, m0
|
|
%else
|
|
; slightly modified version of PALIGNR
|
|
mova m6, m2
|
|
mova m4, m5
|
|
pslldq m6, 14
|
|
pslldq m4, 14
|
|
psrldq m1, 2
|
|
psrldq m2, 2
|
|
por m1, m6
|
|
por m2, m4
|
|
%endif
|
|
psrldq m5, 2
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_hd_32x32, 4, 6, 8, dst, stride, l, a
|
|
mova m0, [lq]
|
|
mova m1, [lq+16]
|
|
movu m2, [aq-1]
|
|
movu m3, [aq+15]
|
|
DEFINE_ARGS dst, stride, stride8, dst8, dst16, dst24
|
|
lea stride8q, [strideq*8]
|
|
lea dst8q, [dstq +stride8q]
|
|
lea dst16q, [dst8q +stride8q]
|
|
lea dst24q, [dst16q+stride8q]
|
|
psrldq m4, m3, 1
|
|
psrldq m5, m3, 2
|
|
LOWPASS 5, 4, 3, 6
|
|
PALIGNR m4, m3, m2, 2, m6
|
|
PALIGNR m3, m2, 1, m6
|
|
LOWPASS 4, 3, 2, 6
|
|
PALIGNR m3, m2, m1, 2, m6
|
|
PALIGNR m2, m1, 1, m6
|
|
LOWPASS 3, 2, 1, 6
|
|
pavgb m2, m1
|
|
PALIGNR m6, m1, m0, 1, m7
|
|
PALIGNR m1, m0, 2, m7
|
|
LOWPASS 1, 6, 0, 7
|
|
pavgb m0, m6
|
|
SBUTTERFLY bw, 2, 3, 6
|
|
SBUTTERFLY bw, 0, 1, 6
|
|
|
|
; m0, m1, m2, m3, m4, m5
|
|
.loop:
|
|
sub stride8q, strideq
|
|
mova [dstq +stride8q+ 0], m3
|
|
mova [dstq +stride8q+16], m4
|
|
mova [dst8q +stride8q+ 0], m2
|
|
mova [dst8q +stride8q+16], m3
|
|
mova [dst16q+stride8q+ 0], m1
|
|
mova [dst16q+stride8q+16], m2
|
|
mova [dst24q+stride8q+ 0], m0
|
|
mova [dst24q+stride8q+16], m1
|
|
%if cpuflag(avx)
|
|
palignr m0, m1, m0, 2
|
|
palignr m1, m2, m1, 2
|
|
palignr m2, m3, m2, 2
|
|
palignr m3, m4, m3, 2
|
|
palignr m4, m5, m4, 2
|
|
psrldq m5, 2
|
|
%elif cpuflag(ssse3)
|
|
psrldq m6, m5, 2
|
|
palignr m5, m4, 2
|
|
palignr m4, m3, 2
|
|
palignr m3, m2, 2
|
|
palignr m2, m1, 2
|
|
palignr m1, m0, 2
|
|
mova m0, m1
|
|
mova m1, m2
|
|
mova m2, m3
|
|
mova m3, m4
|
|
mova m4, m5
|
|
mova m5, m6
|
|
%else
|
|
; sort of a half-integrated version of PALIGNR
|
|
pslldq m7, m4, 14
|
|
pslldq m6, m5, 14
|
|
psrldq m4, 2
|
|
psrldq m5, 2
|
|
por m4, m6
|
|
pslldq m6, m3, 14
|
|
psrldq m3, 2
|
|
por m3, m7
|
|
pslldq m7, m2, 14
|
|
psrldq m2, 2
|
|
por m2, m6
|
|
pslldq m6, m1, 14
|
|
psrldq m1, 2
|
|
por m1, m7
|
|
psrldq m0, 2
|
|
por m0, m6
|
|
%endif
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
HD_XMM_FUNCS
|
|
INIT_XMM ssse3
|
|
HD_XMM_FUNCS
|
|
INIT_XMM avx
|
|
HD_XMM_FUNCS
|
|
|
|
%macro HU_MMX_FUNCS 0
|
|
cglobal vp9_ipred_hu_4x4, 3, 3, 0, dst, stride, l
|
|
movd m0, [lq]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, [pb_0to2_5x3]
|
|
%else
|
|
punpcklbw m1, m0, m0 ; 00112233
|
|
pshufw m1, m1, q3333 ; 33333333
|
|
punpckldq m0, m1 ; 01233333
|
|
%endif
|
|
psrlq m1, m0, 8
|
|
psrlq m2, m1, 8
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
DEFINE_ARGS dst, stride, stride3
|
|
lea stride3q, [strideq*3]
|
|
SBUTTERFLY bw, 1, 2, 0
|
|
PALIGNR m2, m1, 2, m0
|
|
movd [dstq+strideq*0], m1
|
|
movd [dstq+strideq*1], m2
|
|
punpckhdq m1, m1
|
|
punpckhdq m2, m2
|
|
movd [dstq+strideq*2], m1
|
|
movd [dstq+stride3q ], m2
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_MMX mmxext
|
|
HU_MMX_FUNCS
|
|
INIT_MMX ssse3
|
|
HU_MMX_FUNCS
|
|
|
|
%macro HU_XMM_FUNCS 1 ; n_xmm_regs in hu_32x32
|
|
cglobal vp9_ipred_hu_8x8, 3, 4, 4, dst, stride, l
|
|
movq m0, [lq]
|
|
%if cpuflag(ssse3)
|
|
pshufb m0, [pb_0to6_9x7]
|
|
%else
|
|
punpcklbw m1, m0, m0 ; 0011223344556677
|
|
punpckhwd m1, m1 ; 4444555566667777
|
|
shufps m0, m1, q3310 ; 0123456777777777
|
|
%endif
|
|
psrldq m1, m0, 1
|
|
psrldq m2, m1, 1
|
|
LOWPASS 2, 1, 0, 3
|
|
pavgb m1, m0
|
|
DEFINE_ARGS dst, stride, stride3, dst4
|
|
lea stride3q, [strideq*3]
|
|
lea dst4q, [dstq+strideq*4]
|
|
SBUTTERFLY bw, 1, 2, 0
|
|
movq [dstq +strideq*0], m1
|
|
movhps [dst4q+strideq*0], m1
|
|
PALIGNR m0, m2, m1, 2, m3
|
|
movq [dstq +strideq*1], m0
|
|
movhps [dst4q+strideq*1], m0
|
|
PALIGNR m0, m2, m1, 4, m3
|
|
movq [dstq +strideq*2], m0
|
|
movhps [dst4q+strideq*2], m0
|
|
PALIGNR m2, m1, 6, m3
|
|
movq [dstq +stride3q ], m2
|
|
movhps [dst4q+stride3q ], m2
|
|
RET
|
|
|
|
cglobal vp9_ipred_hu_16x16, 3, 4, 5, dst, stride, l
|
|
mova m0, [lq]
|
|
%if cpuflag(ssse3)
|
|
mova m3, [pb_2toE_3xF]
|
|
pshufb m1, m0, [pb_1toE_2xF]
|
|
pshufb m2, m0, m3
|
|
%else
|
|
pand m3, m0, [pb_15x0_1xm1]
|
|
psrldq m1, m0, 1
|
|
por m1, m3
|
|
punpckhbw m3, m3
|
|
psrldq m2, m0, 2
|
|
por m2, m3
|
|
%endif
|
|
LOWPASS 2, 1, 0, 4
|
|
pavgb m1, m0
|
|
DEFINE_ARGS dst, stride, stride9, cnt
|
|
lea stride9q, [strideq*8+strideq]
|
|
mov cntd, 4
|
|
SBUTTERFLY bw, 1, 2, 0
|
|
|
|
.loop:
|
|
mova [dstq+strideq*0], m1
|
|
mova [dstq+strideq*8], m2
|
|
PALIGNR m0, m2, m1, 2, m4
|
|
%if cpuflag(ssse3)
|
|
pshufb m2, m3
|
|
%else
|
|
psrldq m2, 2
|
|
por m2, m3
|
|
%endif
|
|
mova [dstq+strideq*1], m0
|
|
mova [dstq+stride9q ], m2
|
|
PALIGNR m1, m2, m0, 2, m4
|
|
%if cpuflag(ssse3)
|
|
pshufb m2, m3
|
|
%else
|
|
psrldq m2, 2
|
|
por m2, m3
|
|
%endif
|
|
lea dstq, [dstq+strideq*2]
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
|
|
cglobal vp9_ipred_hu_32x32, 3, 7, %1, dst, stride, l
|
|
mova m1, [lq]
|
|
mova m0, [lq+16]
|
|
PALIGNR m2, m0, m1, 1, m5
|
|
PALIGNR m3, m0, m1, 2, m5
|
|
LOWPASS 3, 2, 1, 5
|
|
pavgb m2, m1
|
|
%if cpuflag(ssse3)
|
|
mova m4, [pb_2toE_3xF]
|
|
pshufb m5, m0, [pb_1toE_2xF]
|
|
pshufb m1, m0, m4
|
|
%else
|
|
pand m4, m0, [pb_15x0_1xm1]
|
|
psrldq m5, m0, 1
|
|
por m5, m4
|
|
punpckhbw m4, m4
|
|
psrldq m1, m0, 2
|
|
por m1, m4
|
|
%endif
|
|
LOWPASS 1, 5, 0, 6
|
|
pavgb m0, m5
|
|
DEFINE_ARGS dst, stride, cnt, stride0, dst8, dst16, dst24
|
|
mov cntd, 8
|
|
xor stride0q, stride0q
|
|
lea dst8q, [dstq +strideq*8]
|
|
lea dst16q, [dst8q +strideq*8]
|
|
lea dst24q, [dst16q+strideq*8]
|
|
SBUTTERFLY bw, 0, 1, 5
|
|
SBUTTERFLY bw, 2, 3, 5
|
|
%if cpuflag(ssse3)
|
|
pshufb m6, m1, [pb_15]
|
|
%else
|
|
pshufhw m6, m4, q3333
|
|
punpckhqdq m6, m6
|
|
%endif
|
|
|
|
.loop:
|
|
mova [dstq +stride0q+ 0], m2
|
|
mova [dstq +stride0q+16], m3
|
|
mova [dst8q +stride0q+ 0], m3
|
|
mova [dst8q +stride0q+16], m0
|
|
mova [dst16q+stride0q+ 0], m0
|
|
mova [dst16q+stride0q+16], m1
|
|
mova [dst24q+stride0q+ 0], m1
|
|
mova [dst24q+stride0q+16], m6
|
|
%if cpuflag(avx)
|
|
palignr m2, m3, m2, 2
|
|
palignr m3, m0, m3, 2
|
|
palignr m0, m1, m0, 2
|
|
pshufb m1, m4
|
|
%elif cpuflag(ssse3)
|
|
pshufb m5, m1, m4
|
|
palignr m1, m0, 2
|
|
palignr m0, m3, 2
|
|
palignr m3, m2, 2
|
|
mova m2, m3
|
|
mova m3, m0
|
|
mova m0, m1
|
|
mova m1, m5
|
|
%else
|
|
; half-integrated version of PALIGNR
|
|
pslldq m5, m1, 14
|
|
pslldq m7, m0, 14
|
|
psrldq m1, 2
|
|
psrldq m0, 2
|
|
por m1, m4
|
|
por m0, m5
|
|
pslldq m5, m3, 14
|
|
psrldq m3, 2
|
|
por m3, m7
|
|
psrldq m2, 2
|
|
por m2, m5
|
|
%endif
|
|
add stride0q, strideq
|
|
dec cntd
|
|
jg .loop
|
|
RET
|
|
%endmacro
|
|
|
|
INIT_XMM sse2
|
|
HU_XMM_FUNCS 8
|
|
INIT_XMM ssse3
|
|
HU_XMM_FUNCS 7
|
|
INIT_XMM avx
|
|
HU_XMM_FUNCS 7
|
|
|
|
; FIXME 127, 128, 129 ?
|