ref: 6ef9a030145eef79fdeab6d4d38e00423ae7a83d
dir: /src/x86/cdef_sse.asm/
; Copyright © 2018, VideoLAN and dav1d authors ; Copyright © 2018, Two Orioles, LLC ; Copyright © 2019, VideoLabs ; All rights reserved. ; ; Redistribution and use in source and binary forms, with or without ; modification, are permitted provided that the following conditions are met: ; ; 1. Redistributions of source code must retain the above copyright notice, this ; list of conditions and the following disclaimer. ; ; 2. Redistributions in binary form must reproduce the above copyright notice, ; this list of conditions and the following disclaimer in the documentation ; and/or other materials provided with the distribution. ; ; THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ; ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED ; WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE ; DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ; ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES ; (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; ; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ; ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT ; (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS ; SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. %include "ext/x86/x86inc.asm" SECTION_RODATA 16 %if ARCH_X86_32 pb_0: times 16 db 0 pb_0xFF: times 16 db 0xFF %endif pw_128: times 8 dw 128 pw_256: times 8 dw 256 pw_2048: times 8 dw 2048 %if ARCH_X86_32 pw_0x7FFF: times 8 dw 0x7FFF pw_0x8000: times 8 dw 0x8000 %endif div_table_sse4: dd 840, 420, 280, 210, 168, 140, 120, 105 dd 420, 210, 140, 105, 105, 105, 105, 105 div_table_ssse3: dw 840, 840, 420, 420, 280, 280, 210, 210, 168, 168, 140, 140, 120, 120, 105, 105 dw 420, 420, 210, 210, 140, 140, 105, 105, 105, 105, 105, 105, 105, 105, 105, 105 shufb_lohi: db 0, 8, 1, 9, 2, 10, 3, 11, 4, 12, 5, 13, 6, 14, 7, 15 shufw_6543210x: db 12, 13, 10, 11, 8, 9, 6, 7, 4, 5, 2, 3, 0, 1, 14, 15 tap_table: ; masks for 8-bit shift emulation db 0xFF, 0x7F, 0x3F, 0x1F, 0x0F, 0x07, 0x03, 0x01 ; weights db 4, 2, 3, 3, 2, 1 ; taps indices db -1 * 16 + 1, -2 * 16 + 2 db 0 * 16 + 1, -1 * 16 + 2 db 0 * 16 + 1, 0 * 16 + 2 db 0 * 16 + 1, 1 * 16 + 2 db 1 * 16 + 1, 2 * 16 + 2 db 1 * 16 + 0, 2 * 16 + 1 db 1 * 16 + 0, 2 * 16 + 0 db 1 * 16 + 0, 2 * 16 - 1 ; the last 6 are repeats of the first 6 so we don't need to & 7 db -1 * 16 + 1, -2 * 16 + 2 db 0 * 16 + 1, -1 * 16 + 2 db 0 * 16 + 1, 0 * 16 + 2 db 0 * 16 + 1, 1 * 16 + 2 db 1 * 16 + 1, 2 * 16 + 2 db 1 * 16 + 0, 2 * 16 + 1 SECTION .text %macro movif32 2 %if ARCH_X86_32 mov %1, %2 %endif %endmacro %macro SAVE_ARG 2 ; varname, argnum %define %1_stkloc [rsp+%2*gprsize] %define %1_argnum %2 mov r2, r%2m mov %1_stkloc, r2 %endmacro %macro LOAD_ARG 1-2 0 ; varname, load_to_varname_register %if %2 == 0 mov r %+ %{1}_argnum, %1_stkloc %else mov %1q, %1_stkloc %endif %endmacro %macro LOAD_ARG32 1-2 ; varname, load_to_varname_register %if ARCH_X86_32 %if %0 == 1 LOAD_ARG %1 %else LOAD_ARG %1, %2 %endif %endif %endmacro %if ARCH_X86_32 %define PIC_base_offset $$ %define PIC_sym(sym) (PIC_reg+(sym)-PIC_base_offset) %else %define PIC_sym(sym) sym %endif %macro SAVE_PIC_REG 1 %if ARCH_X86_32 mov [esp+%1], PIC_reg %endif %endmacro %macro LOAD_PIC_REG 1 %if ARCH_X86_32 mov PIC_reg, [esp+%1] %endif %endmacro %macro ACCUMULATE_TAP 7 ; tap_offset, shift, shift_mask, strength, mul_tap, w, stride ; load p0/p1 movsx offq, byte [dirq+kq+%1] ; off1 %if %6 == 4 movq m5, [stkq+offq*2+%7*0] ; p0 movhps m5, [stkq+offq*2+%7*1] %else movu m5, [stkq+offq*2+%7*0] ; p0 %endif neg offq ; -off1 %if %6 == 4 movq m6, [stkq+offq*2+%7*0] ; p1 movhps m6, [stkq+offq*2+%7*1] %else movu m6, [stkq+offq*2+%7*0] ; p1 %endif %if cpuflag(sse4) ; out of bounds values are set to a value that is a both a large unsigned ; value and a negative signed value. ; use signed max and unsigned min to remove them pmaxsw m7, m5 pminuw m8, m5 pmaxsw m7, m6 pminuw m8, m6 %else %if ARCH_X86_64 pcmpeqw m9, m14, m5 pcmpeqw m10, m14, m6 pandn m9, m5 pandn m10, m6 pmaxsw m7, m9 ; max after p0 pminsw m8, m5 ; min after p0 pmaxsw m7, m10 ; max after p1 pminsw m8, m6 ; min after p1 %else pcmpeqw m3, m5, OUT_OF_BOUNDS_MEM pandn m3, m5 pmaxsw m7, m3 ; max after p0 pminsw m8, m5 ; min after p0 pcmpeqw m3, m6, OUT_OF_BOUNDS_MEM pandn m3, m6 pmaxsw m7, m3 ; max after p1 pminsw m8, m6 ; min after p1 %endif %endif ; accumulate sum[m13] over p0/p1 psubw m5, m4 ; diff_p0(p0 - px) psubw m6, m4 ; diff_p1(p1 - px) packsswb m5, m6 ; convert pixel diff to 8-bit %if ARCH_X86_64 && cpuflag(sse4) pshufb m5, m14 ; group diffs p0 and p1 into pairs %else pshufb m5, [PIC_sym(shufb_lohi)] %endif pabsb m6, m5 psignb m9, %5, m5 %if ARCH_X86_64 psrlw m10, m6, %2 ; emulate 8-bit shift pand m10, %3 psubusb m5, %4, m10 %else psrlw m5, m6, %2 ; emulate 8-bit shift pand m5, %3 paddusb m5, %4 pxor m5, [PIC_sym(pb_0xFF)] %endif pminub m5, m6 ; constrain(diff_p) pmaddubsw m5, m9 ; constrain(diff_p) * taps paddw m13, m5 %endmacro %macro PMOVZXBW 2-3 0 ; %3 = half %if %3 == 1 movd %1, %2 %else movq %1, %2 %endif punpcklbw %1, m15 %endmacro %macro LOAD_BODY 4 ; dst, src, block_width, tmp_stride %if %3 == 4 PMOVZXBW m0, [%2+strideq*0] PMOVZXBW m1, [%2+strideq*1] PMOVZXBW m2, [%2+strideq*2] PMOVZXBW m3, [%2+stride3q] %else movu m0, [%2+strideq*0] movu m1, [%2+strideq*1] movu m2, [%2+strideq*2] movu m3, [%2+stride3q] punpckhbw m4, m0, m15 punpcklbw m0, m15 punpckhbw m5, m1, m15 punpcklbw m1, m15 punpckhbw m6, m2, m15 punpcklbw m2, m15 punpckhbw m7, m3, m15 punpcklbw m3, m15 %endif mova [%1+0*%4], m0 mova [%1+1*%4], m1 mova [%1+2*%4], m2 mova [%1+3*%4], m3 %if %3 == 8 mova [%1+0*%4+2*8], m4 mova [%1+1*%4+2*8], m5 mova [%1+2*%4+2*8], m6 mova [%1+3*%4+2*8], m7 %endif %endmacro %macro CDEF_FILTER 3 ; w, h, stride %if cpuflag(sse4) %define OUT_OF_BOUNDS 0x80008000 %else %define OUT_OF_BOUNDS 0x7FFF7FFF %endif %if ARCH_X86_64 cglobal cdef_filter_%1x%2, 4, 9, 16, 3 * 16 + (%2+4)*%3, \ dst, stride, left, top, pri, sec, stride3, dst4, edge pcmpeqw m14, m14 %if cpuflag(sse4) psllw m14, 15 ; 0x8000 %else psrlw m14, 1 ; 0x7FFF %endif pxor m15, m15 %define px rsp+3*16+2*%3 %else cglobal cdef_filter_%1x%2, 2, 7, 8, - 7 * 16 - (%2+4)*%3, \ dst, stride, left, top, stride3, dst4, edge SAVE_ARG left, 2 SAVE_ARG top, 3 SAVE_ARG pri, 4 SAVE_ARG sec, 5 SAVE_ARG dir, 6 SAVE_ARG damping, 7 %define PIC_reg r2 LEA PIC_reg, PIC_base_offset %if cpuflag(sse4) %define OUT_OF_BOUNDS_MEM [PIC_sym(pw_0x8000)] %else %define OUT_OF_BOUNDS_MEM [PIC_sym(pw_0x7FFF)] %endif %define m15 [PIC_sym(pb_0)] %define px esp+7*16+2*%3 %endif mov edged, r8m ; prepare pixel buffers - body/right %if %2 == 8 lea dst4q, [dstq+strideq*4] %endif lea stride3q, [strideq*3] test edged, 2 ; have_right jz .no_right LOAD_BODY px, dstq, %1, %3 %if %2 == 8 LOAD_BODY px+4*%3, dst4q, %1, %3 %endif jmp .body_done .no_right: PMOVZXBW m0, [dstq+strideq*0], %1 == 4 PMOVZXBW m1, [dstq+strideq*1], %1 == 4 PMOVZXBW m2, [dstq+strideq*2], %1 == 4 PMOVZXBW m3, [dstq+stride3q ], %1 == 4 %if %2 == 8 PMOVZXBW m4, [dst4q+strideq*0], %1 == 4 PMOVZXBW m5, [dst4q+strideq*1], %1 == 4 PMOVZXBW m6, [dst4q+strideq*2], %1 == 4 PMOVZXBW m7, [dst4q+stride3q ], %1 == 4 %endif mova [px+0*%3], m0 mova [px+1*%3], m1 mova [px+2*%3], m2 mova [px+3*%3], m3 %if %2 == 8 mova [px+4*%3], m4 mova [px+5*%3], m5 mova [px+6*%3], m6 mova [px+7*%3], m7 mov dword [px+4*%3+%1*2], OUT_OF_BOUNDS mov dword [px+5*%3+%1*2], OUT_OF_BOUNDS mov dword [px+6*%3+%1*2], OUT_OF_BOUNDS mov dword [px+7*%3+%1*2], OUT_OF_BOUNDS %endif mov dword [px+0*%3+%1*2], OUT_OF_BOUNDS mov dword [px+1*%3+%1*2], OUT_OF_BOUNDS mov dword [px+2*%3+%1*2], OUT_OF_BOUNDS mov dword [px+3*%3+%1*2], OUT_OF_BOUNDS .body_done: ; top %if ARCH_X86_64 DEFINE_ARGS dst, stride, left, top2, pri, sec, stride3, top1, edge %else DEFINE_ARGS dst, stride, left, top2, stride3, top1, edge %endif LOAD_ARG32 top test edged, 4 ; have_top jz .no_top mov top1q, [top2q+0*gprsize] mov top2q, [top2q+1*gprsize] test edged, 1 ; have_left jz .top_no_left test edged, 2 ; have_right jz .top_no_right %if %1 == 4 PMOVZXBW m0, [top1q-2] PMOVZXBW m1, [top2q-2] %else movu m0, [top1q-4] movu m1, [top2q-4] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 movu [px-2*%3+8], m2 movu [px-1*%3+8], m3 %endif movu [px-2*%3-%1], m0 movu [px-1*%3-%1], m1 jmp .top_done .top_no_right: %if %1 == 4 PMOVZXBW m0, [top1q-%1] PMOVZXBW m1, [top2q-%1] movu [px-2*%3-4*2], m0 movu [px-1*%3-4*2], m1 %else movu m0, [top1q-%1] movu m1, [top2q-%2] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 mova [px-2*%3-8*2], m0 mova [px-2*%3-0*2], m2 mova [px-1*%3-8*2], m1 mova [px-1*%3-0*2], m3 %endif mov dword [px-2*%3+%1*2], OUT_OF_BOUNDS mov dword [px-1*%3+%1*2], OUT_OF_BOUNDS jmp .top_done .top_no_left: test edged, 2 ; have_right jz .top_no_left_right %if %1 == 4 PMOVZXBW m0, [top1q] PMOVZXBW m1, [top2q] %else movu m0, [top1q] movu m1, [top2q] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 movd [px-2*%3+8*2], m2 movd [px-1*%3+8*2], m3 %endif mova [px-2*%3], m0 mova [px-1*%3], m1 mov dword [px-2*%3-4], OUT_OF_BOUNDS mov dword [px-1*%3-4], OUT_OF_BOUNDS jmp .top_done .top_no_left_right: PMOVZXBW m0, [top1q], %1 == 4 PMOVZXBW m1, [top2q], %1 == 4 mova [px-2*%3], m0 mova [px-1*%3], m1 mov dword [px-2*%3+%1*2], OUT_OF_BOUNDS mov dword [px-1*%3+%1*2], OUT_OF_BOUNDS mov dword [px-2*%3-4], OUT_OF_BOUNDS mov dword [px-1*%3-4], OUT_OF_BOUNDS jmp .top_done .no_top: %if ARCH_X86_64 SWAP m0, m14 %else mova m0, OUT_OF_BOUNDS_MEM %endif movu [px-2*%3-4], m0 movu [px-1*%3-4], m0 %if %1 == 8 movq [px-2*%3+12], m0 movq [px-1*%3+12], m0 %endif %if ARCH_X86_64 SWAP m0, m14 %endif .top_done: ; left test edged, 1 ; have_left jz .no_left SAVE_PIC_REG 0 LOAD_ARG32 left %if %2 == 4 movq m0, [leftq] %else movu m0, [leftq] %endif LOAD_PIC_REG 0 %if %2 == 4 punpcklbw m0, m15 %else punpckhbw m1, m0, m15 punpcklbw m0, m15 movhlps m3, m1 movd [px+4*%3-4], m1 movd [px+6*%3-4], m3 psrlq m1, 32 psrlq m3, 32 movd [px+5*%3-4], m1 movd [px+7*%3-4], m3 %endif movhlps m2, m0 movd [px+0*%3-4], m0 movd [px+2*%3-4], m2 psrlq m0, 32 psrlq m2, 32 movd [px+1*%3-4], m0 movd [px+3*%3-4], m2 jmp .left_done .no_left: mov dword [px+0*%3-4], OUT_OF_BOUNDS mov dword [px+1*%3-4], OUT_OF_BOUNDS mov dword [px+2*%3-4], OUT_OF_BOUNDS mov dword [px+3*%3-4], OUT_OF_BOUNDS %if %2 == 8 mov dword [px+4*%3-4], OUT_OF_BOUNDS mov dword [px+5*%3-4], OUT_OF_BOUNDS mov dword [px+6*%3-4], OUT_OF_BOUNDS mov dword [px+7*%3-4], OUT_OF_BOUNDS %endif .left_done: ; bottom %if ARCH_X86_64 DEFINE_ARGS dst, stride, dummy1, dst8, pri, sec, stride3, dummy2, edge %else DEFINE_ARGS dst, stride, dummy1, dst8, stride3, dummy2, edge %endif test edged, 8 ; have_bottom jz .no_bottom lea dst8q, [dstq+%2*strideq] test edged, 1 ; have_left jz .bottom_no_left test edged, 2 ; have_right jz .bottom_no_right %if %1 == 4 PMOVZXBW m0, [dst8q-(%1/2)] PMOVZXBW m1, [dst8q+strideq-(%1/2)] %else movu m0, [dst8q-4] movu m1, [dst8q+strideq-4] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 movu [px+(%2+0)*%3+8], m2 movu [px+(%2+1)*%3+8], m3 %endif movu [px+(%2+0)*%3-%1], m0 movu [px+(%2+1)*%3-%1], m1 jmp .bottom_done .bottom_no_right: %if %1 == 4 PMOVZXBW m0, [dst8q-4] PMOVZXBW m1, [dst8q+strideq-4] movu [px+(%2+0)*%3-4*2], m0 movu [px+(%2+1)*%3-4*2], m1 %else movu m0, [dst8q-8] movu m1, [dst8q+strideq-8] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 mova [px+(%2+0)*%3-8*2], m0 mova [px+(%2+0)*%3-0*2], m2 mova [px+(%2+1)*%3-8*2], m1 mova [px+(%2+1)*%3-0*2], m3 mov dword [px+(%2-1)*%3+8*2], OUT_OF_BOUNDS ; overwritten by first mova %endif mov dword [px+(%2+0)*%3+%1*2], OUT_OF_BOUNDS mov dword [px+(%2+1)*%3+%1*2], OUT_OF_BOUNDS jmp .bottom_done .bottom_no_left: test edged, 2 ; have_right jz .bottom_no_left_right %if %1 == 4 PMOVZXBW m0, [dst8q] PMOVZXBW m1, [dst8q+strideq] %else movu m0, [dst8q] movu m1, [dst8q+strideq] punpckhbw m2, m0, m15 punpcklbw m0, m15 punpckhbw m3, m1, m15 punpcklbw m1, m15 mova [px+(%2+0)*%3+8*2], m2 mova [px+(%2+1)*%3+8*2], m3 %endif mova [px+(%2+0)*%3], m0 mova [px+(%2+1)*%3], m1 mov dword [px+(%2+0)*%3-4], OUT_OF_BOUNDS mov dword [px+(%2+1)*%3-4], OUT_OF_BOUNDS jmp .bottom_done .bottom_no_left_right: PMOVZXBW m0, [dst8q+strideq*0], %1 == 4 PMOVZXBW m1, [dst8q+strideq*1], %1 == 4 mova [px+(%2+0)*%3], m0 mova [px+(%2+1)*%3], m1 mov dword [px+(%2+0)*%3+%1*2], OUT_OF_BOUNDS mov dword [px+(%2+1)*%3+%1*2], OUT_OF_BOUNDS mov dword [px+(%2+0)*%3-4], OUT_OF_BOUNDS mov dword [px+(%2+1)*%3-4], OUT_OF_BOUNDS jmp .bottom_done .no_bottom: %if ARCH_X86_64 SWAP m0, m14 %else mova m0, OUT_OF_BOUNDS_MEM %endif movu [px+(%2+0)*%3-4], m0 movu [px+(%2+1)*%3-4], m0 %if %1 == 8 movq [px+(%2+0)*%3+12], m0 movq [px+(%2+1)*%3+12], m0 %endif %if ARCH_X86_64 SWAP m0, m14 %endif .bottom_done: ; actual filter DEFINE_ARGS dst, stride, pridmp, damping, pri, sec, secdmp %if ARCH_X86_64 movifnidn prid, prim movifnidn secd, secm mov dampingd, r7m %else LOAD_ARG pri LOAD_ARG sec LOAD_ARG damping, 1 %endif SAVE_PIC_REG 8 mov pridmpd, prid mov secdmpd, secd or pridmpd, 1 or secdmpd, 1 bsr pridmpd, pridmpd bsr secdmpd, secdmpd sub pridmpd, dampingd sub secdmpd, dampingd xor dampingd, dampingd neg pridmpd cmovl pridmpd, dampingd neg secdmpd cmovl secdmpd, dampingd %if ARCH_X86_64 mov [rsp+ 0], pridmpq ; pri_shift mov [rsp+16], secdmpq ; sec_shift %else mov [esp+0x00], pridmpd mov [esp+0x30], secdmpd mov dword [esp+0x04], 0 ; zero upper 32 bits of psrlw mov dword [esp+0x34], 0 ; source operand in ACCUMULATE_TAP %define PIC_reg r4 LOAD_PIC_REG 8 %endif DEFINE_ARGS dst, stride, pridmp, table, pri, sec, secdmp lea tableq, [PIC_sym(tap_table)] %if ARCH_X86_64 SWAP m2, m11 SWAP m3, m12 %endif movd m2, [tableq+pridmpq] movd m3, [tableq+secdmpq] pshufb m2, m15 ; pri_shift_mask pshufb m3, m15 ; sec_shift_mask %if ARCH_X86_64 SWAP m2, m11 SWAP m3, m12 %else %define PIC_reg r6 mov PIC_reg, r4 DEFINE_ARGS dst, stride, dir, table, pri, sec, secdmp LOAD_ARG pri LOAD_ARG dir, 1 mova [esp+0x10], m2 mova [esp+0x40], m3 %endif ; pri/sec_taps[k] [4 total] DEFINE_ARGS dst, stride, dummy, tap, pri, sec movd m0, prid movd m1, secd %if ARCH_X86_64 pshufb m0, m15 pshufb m1, m15 %else mova m2, m15 mova m3, [PIC_sym(pb_0xFF)] pshufb m0, m2 pshufb m1, m2 pxor m0, m3 pxor m1, m3 mova [esp+0x20], m0 mova [esp+0x50], m1 %endif and prid, 1 lea priq, [tapq+8+priq*2] ; pri_taps lea secq, [tapq+12] ; sec_taps %if ARCH_X86_64 && cpuflag(sse4) mova m14, [shufb_lohi] %endif ; off1/2/3[k] [6 total] from [tapq+12+(dir+0/2/6)*2+k] DEFINE_ARGS dst, stride, dir, tap, pri, sec %if ARCH_X86_64 mov dird, r6m lea dirq, [tapq+14+dirq*2] DEFINE_ARGS dst, stride, dir, stk, pri, sec, h, off, k %else lea dird, [tapd+14+dird*2] DEFINE_ARGS dst, stride, dir, stk, pri, sec %define hd dword [esp+8] %define offq dstq %define kq strideq %endif mov hd, %1*%2*2/mmsize lea stkq, [px] movif32 [esp+0x3C], strided .v_loop: movif32 [esp+0x38], dstd mov kq, 1 %if %1 == 4 movq m4, [stkq+%3*0] movhps m4, [stkq+%3*1] %else mova m4, [stkq+%3*0] ; px %endif %if ARCH_X86_32 %xdefine m9 m3 %xdefine m13 m7 %xdefine m7 m0 %xdefine m8 m1 %endif pxor m13, m13 ; sum mova m7, m4 ; max mova m8, m4 ; min .k_loop: %if ARCH_X86_64 movd m2, [priq+kq] ; pri_taps movd m3, [secq+kq] ; sec_taps pshufb m2, m15 pshufb m3, m15 ACCUMULATE_TAP 0*2, [rsp+ 0], m11, m0, m2, %1, %3 ACCUMULATE_TAP 2*2, [rsp+16], m12, m1, m3, %1, %3 ACCUMULATE_TAP 6*2, [rsp+16], m12, m1, m3, %1, %3 %else movd m2, [priq+kq] ; pri_taps pshufb m2, m15 ACCUMULATE_TAP 0*2, [esp+0x00], [esp+0x10], [esp+0x20], m2, %1, %3 movd m2, [secq+kq] ; sec_taps pshufb m2, m15 ACCUMULATE_TAP 2*2, [esp+0x30], [esp+0x40], [esp+0x50], m2, %1, %3 ACCUMULATE_TAP 6*2, [esp+0x30], [esp+0x40], [esp+0x50], m2, %1, %3 %endif dec kq jge .k_loop %if cpuflag(sse4) pcmpgtw m6, m15, m13 %else pxor m6, m6 pcmpgtw m6, m13 %endif paddw m13, m6 pmulhrsw m13, [PIC_sym(pw_2048)] paddw m4, m13 pminsw m4, m7 pmaxsw m4, m8 packuswb m4, m4 movif32 dstd, [esp+0x38] movif32 strided, [esp+0x3C] %if %1 == 4 movd [dstq+strideq*0], m4 psrlq m4, 32 movd [dstq+strideq*1], m4 %else movq [dstq], m4 %endif %if %1 == 4 %define vloop_lines (mmsize/(%1*2)) lea dstq, [dstq+strideq*vloop_lines] add stkq, %3*vloop_lines %else lea dstq, [dstq+strideq] add stkq, %3 %endif dec hd jg .v_loop RET %endmacro %macro MULLD 2 %if cpuflag(sse4) pmulld %1, %2 %else %if ARCH_X86_32 %define m15 m1 %endif pmulhuw m15, %1, %2 pmullw %1, %2 pslld m15, 16 paddd %1, m15 %endif %endmacro %macro CDEF_DIR 0 %if ARCH_X86_64 cglobal cdef_dir, 3, 5, 16, 32, src, stride, var, stride3 lea stride3q, [strideq*3] movq m1, [srcq+strideq*0] movhps m1, [srcq+strideq*1] movq m3, [srcq+strideq*2] movhps m3, [srcq+stride3q] lea srcq, [srcq+strideq*4] movq m5, [srcq+strideq*0] movhps m5, [srcq+strideq*1] movq m7, [srcq+strideq*2] movhps m7, [srcq+stride3q] pxor m8, m8 psadbw m0, m1, m8 psadbw m2, m3, m8 psadbw m4, m5, m8 psadbw m6, m7, m8 packssdw m0, m2 packssdw m4, m6 packssdw m0, m4 SWAP m0, m9 punpcklbw m0, m1, m8 punpckhbw m1, m8 punpcklbw m2, m3, m8 punpckhbw m3, m8 punpcklbw m4, m5, m8 punpckhbw m5, m8 punpcklbw m6, m7, m8 punpckhbw m7, m8 mova m8, [pw_128] psubw m0, m8 psubw m1, m8 psubw m2, m8 psubw m3, m8 psubw m4, m8 psubw m5, m8 psubw m6, m8 psubw m7, m8 psllw m8, 3 psubw m9, m8 ; partial_sum_hv[0] paddw m8, m0, m1 paddw m10, m2, m3 paddw m8, m4 paddw m10, m5 paddw m8, m6 paddw m10, m7 paddw m8, m10 ; partial_sum_hv[1] pmaddwd m8, m8 pmaddwd m9, m9 phaddd m9, m8 SWAP m8, m9 MULLD m8, [div_table%+SUFFIX+48] pslldq m9, m1, 2 psrldq m10, m1, 14 pslldq m11, m2, 4 psrldq m12, m2, 12 pslldq m13, m3, 6 psrldq m14, m3, 10 paddw m9, m0 paddw m10, m12 paddw m11, m13 paddw m10, m14 ; partial_sum_diag[0] top/right half paddw m9, m11 ; partial_sum_diag[0] top/left half pslldq m11, m4, 8 psrldq m12, m4, 8 pslldq m13, m5, 10 psrldq m14, m5, 6 paddw m9, m11 paddw m10, m12 paddw m9, m13 paddw m10, m14 pslldq m11, m6, 12 psrldq m12, m6, 4 pslldq m13, m7, 14 psrldq m14, m7, 2 paddw m9, m11 paddw m10, m12 paddw m9, m13 ; partial_sum_diag[0][0-7] paddw m10, m14 ; partial_sum_diag[0][8-14,zero] pshufb m10, [shufw_6543210x] punpckhwd m11, m9, m10 punpcklwd m9, m10 pmaddwd m11, m11 pmaddwd m9, m9 MULLD m11, [div_table%+SUFFIX+16] MULLD m9, [div_table%+SUFFIX+0] paddd m9, m11 ; cost[0a-d] pslldq m10, m0, 14 psrldq m11, m0, 2 pslldq m12, m1, 12 psrldq m13, m1, 4 pslldq m14, m2, 10 psrldq m15, m2, 6 paddw m10, m12 paddw m11, m13 paddw m10, m14 paddw m11, m15 pslldq m12, m3, 8 psrldq m13, m3, 8 pslldq m14, m4, 6 psrldq m15, m4, 10 paddw m10, m12 paddw m11, m13 paddw m10, m14 paddw m11, m15 pslldq m12, m5, 4 psrldq m13, m5, 12 pslldq m14, m6, 2 psrldq m15, m6, 14 paddw m10, m12 paddw m11, m13 paddw m10, m14 paddw m11, m15 ; partial_sum_diag[1][8-14,zero] paddw m10, m7 ; partial_sum_diag[1][0-7] pshufb m11, [shufw_6543210x] punpckhwd m12, m10, m11 punpcklwd m10, m11 pmaddwd m12, m12 pmaddwd m10, m10 MULLD m12, [div_table%+SUFFIX+16] MULLD m10, [div_table%+SUFFIX+0] paddd m10, m12 ; cost[4a-d] phaddd m9, m10 ; cost[0a/b,4a/b] paddw m10, m0, m1 paddw m11, m2, m3 paddw m12, m4, m5 paddw m13, m6, m7 phaddw m0, m4 phaddw m1, m5 phaddw m2, m6 phaddw m3, m7 ; m0-3 are horizontal sums (x >> 1), m10-13 are vertical sums (y >> 1) pslldq m4, m11, 2 psrldq m5, m11, 14 pslldq m6, m12, 4 psrldq m7, m12, 12 pslldq m14, m13, 6 psrldq m15, m13, 10 paddw m4, m10 paddw m5, m7 paddw m4, m6 paddw m5, m15 ; partial_sum_alt[3] right paddw m4, m14 ; partial_sum_alt[3] left pshuflw m6, m5, q3012 punpckhwd m5, m4 punpcklwd m4, m6 pmaddwd m5, m5 pmaddwd m4, m4 MULLD m5, [div_table%+SUFFIX+48] MULLD m4, [div_table%+SUFFIX+32] paddd m4, m5 ; cost[7a-d] pslldq m5, m10, 6 psrldq m6, m10, 10 pslldq m7, m11, 4 psrldq m10, m11, 12 pslldq m11, m12, 2 psrldq m12, 14 paddw m5, m7 paddw m6, m10 paddw m5, m11 paddw m6, m12 paddw m5, m13 pshuflw m7, m6, q3012 punpckhwd m6, m5 punpcklwd m5, m7 pmaddwd m6, m6 pmaddwd m5, m5 MULLD m6, [div_table%+SUFFIX+48] MULLD m5, [div_table%+SUFFIX+32] paddd m5, m6 ; cost[5a-d] pslldq m6, m1, 2 psrldq m7, m1, 14 pslldq m10, m2, 4 psrldq m11, m2, 12 pslldq m12, m3, 6 psrldq m13, m3, 10 paddw m6, m0 paddw m7, m11 paddw m6, m10 paddw m7, m13 ; partial_sum_alt[3] right paddw m6, m12 ; partial_sum_alt[3] left pshuflw m10, m7, q3012 punpckhwd m7, m6 punpcklwd m6, m10 pmaddwd m7, m7 pmaddwd m6, m6 MULLD m7, [div_table%+SUFFIX+48] MULLD m6, [div_table%+SUFFIX+32] paddd m6, m7 ; cost[1a-d] pshufd m0, m0, q1032 pshufd m1, m1, q1032 pshufd m2, m2, q1032 pshufd m3, m3, q1032 pslldq m10, m0, 6 psrldq m11, m0, 10 pslldq m12, m1, 4 psrldq m13, m1, 12 pslldq m14, m2, 2 psrldq m2, 14 paddw m10, m12 paddw m11, m13 paddw m10, m14 paddw m11, m2 paddw m10, m3 pshuflw m12, m11, q3012 punpckhwd m11, m10 punpcklwd m10, m12 pmaddwd m11, m11 pmaddwd m10, m10 MULLD m11, [div_table%+SUFFIX+48] MULLD m10, [div_table%+SUFFIX+32] paddd m10, m11 ; cost[3a-d] phaddd m9, m8 ; cost[0,4,2,6] phaddd m6, m10 phaddd m5, m4 phaddd m6, m5 ; cost[1,3,5,7] pshufd m4, m9, q3120 ; now find the best cost %if cpuflag(sse4) pmaxsd m9, m6 pshufd m0, m9, q1032 pmaxsd m0, m9 pshufd m1, m0, q2301 pmaxsd m0, m1 ; best cost %else pcmpgtd m0, m9, m6 pand m9, m0 pandn m0, m6 por m9, m0 pshufd m1, m9, q1032 pcmpgtd m0, m9, m1 pand m9, m0 pandn m0, m1 por m9, m0 pshufd m1, m9, q2301 pcmpgtd m0, m9, m1 pand m9, m0 pandn m0, m1 por m0, m9 %endif ; get direction and variance punpckhdq m1, m4, m6 punpckldq m4, m6 psubd m2, m0, m1 psubd m3, m0, m4 mova [rsp+0x00], m2 ; emulate ymm in stack mova [rsp+0x10], m3 pcmpeqd m1, m0 ; compute best cost mask pcmpeqd m4, m0 packssdw m4, m1 pmovmskb eax, m4 ; get byte-idx from mask tzcnt eax, eax mov r1d, [rsp+rax*2] ; get idx^4 complement from emulated ymm shr eax, 1 ; get direction by converting byte-idx to word-idx shr r1d, 10 mov [varq], r1d %else cglobal cdef_dir, 3, 5, 16, 96, src, stride, var, stride3 %define PIC_reg r4 LEA PIC_reg, PIC_base_offset pxor m0, m0 mova m1, [PIC_sym(pw_128)] lea stride3q, [strideq*3] movq m5, [srcq+strideq*0] movhps m5, [srcq+strideq*1] movq m7, [srcq+strideq*2] movhps m7, [srcq+stride3q] psadbw m2, m5, m0 psadbw m3, m7, m0 packssdw m2, m3 punpcklbw m4, m5, m0 punpckhbw m5, m0 punpcklbw m6, m7, m0 punpckhbw m7, m0 psubw m4, m1 psubw m5, m1 psubw m6, m1 psubw m7, m1 mova [esp+0x00], m4 mova [esp+0x10], m5 mova [esp+0x20], m6 mova [esp+0x50], m7 lea srcq, [srcq+strideq*4] movq m5, [srcq+strideq*0] movhps m5, [srcq+strideq*1] movq m7, [srcq+strideq*2] movhps m7, [srcq+stride3q] psadbw m3, m5, m0 psadbw m0, m7, m0 packssdw m3, m0 pxor m0, m0 packssdw m2, m3 punpcklbw m4, m5, m0 punpckhbw m5, m0 punpcklbw m6, m7, m0 punpckhbw m7, m0 psubw m4, m1 psubw m5, m1 psubw m6, m1 psubw m7, m1 psllw m1, 3 psubw m2, m1 ; partial_sum_hv[0] pmaddwd m2, m2 mova m3, [esp+0x50] mova m0, [esp+0x00] paddw m0, [esp+0x10] paddw m1, m3, [esp+0x20] paddw m0, m4 paddw m1, m5 paddw m0, m6 paddw m1, m7 paddw m0, m1 ; partial_sum_hv[1] pmaddwd m0, m0 phaddd m2, m0 MULLD m2, [PIC_sym(div_table%+SUFFIX)+48] mova [esp+0x30], m2 mova m1, [esp+0x10] pslldq m0, m1, 2 psrldq m1, 14 paddw m0, [esp+0x00] pslldq m2, m3, 6 psrldq m3, 10 paddw m0, m2 paddw m1, m3 mova m3, [esp+0x20] pslldq m2, m3, 4 psrldq m3, 12 paddw m0, m2 ; partial_sum_diag[0] top/left half paddw m1, m3 ; partial_sum_diag[0] top/right half pslldq m2, m4, 8 psrldq m3, m4, 8 paddw m0, m2 paddw m1, m3 pslldq m2, m5, 10 psrldq m3, m5, 6 paddw m0, m2 paddw m1, m3 pslldq m2, m6, 12 psrldq m3, m6, 4 paddw m0, m2 paddw m1, m3 pslldq m2, m7, 14 psrldq m3, m7, 2 paddw m0, m2 ; partial_sum_diag[0][0-7] paddw m1, m3 ; partial_sum_diag[0][8-14,zero] mova m3, [esp+0x50] pshufb m1, [PIC_sym(shufw_6543210x)] punpckhwd m2, m0, m1 punpcklwd m0, m1 pmaddwd m2, m2 pmaddwd m0, m0 MULLD m2, [PIC_sym(div_table%+SUFFIX)+16] MULLD m0, [PIC_sym(div_table%+SUFFIX)+0] paddd m0, m2 ; cost[0a-d] mova [esp+0x40], m0 mova m1, [esp+0x00] pslldq m0, m1, 14 psrldq m1, 2 paddw m0, m7 pslldq m2, m3, 8 psrldq m3, 8 paddw m0, m2 paddw m1, m3 mova m3, [esp+0x20] pslldq m2, m3, 10 psrldq m3, 6 paddw m0, m2 paddw m1, m3 mova m3, [esp+0x10] pslldq m2, m3, 12 psrldq m3, 4 paddw m0, m2 paddw m1, m3 pslldq m2, m4, 6 psrldq m3, m4, 10 paddw m0, m2 paddw m1, m3 pslldq m2, m5, 4 psrldq m3, m5, 12 paddw m0, m2 paddw m1, m3 pslldq m2, m6, 2 psrldq m3, m6, 14 paddw m0, m2 ; partial_sum_diag[1][0-7] paddw m1, m3 ; partial_sum_diag[1][8-14,zero] mova m3, [esp+0x50] pshufb m1, [PIC_sym(shufw_6543210x)] punpckhwd m2, m0, m1 punpcklwd m0, m1 pmaddwd m2, m2 pmaddwd m0, m0 MULLD m2, [PIC_sym(div_table%+SUFFIX)+16] MULLD m0, [PIC_sym(div_table%+SUFFIX)+0] paddd m0, m2 ; cost[4a-d] phaddd m1, [esp+0x40], m0 ; cost[0a/b,4a/b] phaddd m1, [esp+0x30] ; cost[0,4,2,6] mova [esp+0x30], m1 phaddw m0, [esp+0x00], m4 phaddw m1, [esp+0x10], m5 paddw m4, m5 mova m2, [esp+0x20] paddw m5, m2, m3 phaddw m2, m6 paddw m6, m7 phaddw m3, m7 mova m7, [esp+0x00] paddw m7, [esp+0x10] mova [esp+0x00], m0 mova [esp+0x10], m1 mova [esp+0x20], m2 pslldq m1, m4, 4 pslldq m2, m6, 6 pslldq m0, m5, 2 paddw m1, m2 paddw m0, m7 psrldq m2, m5, 14 paddw m0, m1 ; partial_sum_alt[3] left psrldq m1, m4, 12 paddw m1, m2 psrldq m2, m6, 10 paddw m1, m2 ; partial_sum_alt[3] right pshuflw m1, m1, q3012 punpckhwd m2, m0, m1 punpcklwd m0, m1 pmaddwd m2, m2 pmaddwd m0, m0 MULLD m2, [PIC_sym(div_table%+SUFFIX)+48] MULLD m0, [PIC_sym(div_table%+SUFFIX)+32] paddd m0, m2 ; cost[7a-d] mova [esp+0x40], m0 pslldq m0, m7, 6 psrldq m7, 10 pslldq m1, m5, 4 psrldq m5, 12 pslldq m2, m4, 2 psrldq m4, 14 paddw m0, m6 paddw m7, m5 paddw m0, m1 paddw m7, m4 paddw m0, m2 pshuflw m2, m7, q3012 punpckhwd m7, m0 punpcklwd m0, m2 pmaddwd m7, m7 pmaddwd m0, m0 MULLD m7, [PIC_sym(div_table%+SUFFIX)+48] MULLD m0, [PIC_sym(div_table%+SUFFIX)+32] paddd m0, m7 ; cost[5a-d] mova [esp+0x50], m0 mova m7, [esp+0x10] mova m2, [esp+0x20] pslldq m0, m7, 2 psrldq m7, 14 pslldq m4, m2, 4 psrldq m2, 12 pslldq m5, m3, 6 psrldq m6, m3, 10 paddw m0, [esp+0x00] paddw m7, m2 paddw m4, m5 paddw m7, m6 ; partial_sum_alt[3] right paddw m0, m4 ; partial_sum_alt[3] left pshuflw m2, m7, q3012 punpckhwd m7, m0 punpcklwd m0, m2 pmaddwd m7, m7 pmaddwd m0, m0 MULLD m7, [PIC_sym(div_table%+SUFFIX)+48] MULLD m0, [PIC_sym(div_table%+SUFFIX)+32] paddd m0, m7 ; cost[1a-d] SWAP m0, m4 pshufd m0, [esp+0x00], q1032 pshufd m1, [esp+0x10], q1032 pshufd m2, [esp+0x20], q1032 pshufd m3, m3, q1032 mova [esp+0x00], m4 pslldq m4, m0, 6 psrldq m0, 10 pslldq m5, m1, 4 psrldq m1, 12 pslldq m6, m2, 2 psrldq m2, 14 paddw m4, m3 paddw m0, m1 paddw m5, m6 paddw m0, m2 paddw m4, m5 pshuflw m2, m0, q3012 punpckhwd m0, m4 punpcklwd m4, m2 pmaddwd m0, m0 pmaddwd m4, m4 MULLD m0, [PIC_sym(div_table%+SUFFIX)+48] MULLD m4, [PIC_sym(div_table%+SUFFIX)+32] paddd m4, m0 ; cost[3a-d] mova m1, [esp+0x00] mova m2, [esp+0x50] mova m0, [esp+0x30] ; cost[0,4,2,6] phaddd m1, m4 phaddd m2, [esp+0x40] ; cost[1,3,5,7] phaddd m1, m2 pshufd m2, m0, q3120 ; now find the best cost %if cpuflag(sse4) pmaxsd m0, m1 pshufd m3, m0, q1032 pmaxsd m3, m0 pshufd m0, m3, q2301 pmaxsd m0, m3 %else pcmpgtd m3, m0, m1 pand m0, m3 pandn m3, m1 por m0, m3 pshufd m4, m0, q1032 pcmpgtd m3, m0, m4 pand m0, m3 pandn m3, m4 por m0, m3 pshufd m4, m0, q2301 pcmpgtd m3, m0, m4 pand m0, m3 pandn m3, m4 por m0, m3 %endif ; get direction and variance punpckhdq m3, m2, m1 punpckldq m2, m1 psubd m1, m0, m3 psubd m4, m0, m2 mova [esp+0x00], m1 ; emulate ymm in stack mova [esp+0x10], m4 pcmpeqd m3, m0 ; compute best cost mask pcmpeqd m2, m0 packssdw m2, m3 pmovmskb eax, m2 ; get byte-idx from mask tzcnt eax, eax mov r1d, [esp+eax*2] ; get idx^4 complement from emulated ymm shr eax, 1 ; get direction by converting byte-idx to word-idx shr r1d, 10 mov [vard], r1d %endif RET %endmacro INIT_XMM sse4 CDEF_FILTER 8, 8, 32 CDEF_FILTER 4, 8, 32 CDEF_FILTER 4, 4, 32 CDEF_DIR INIT_XMM ssse3 CDEF_FILTER 8, 8, 32 CDEF_FILTER 4, 8, 32 CDEF_FILTER 4, 4, 32 CDEF_DIR