ref: 3b79000122ab520ebbe7de226a5a755687d59224
dir: /vp8/common/x86/loopfilter_mmx.asm/
; ; Copyright (c) 2010 The WebM project authors. All Rights Reserved. ; ; Use of this source code is governed by a BSD-style license ; that can be found in the LICENSE file in the root of the source ; tree. An additional intellectual property rights grant can be found ; in the file PATENTS. All contributing project authors may ; be found in the AUTHORS file in the root of the source tree. ; %include "vpx_ports/x86_abi_support.asm" ;void vp8_loop_filter_horizontal_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit, ; const char *limit, ; const char *thresh, ; int count ;) global sym(vp8_loop_filter_horizontal_edge_mmx) PRIVATE sym(vp8_loop_filter_horizontal_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 6 GET_GOT rbx push rsi push rdi ; end prolog ALIGN_STACK 16, rax sub rsp, 32 ; reserve 32 bytes %define t0 [rsp + 0] ;__declspec(align(16)) char t0[8]; %define t1 [rsp + 16] ;__declspec(align(16)) char t1[8]; mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? movsxd rcx, dword ptr arg(5) ;count .next8_h: mov rdx, arg(3) ;limit movq mm7, [rdx] mov rdi, rsi ; rdi points to row +1 for indirect addressing add rdi, rax ; calculate breakout conditions movq mm2, [rdi+2*rax] ; q3 movq mm1, [rsi+2*rax] ; q2 movq mm6, mm1 ; q2 psubusb mm1, mm2 ; q2-=q3 psubusb mm2, mm6 ; q3-=q2 por mm1, mm2 ; abs(q3-q2) psubusb mm1, mm7 ; movq mm4, [rsi+rax] ; q1 movq mm3, mm4 ; q1 psubusb mm4, mm6 ; q1-=q2 psubusb mm6, mm3 ; q2-=q1 por mm4, mm6 ; abs(q2-q1) psubusb mm4, mm7 por mm1, mm4 movq mm4, [rsi] ; q0 movq mm0, mm4 ; q0 psubusb mm4, mm3 ; q0-=q1 psubusb mm3, mm0 ; q1-=q0 por mm4, mm3 ; abs(q0-q1) movq t0, mm4 ; save to t0 psubusb mm4, mm7 por mm1, mm4 neg rax ; negate pitch to deal with above border movq mm2, [rsi+4*rax] ; p3 movq mm4, [rdi+4*rax] ; p2 movq mm5, mm4 ; p2 psubusb mm4, mm2 ; p2-=p3 psubusb mm2, mm5 ; p3-=p2 por mm4, mm2 ; abs(p3 - p2) psubusb mm4, mm7 por mm1, mm4 movq mm4, [rsi+2*rax] ; p1 movq mm3, mm4 ; p1 psubusb mm4, mm5 ; p1-=p2 psubusb mm5, mm3 ; p2-=p1 por mm4, mm5 ; abs(p2 - p1) psubusb mm4, mm7 por mm1, mm4 movq mm2, mm3 ; p1 movq mm4, [rsi+rax] ; p0 movq mm5, mm4 ; p0 psubusb mm4, mm3 ; p0-=p1 psubusb mm3, mm5 ; p1-=p0 por mm4, mm3 ; abs(p1 - p0) movq t1, mm4 ; save to t1 psubusb mm4, mm7 por mm1, mm4 movq mm3, [rdi] ; q1 movq mm4, mm3 ; q1 psubusb mm3, mm2 ; q1-=p1 psubusb mm2, mm4 ; p1-=q1 por mm2, mm3 ; abs(p1-q1) pand mm2, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm2, 1 ; abs(p1-q1)/2 movq mm6, mm5 ; p0 movq mm3, [rsi] ; q0 psubusb mm5, mm3 ; p0-=q0 psubusb mm3, mm6 ; q0-=p0 por mm5, mm3 ; abs(p0 - q0) paddusb mm5, mm5 ; abs(p0-q0)*2 paddusb mm5, mm2 ; abs (p0 - q0) *2 + abs(p1-q1)/2 mov rdx, arg(2) ;blimit ; get blimit movq mm7, [rdx] ; blimit psubusb mm5, mm7 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit por mm1, mm5 pxor mm5, mm5 pcmpeqb mm1, mm5 ; mask mm1 ; calculate high edge variance mov rdx, arg(4) ;thresh ; get thresh movq mm7, [rdx] ; movq mm4, t0 ; get abs (q1 - q0) psubusb mm4, mm7 movq mm3, t1 ; get abs (p1 - p0) psubusb mm3, mm7 paddb mm4, mm3 ; abs(q1 - q0) > thresh || abs(p1 - p0) > thresh pcmpeqb mm4, mm5 pcmpeqb mm5, mm5 pxor mm4, mm5 ; start work on filters movq mm2, [rsi+2*rax] ; p1 movq mm7, [rdi] ; q1 pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm2, mm7 ; p1 - q1 pand mm2, mm4 ; high var mask (hvm)(p1 - q1) pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm0 ; q0 psubsb mm0, mm6 ; q0 - p0 paddsb mm2, mm0 ; 1 * (q0 - p0) + hvm(p1 - q1) paddsb mm2, mm0 ; 2 * (q0 - p0) + hvm(p1 - q1) paddsb mm2, mm0 ; 3 * (q0 - p0) + hvm(p1 - q1) pand mm1, mm2 ; mask filter values we don't care about movq mm2, mm1 paddsb mm1, [GLOBAL(t4)] ; 3* (q0 - p0) + hvm(p1 - q1) + 4 paddsb mm2, [GLOBAL(t3)] ; 3* (q0 - p0) + hvm(p1 - q1) + 3 pxor mm0, mm0 ; pxor mm5, mm5 punpcklbw mm0, mm2 ; punpckhbw mm5, mm2 ; psraw mm0, 11 ; psraw mm5, 11 packsswb mm0, mm5 movq mm2, mm0 ; (3* (q0 - p0) + hvm(p1 - q1) + 3) >> 3; pxor mm0, mm0 ; 0 movq mm5, mm1 ; abcdefgh punpcklbw mm0, mm1 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 pxor mm1, mm1 ; 0 punpckhbw mm1, mm5 ; a0b0c0d0 psraw mm1, 11 ; sign extended shift right by 3 movq mm5, mm0 ; save results packsswb mm0, mm1 ; (3* (q0 - p0) + hvm(p1 - q1) + 4) >>3 paddsw mm5, [GLOBAL(ones)] paddsw mm1, [GLOBAL(ones)] psraw mm5, 1 ; partial shifted one more time for 2nd tap psraw mm1, 1 ; partial shifted one more time for 2nd tap packsswb mm5, mm1 ; (3* (q0 - p0) + hvm(p1 - q1) + 4) >>4 pandn mm4, mm5 ; high edge variance additive paddsb mm6, mm2 ; p0+= p0 add pxor mm6, [GLOBAL(t80)] ; unoffset movq [rsi+rax], mm6 ; write back movq mm6, [rsi+2*rax] ; p1 pxor mm6, [GLOBAL(t80)] ; reoffset paddsb mm6, mm4 ; p1+= p1 add pxor mm6, [GLOBAL(t80)] ; unoffset movq [rsi+2*rax], mm6 ; write back psubsb mm3, mm0 ; q0-= q0 add pxor mm3, [GLOBAL(t80)] ; unoffset movq [rsi], mm3 ; write back psubsb mm7, mm4 ; q1-= q1 add pxor mm7, [GLOBAL(t80)] ; unoffset movq [rdi], mm7 ; write back add rsi,8 neg rax dec rcx jnz .next8_h add rsp, 32 pop rsp ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void vp8_loop_filter_vertical_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit, ; const char *limit, ; const char *thresh, ; int count ;) global sym(vp8_loop_filter_vertical_edge_mmx) PRIVATE sym(vp8_loop_filter_vertical_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 6 GET_GOT rbx push rsi push rdi ; end prolog ALIGN_STACK 16, rax sub rsp, 64 ; reserve 64 bytes %define t0 [rsp + 0] ;__declspec(align(16)) char t0[8]; %define t1 [rsp + 16] ;__declspec(align(16)) char t1[8]; %define srct [rsp + 32] ;__declspec(align(16)) char srct[32]; mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? lea rsi, [rsi + rax*4 - 4] movsxd rcx, dword ptr arg(5) ;count .next8_v: mov rdi, rsi ; rdi points to row +1 for indirect addressing add rdi, rax ;transpose movq mm6, [rsi+2*rax] ; 67 66 65 64 63 62 61 60 movq mm7, mm6 ; 77 76 75 74 73 72 71 70 punpckhbw mm7, [rdi+2*rax] ; 77 67 76 66 75 65 74 64 punpcklbw mm6, [rdi+2*rax] ; 73 63 72 62 71 61 70 60 movq mm4, [rsi] ; 47 46 45 44 43 42 41 40 movq mm5, mm4 ; 47 46 45 44 43 42 41 40 punpckhbw mm5, [rsi+rax] ; 57 47 56 46 55 45 54 44 punpcklbw mm4, [rsi+rax] ; 53 43 52 42 51 41 50 40 movq mm3, mm5 ; 57 47 56 46 55 45 54 44 punpckhwd mm5, mm7 ; 77 67 57 47 76 66 56 46 punpcklwd mm3, mm7 ; 75 65 55 45 74 64 54 44 movq mm2, mm4 ; 53 43 52 42 51 41 50 40 punpckhwd mm4, mm6 ; 73 63 53 43 72 62 52 42 punpcklwd mm2, mm6 ; 71 61 51 41 70 60 50 40 neg rax movq mm6, [rsi+rax*2] ; 27 26 25 24 23 22 21 20 movq mm1, mm6 ; 27 26 25 24 23 22 21 20 punpckhbw mm6, [rsi+rax] ; 37 27 36 36 35 25 34 24 punpcklbw mm1, [rsi+rax] ; 33 23 32 22 31 21 30 20 movq mm7, [rsi+rax*4]; ; 07 06 05 04 03 02 01 00 punpckhbw mm7, [rdi+rax*4] ; 17 07 16 06 15 05 14 04 movq mm0, mm7 ; 17 07 16 06 15 05 14 04 punpckhwd mm7, mm6 ; 37 27 17 07 36 26 16 06 punpcklwd mm0, mm6 ; 35 25 15 05 34 24 14 04 movq mm6, mm7 ; 37 27 17 07 36 26 16 06 punpckhdq mm7, mm5 ; 77 67 57 47 37 27 17 07 = q3 punpckldq mm6, mm5 ; 76 66 56 46 36 26 16 06 = q2 movq mm5, mm6 ; 76 66 56 46 36 26 16 06 psubusb mm5, mm7 ; q2-q3 psubusb mm7, mm6 ; q3-q2 por mm7, mm5; ; mm7=abs (q3-q2) movq mm5, mm0 ; 35 25 15 05 34 24 14 04 punpckhdq mm5, mm3 ; 75 65 55 45 35 25 15 05 = q1 punpckldq mm0, mm3 ; 74 64 54 44 34 24 15 04 = q0 movq mm3, mm5 ; 75 65 55 45 35 25 15 05 = q1 psubusb mm3, mm6 ; q1-q2 psubusb mm6, mm5 ; q2-q1 por mm6, mm3 ; mm6=abs(q2-q1) lea rdx, srct movq [rdx+24], mm5 ; save q1 movq [rdx+16], mm0 ; save q0 movq mm3, [rsi+rax*4] ; 07 06 05 04 03 02 01 00 punpcklbw mm3, [rdi+rax*4] ; 13 03 12 02 11 01 10 00 movq mm0, mm3 ; 13 03 12 02 11 01 10 00 punpcklwd mm0, mm1 ; 31 21 11 01 30 20 10 00 punpckhwd mm3, mm1 ; 33 23 13 03 32 22 12 02 movq mm1, mm0 ; 31 21 11 01 30 20 10 00 punpckldq mm0, mm2 ; 70 60 50 40 30 20 10 00 =p3 punpckhdq mm1, mm2 ; 71 61 51 41 31 21 11 01 =p2 movq mm2, mm1 ; 71 61 51 41 31 21 11 01 =p2 psubusb mm2, mm0 ; p2-p3 psubusb mm0, mm1 ; p3-p2 por mm0, mm2 ; mm0=abs(p3-p2) movq mm2, mm3 ; 33 23 13 03 32 22 12 02 punpckldq mm2, mm4 ; 72 62 52 42 32 22 12 02 = p1 punpckhdq mm3, mm4 ; 73 63 53 43 33 23 13 03 = p0 movq [rdx+8], mm3 ; save p0 movq [rdx], mm2 ; save p1 movq mm5, mm2 ; mm5 = p1 psubusb mm2, mm1 ; p1-p2 psubusb mm1, mm5 ; p2-p1 por mm1, mm2 ; mm1=abs(p2-p1) mov rdx, arg(3) ;limit movq mm4, [rdx] ; mm4 = limit psubusb mm7, mm4 psubusb mm0, mm4 psubusb mm1, mm4 psubusb mm6, mm4 por mm7, mm6 por mm0, mm1 por mm0, mm7 ; abs(q3-q2) > limit || abs(p3-p2) > limit ||abs(p2-p1) > limit || abs(q2-q1) > limit movq mm1, mm5 ; p1 movq mm7, mm3 ; mm3=mm7=p0 psubusb mm7, mm5 ; p0 - p1 psubusb mm5, mm3 ; p1 - p0 por mm5, mm7 ; abs(p1-p0) movq t0, mm5 ; save abs(p1-p0) lea rdx, srct psubusb mm5, mm4 por mm0, mm5 ; mm0=mask movq mm5, [rdx+16] ; mm5=q0 movq mm7, [rdx+24] ; mm7=q1 movq mm6, mm5 ; mm6=q0 movq mm2, mm7 ; q1 psubusb mm5, mm7 ; q0-q1 psubusb mm7, mm6 ; q1-q0 por mm7, mm5 ; abs(q1-q0) movq t1, mm7 ; save abs(q1-q0) psubusb mm7, mm4 por mm0, mm7 ; mask movq mm5, mm2 ; q1 psubusb mm5, mm1 ; q1-=p1 psubusb mm1, mm2 ; p1-=q1 por mm5, mm1 ; abs(p1-q1) pand mm5, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm5, 1 ; abs(p1-q1)/2 mov rdx, arg(2) ;blimit ; movq mm4, [rdx] ;blimit movq mm1, mm3 ; mm1=mm3=p0 movq mm7, mm6 ; mm7=mm6=q0 psubusb mm1, mm7 ; p0-q0 psubusb mm7, mm3 ; q0-p0 por mm1, mm7 ; abs(q0-p0) paddusb mm1, mm1 ; abs(q0-p0)*2 paddusb mm1, mm5 ; abs (p0 - q0) *2 + abs(p1-q1)/2 psubusb mm1, mm4 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit por mm1, mm0; ; mask pxor mm0, mm0 pcmpeqb mm1, mm0 ; calculate high edge variance mov rdx, arg(4) ;thresh ; get thresh movq mm7, [rdx] ; movq mm4, t0 ; get abs (q1 - q0) psubusb mm4, mm7 movq mm3, t1 ; get abs (p1 - p0) psubusb mm3, mm7 por mm4, mm3 ; abs(q1 - q0) > thresh || abs(p1 - p0) > thresh pcmpeqb mm4, mm0 pcmpeqb mm0, mm0 pxor mm4, mm0 ; start work on filters lea rdx, srct movq mm2, [rdx] ; p1 movq mm7, [rdx+24] ; q1 movq mm6, [rdx+8] ; p0 movq mm0, [rdx+16] ; q0 pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm2, mm7 ; p1 - q1 pand mm2, mm4 ; high var mask (hvm)(p1 - q1) pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm0 ; q0 psubsb mm0, mm6 ; q0 - p0 paddsb mm2, mm0 ; 1 * (q0 - p0) + hvm(p1 - q1) paddsb mm2, mm0 ; 2 * (q0 - p0) + hvm(p1 - q1) paddsb mm2, mm0 ; 3 * (q0 - p0) + hvm(p1 - q1) pand mm1, mm2 ; mask filter values we don't care about movq mm2, mm1 paddsb mm1, [GLOBAL(t4)] ; 3* (q0 - p0) + hvm(p1 - q1) + 4 paddsb mm2, [GLOBAL(t3)] ; 3* (q0 - p0) + hvm(p1 - q1) + 3 pxor mm0, mm0 ; pxor mm5, mm5 punpcklbw mm0, mm2 ; punpckhbw mm5, mm2 ; psraw mm0, 11 ; psraw mm5, 11 packsswb mm0, mm5 movq mm2, mm0 ; (3* (q0 - p0) + hvm(p1 - q1) + 3) >> 3; pxor mm0, mm0 ; 0 movq mm5, mm1 ; abcdefgh punpcklbw mm0, mm1 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 pxor mm1, mm1 ; 0 punpckhbw mm1, mm5 ; a0b0c0d0 psraw mm1, 11 ; sign extended shift right by 3 movq mm5, mm0 ; save results packsswb mm0, mm1 ; (3* (q0 - p0) + hvm(p1 - q1) + 4) >>3 paddsw mm5, [GLOBAL(ones)] paddsw mm1, [GLOBAL(ones)] psraw mm5, 1 ; partial shifted one more time for 2nd tap psraw mm1, 1 ; partial shifted one more time for 2nd tap packsswb mm5, mm1 ; (3* (q0 - p0) + hvm(p1 - q1) + 4) >>4 pandn mm4, mm5 ; high edge variance additive paddsb mm6, mm2 ; p0+= p0 add pxor mm6, [GLOBAL(t80)] ; unoffset ; mm6=p0 ; movq mm1, [rdx] ; p1 pxor mm1, [GLOBAL(t80)] ; reoffset paddsb mm1, mm4 ; p1+= p1 add pxor mm1, [GLOBAL(t80)] ; unoffset ; mm6 = p0 mm1 = p1 psubsb mm3, mm0 ; q0-= q0 add pxor mm3, [GLOBAL(t80)] ; unoffset ; mm3 = q0 psubsb mm7, mm4 ; q1-= q1 add pxor mm7, [GLOBAL(t80)] ; unoffset ; mm7 = q1 ; tranpose and write back ; mm1 = 72 62 52 42 32 22 12 02 ; mm6 = 73 63 53 43 33 23 13 03 ; mm3 = 74 64 54 44 34 24 14 04 ; mm7 = 75 65 55 45 35 25 15 05 movq mm2, mm1 ; 72 62 52 42 32 22 12 02 punpcklbw mm2, mm6 ; 33 32 23 22 13 12 03 02 movq mm4, mm3 ; 74 64 54 44 34 24 14 04 punpckhbw mm1, mm6 ; 73 72 63 62 53 52 43 42 punpcklbw mm4, mm7 ; 35 34 25 24 15 14 05 04 punpckhbw mm3, mm7 ; 75 74 65 64 55 54 45 44 movq mm6, mm2 ; 33 32 23 22 13 12 03 02 punpcklwd mm2, mm4 ; 15 14 13 12 05 04 03 02 punpckhwd mm6, mm4 ; 35 34 33 32 25 24 23 22 movq mm5, mm1 ; 73 72 63 62 53 52 43 42 punpcklwd mm1, mm3 ; 55 54 53 52 45 44 43 42 punpckhwd mm5, mm3 ; 75 74 73 72 65 64 63 62 ; mm2 = 15 14 13 12 05 04 03 02 ; mm6 = 35 34 33 32 25 24 23 22 ; mm5 = 55 54 53 52 45 44 43 42 ; mm1 = 75 74 73 72 65 64 63 62 movd [rsi+rax*4+2], mm2 psrlq mm2, 32 movd [rdi+rax*4+2], mm2 movd [rsi+rax*2+2], mm6 psrlq mm6, 32 movd [rsi+rax+2],mm6 movd [rsi+2], mm1 psrlq mm1, 32 movd [rdi+2], mm1 neg rax movd [rdi+rax+2],mm5 psrlq mm5, 32 movd [rdi+rax*2+2], mm5 lea rsi, [rsi+rax*8] dec rcx jnz .next8_v add rsp, 64 pop rsp ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void vp8_mbloop_filter_horizontal_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit, ; const char *limit, ; const char *thresh, ; int count ;) global sym(vp8_mbloop_filter_horizontal_edge_mmx) PRIVATE sym(vp8_mbloop_filter_horizontal_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 6 GET_GOT rbx push rsi push rdi ; end prolog ALIGN_STACK 16, rax sub rsp, 32 ; reserve 32 bytes %define t0 [rsp + 0] ;__declspec(align(16)) char t0[8]; %define t1 [rsp + 16] ;__declspec(align(16)) char t1[8]; mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? movsxd rcx, dword ptr arg(5) ;count .next8_mbh: mov rdx, arg(3) ;limit movq mm7, [rdx] mov rdi, rsi ; rdi points to row +1 for indirect addressing add rdi, rax ; calculate breakout conditions movq mm2, [rdi+2*rax] ; q3 movq mm1, [rsi+2*rax] ; q2 movq mm6, mm1 ; q2 psubusb mm1, mm2 ; q2-=q3 psubusb mm2, mm6 ; q3-=q2 por mm1, mm2 ; abs(q3-q2) psubusb mm1, mm7 ; mm1 = abs(q3-q2), mm6 =q2, mm7 = limit movq mm4, [rsi+rax] ; q1 movq mm3, mm4 ; q1 psubusb mm4, mm6 ; q1-=q2 psubusb mm6, mm3 ; q2-=q1 por mm4, mm6 ; abs(q2-q1) psubusb mm4, mm7 por mm1, mm4 ; mm1 = mask, mm3=q1, mm7 = limit movq mm4, [rsi] ; q0 movq mm0, mm4 ; q0 psubusb mm4, mm3 ; q0-=q1 psubusb mm3, mm0 ; q1-=q0 por mm4, mm3 ; abs(q0-q1) movq t0, mm4 ; save to t0 psubusb mm4, mm7 por mm1, mm4 ; mm1 = mask, mm0=q0, mm7 = limit, t0 = abs(q0-q1) neg rax ; negate pitch to deal with above border movq mm2, [rsi+4*rax] ; p3 movq mm4, [rdi+4*rax] ; p2 movq mm5, mm4 ; p2 psubusb mm4, mm2 ; p2-=p3 psubusb mm2, mm5 ; p3-=p2 por mm4, mm2 ; abs(p3 - p2) psubusb mm4, mm7 por mm1, mm4 ; mm1 = mask, mm0=q0, mm7 = limit, t0 = abs(q0-q1) movq mm4, [rsi+2*rax] ; p1 movq mm3, mm4 ; p1 psubusb mm4, mm5 ; p1-=p2 psubusb mm5, mm3 ; p2-=p1 por mm4, mm5 ; abs(p2 - p1) psubusb mm4, mm7 por mm1, mm4 movq mm2, mm3 ; p1 ; mm1 = mask, mm0=q0, mm7 = limit, t0 = abs(q0-q1) movq mm4, [rsi+rax] ; p0 movq mm5, mm4 ; p0 psubusb mm4, mm3 ; p0-=p1 psubusb mm3, mm5 ; p1-=p0 por mm4, mm3 ; abs(p1 - p0) movq t1, mm4 ; save to t1 psubusb mm4, mm7 por mm1, mm4 ; mm1 = mask, mm0=q0, mm7 = limit, t0 = abs(q0-q1) t1 = abs(p1-p0) ; mm5 = p0 movq mm3, [rdi] ; q1 movq mm4, mm3 ; q1 psubusb mm3, mm2 ; q1-=p1 psubusb mm2, mm4 ; p1-=q1 por mm2, mm3 ; abs(p1-q1) pand mm2, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm2, 1 ; abs(p1-q1)/2 movq mm6, mm5 ; p0 movq mm3, mm0 ; q0 psubusb mm5, mm3 ; p0-=q0 psubusb mm3, mm6 ; q0-=p0 por mm5, mm3 ; abs(p0 - q0) paddusb mm5, mm5 ; abs(p0-q0)*2 paddusb mm5, mm2 ; abs (p0 - q0) *2 + abs(p1-q1)/2 mov rdx, arg(2) ;blimit ; get blimit movq mm7, [rdx] ; blimit psubusb mm5, mm7 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit por mm1, mm5 pxor mm5, mm5 pcmpeqb mm1, mm5 ; mask mm1 ; mm1 = mask, mm0=q0, mm7 = blimit, t0 = abs(q0-q1) t1 = abs(p1-p0) ; mm6 = p0, ; calculate high edge variance mov rdx, arg(4) ;thresh ; get thresh movq mm7, [rdx] ; movq mm4, t0 ; get abs (q1 - q0) psubusb mm4, mm7 movq mm3, t1 ; get abs (p1 - p0) psubusb mm3, mm7 paddb mm4, mm3 ; abs(q1 - q0) > thresh || abs(p1 - p0) > thresh pcmpeqb mm4, mm5 pcmpeqb mm5, mm5 pxor mm4, mm5 ; mm1 = mask, mm0=q0, mm7 = thresh, t0 = abs(q0-q1) t1 = abs(p1-p0) ; mm6 = p0, mm4=hev ; start work on filters movq mm2, [rsi+2*rax] ; p1 movq mm7, [rdi] ; q1 pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm2, mm7 ; p1 - q1 pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm0 ; q0 psubsb mm0, mm6 ; q0 - p0 paddsb mm2, mm0 ; 1 * (q0 - p0) + (p1 - q1) paddsb mm2, mm0 ; 2 * (q0 - p0) paddsb mm2, mm0 ; 3 * (q0 - p0) + (p1 - q1) pand mm1, mm2 ; mask filter values we don't care about ; mm1 = vp8_filter, mm4=hev, mm6=ps0, mm3=qs0 movq mm2, mm1 ; vp8_filter pand mm2, mm4; ; Filter2 = vp8_filter & hev movq mm5, mm2 ; paddsb mm5, [GLOBAL(t3)]; pxor mm0, mm0 ; 0 pxor mm7, mm7 ; 0 punpcklbw mm0, mm5 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 punpckhbw mm7, mm5 ; a0b0c0d0 psraw mm7, 11 ; sign extended shift right by 3 packsswb mm0, mm7 ; Filter2 >>=3; movq mm5, mm0 ; Filter2 paddsb mm2, [GLOBAL(t4)] ; vp8_signed_char_clamp(Filter2 + 4) pxor mm0, mm0 ; 0 pxor mm7, mm7 ; 0 punpcklbw mm0, mm2 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 punpckhbw mm7, mm2 ; a0b0c0d0 psraw mm7, 11 ; sign extended shift right by 3 packsswb mm0, mm7 ; Filter2 >>=3; ; mm0= filter2 mm1 = vp8_filter, mm3 =qs0 mm5=s mm4 =hev mm6=ps0 psubsb mm3, mm0 ; qs0 =qs0 - filter1 paddsb mm6, mm5 ; ps0 =ps0 + Fitler2 ; mm1=vp8_filter, mm3=qs0, mm4 =hev mm6=ps0 ; vp8_filter &= ~hev; ; Filter2 = vp8_filter; pandn mm4, mm1 ; vp8_filter&=~hev ; mm3=qs0, mm4=filter2, mm6=ps0 ; u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7); ; s = vp8_signed_char_clamp(qs0 - u); ; *oq0 = s^0x80; ; s = vp8_signed_char_clamp(ps0 + u); ; *op0 = s^0x80; pxor mm0, mm0 pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s27)] pmulhw mm2, [GLOBAL(s27)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 psubsb mm3, mm1 paddsb mm6, mm1 pxor mm3, [GLOBAL(t80)] pxor mm6, [GLOBAL(t80)] movq [rsi+rax], mm6 movq [rsi], mm3 ; roughly 2/7th difference across boundary ; u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7); ; s = vp8_signed_char_clamp(qs1 - u); ; *oq1 = s^0x80; ; s = vp8_signed_char_clamp(ps1 + u); ; *op1 = s^0x80; pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s18)] pmulhw mm2, [GLOBAL(s18)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 movq mm3, [rdi] movq mm6, [rsi+rax*2] ; p1 pxor mm3, [GLOBAL(t80)] pxor mm6, [GLOBAL(t80)] paddsb mm6, mm1 psubsb mm3, mm1 pxor mm6, [GLOBAL(t80)] pxor mm3, [GLOBAL(t80)] movq [rdi], mm3 movq [rsi+rax*2], mm6 ; roughly 1/7th difference across boundary ; u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7); ; s = vp8_signed_char_clamp(qs2 - u); ; *oq2 = s^0x80; ; s = vp8_signed_char_clamp(ps2 + u); ; *op2 = s^0x80; pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s9)] pmulhw mm2, [GLOBAL(s9)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 movq mm6, [rdi+rax*4] neg rax movq mm3, [rdi+rax ] pxor mm6, [GLOBAL(t80)] pxor mm3, [GLOBAL(t80)] paddsb mm6, mm1 psubsb mm3, mm1 pxor mm6, [GLOBAL(t80)] pxor mm3, [GLOBAL(t80)] movq [rdi+rax ], mm3 neg rax movq [rdi+rax*4], mm6 ;EARLY_BREAK_OUT: neg rax add rsi,8 dec rcx jnz .next8_mbh add rsp, 32 pop rsp ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void vp8_mbloop_filter_vertical_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit, ; const char *limit, ; const char *thresh, ; int count ;) global sym(vp8_mbloop_filter_vertical_edge_mmx) PRIVATE sym(vp8_mbloop_filter_vertical_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 6 GET_GOT rbx push rsi push rdi ; end prolog ALIGN_STACK 16, rax sub rsp, 96 ; reserve 96 bytes %define t0 [rsp + 0] ;__declspec(align(16)) char t0[8]; %define t1 [rsp + 16] ;__declspec(align(16)) char t1[8]; %define srct [rsp + 32] ;__declspec(align(16)) char srct[64]; mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? lea rsi, [rsi + rax*4 - 4] movsxd rcx, dword ptr arg(5) ;count .next8_mbv: lea rdi, [rsi + rax] ; rdi points to row +1 for indirect addressing ;transpose movq mm0, [rdi+2*rax] ; 77 76 75 74 73 72 71 70 movq mm6, [rsi+2*rax] ; 67 66 65 64 63 62 61 60 movq mm7, mm6 ; 77 76 75 74 73 72 71 70 punpckhbw mm7, mm0 ; 77 67 76 66 75 65 74 64 punpcklbw mm6, mm0 ; 73 63 72 62 71 61 70 60 movq mm0, [rsi+rax] ; 57 56 55 54 53 52 51 50 movq mm4, [rsi] ; 47 46 45 44 43 42 41 40 movq mm5, mm4 ; 47 46 45 44 43 42 41 40 punpckhbw mm5, mm0 ; 57 47 56 46 55 45 54 44 punpcklbw mm4, mm0 ; 53 43 52 42 51 41 50 40 movq mm3, mm5 ; 57 47 56 46 55 45 54 44 punpckhwd mm5, mm7 ; 77 67 57 47 76 66 56 46 punpcklwd mm3, mm7 ; 75 65 55 45 74 64 54 44 movq mm2, mm4 ; 53 43 52 42 51 41 50 40 punpckhwd mm4, mm6 ; 73 63 53 43 72 62 52 42 punpcklwd mm2, mm6 ; 71 61 51 41 70 60 50 40 neg rax movq mm7, [rsi+rax] ; 37 36 35 34 33 32 31 30 movq mm6, [rsi+rax*2] ; 27 26 25 24 23 22 21 20 movq mm1, mm6 ; 27 26 25 24 23 22 21 20 punpckhbw mm6, mm7 ; 37 27 36 36 35 25 34 24 punpcklbw mm1, mm7 ; 33 23 32 22 31 21 30 20 movq mm7, [rsi+rax*4]; ; 07 06 05 04 03 02 01 00 punpckhbw mm7, [rdi+rax*4] ; 17 07 16 06 15 05 14 04 movq mm0, mm7 ; 17 07 16 06 15 05 14 04 punpckhwd mm7, mm6 ; 37 27 17 07 36 26 16 06 punpcklwd mm0, mm6 ; 35 25 15 05 34 24 14 04 movq mm6, mm7 ; 37 27 17 07 36 26 16 06 punpckhdq mm7, mm5 ; 77 67 57 47 37 27 17 07 = q3 punpckldq mm6, mm5 ; 76 66 56 46 36 26 16 06 = q2 lea rdx, srct movq mm5, mm6 ; 76 66 56 46 36 26 16 06 movq [rdx+56], mm7 psubusb mm5, mm7 ; q2-q3 movq [rdx+48], mm6 psubusb mm7, mm6 ; q3-q2 por mm7, mm5; ; mm7=abs (q3-q2) movq mm5, mm0 ; 35 25 15 05 34 24 14 04 punpckhdq mm5, mm3 ; 75 65 55 45 35 25 15 05 = q1 punpckldq mm0, mm3 ; 74 64 54 44 34 24 15 04 = q0 movq mm3, mm5 ; 75 65 55 45 35 25 15 05 = q1 psubusb mm3, mm6 ; q1-q2 psubusb mm6, mm5 ; q2-q1 por mm6, mm3 ; mm6=abs(q2-q1) movq [rdx+40], mm5 ; save q1 movq [rdx+32], mm0 ; save q0 movq mm3, [rsi+rax*4] ; 07 06 05 04 03 02 01 00 punpcklbw mm3, [rdi+rax*4] ; 13 03 12 02 11 01 10 00 movq mm0, mm3 ; 13 03 12 02 11 01 10 00 punpcklwd mm0, mm1 ; 31 21 11 01 30 20 10 00 punpckhwd mm3, mm1 ; 33 23 13 03 32 22 12 02 movq mm1, mm0 ; 31 21 11 01 30 20 10 00 punpckldq mm0, mm2 ; 70 60 50 40 30 20 10 00 =p3 punpckhdq mm1, mm2 ; 71 61 51 41 31 21 11 01 =p2 movq [rdx], mm0 ; save p3 movq [rdx+8], mm1 ; save p2 movq mm2, mm1 ; 71 61 51 41 31 21 11 01 =p2 psubusb mm2, mm0 ; p2-p3 psubusb mm0, mm1 ; p3-p2 por mm0, mm2 ; mm0=abs(p3-p2) movq mm2, mm3 ; 33 23 13 03 32 22 12 02 punpckldq mm2, mm4 ; 72 62 52 42 32 22 12 02 = p1 punpckhdq mm3, mm4 ; 73 63 53 43 33 23 13 03 = p0 movq [rdx+24], mm3 ; save p0 movq [rdx+16], mm2 ; save p1 movq mm5, mm2 ; mm5 = p1 psubusb mm2, mm1 ; p1-p2 psubusb mm1, mm5 ; p2-p1 por mm1, mm2 ; mm1=abs(p2-p1) mov rdx, arg(3) ;limit movq mm4, [rdx] ; mm4 = limit psubusb mm7, mm4 ; abs(q3-q2) > limit psubusb mm0, mm4 ; abs(p3-p2) > limit psubusb mm1, mm4 ; abs(p2-p1) > limit psubusb mm6, mm4 ; abs(q2-q1) > limit por mm7, mm6 ; or por mm0, mm1 ; por mm0, mm7 ; abs(q3-q2) > limit || abs(p3-p2) > limit ||abs(p2-p1) > limit || abs(q2-q1) > limit movq mm1, mm5 ; p1 movq mm7, mm3 ; mm3=mm7=p0 psubusb mm7, mm5 ; p0 - p1 psubusb mm5, mm3 ; p1 - p0 por mm5, mm7 ; abs(p1-p0) movq t0, mm5 ; save abs(p1-p0) lea rdx, srct psubusb mm5, mm4 ; mm5 = abs(p1-p0) > limit por mm0, mm5 ; mm0=mask movq mm5, [rdx+32] ; mm5=q0 movq mm7, [rdx+40] ; mm7=q1 movq mm6, mm5 ; mm6=q0 movq mm2, mm7 ; q1 psubusb mm5, mm7 ; q0-q1 psubusb mm7, mm6 ; q1-q0 por mm7, mm5 ; abs(q1-q0) movq t1, mm7 ; save abs(q1-q0) psubusb mm7, mm4 ; mm7=abs(q1-q0)> limit por mm0, mm7 ; mask movq mm5, mm2 ; q1 psubusb mm5, mm1 ; q1-=p1 psubusb mm1, mm2 ; p1-=q1 por mm5, mm1 ; abs(p1-q1) pand mm5, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm5, 1 ; abs(p1-q1)/2 mov rdx, arg(2) ;blimit ; movq mm4, [rdx] ;blimit movq mm1, mm3 ; mm1=mm3=p0 movq mm7, mm6 ; mm7=mm6=q0 psubusb mm1, mm7 ; p0-q0 psubusb mm7, mm3 ; q0-p0 por mm1, mm7 ; abs(q0-p0) paddusb mm1, mm1 ; abs(q0-p0)*2 paddusb mm1, mm5 ; abs (p0 - q0) *2 + abs(p1-q1)/2 psubusb mm1, mm4 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit por mm1, mm0; ; mask pxor mm0, mm0 pcmpeqb mm1, mm0 ; calculate high edge variance mov rdx, arg(4) ;thresh ; get thresh movq mm7, [rdx] ; movq mm4, t0 ; get abs (q1 - q0) psubusb mm4, mm7 ; abs(q1 - q0) > thresh movq mm3, t1 ; get abs (p1 - p0) psubusb mm3, mm7 ; abs(p1 - p0)> thresh por mm4, mm3 ; abs(q1 - q0) > thresh || abs(p1 - p0) > thresh pcmpeqb mm4, mm0 pcmpeqb mm0, mm0 pxor mm4, mm0 ; start work on filters lea rdx, srct ; start work on filters movq mm2, [rdx+16] ; p1 movq mm7, [rdx+40] ; q1 pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm2, mm7 ; p1 - q1 movq mm6, [rdx+24] ; p0 movq mm0, [rdx+32] ; q0 pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm0 ; q0 psubsb mm0, mm6 ; q0 - p0 paddsb mm2, mm0 ; 1 * (q0 - p0) + (p1 - q1) paddsb mm2, mm0 ; 2 * (q0 - p0) paddsb mm2, mm0 ; 3 * (q0 - p0) + (p1 - q1) pand mm1, mm2 ; mask filter values we don't care about ; mm1 = vp8_filter, mm4=hev, mm6=ps0, mm3=qs0 movq mm2, mm1 ; vp8_filter pand mm2, mm4; ; Filter2 = vp8_filter & hev movq mm5, mm2 ; paddsb mm5, [GLOBAL(t3)]; pxor mm0, mm0 ; 0 pxor mm7, mm7 ; 0 punpcklbw mm0, mm5 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 punpckhbw mm7, mm5 ; a0b0c0d0 psraw mm7, 11 ; sign extended shift right by 3 packsswb mm0, mm7 ; Filter2 >>=3; movq mm5, mm0 ; Filter2 paddsb mm2, [GLOBAL(t4)] ; vp8_signed_char_clamp(Filter2 + 4) pxor mm0, mm0 ; 0 pxor mm7, mm7 ; 0 punpcklbw mm0, mm2 ; e0f0g0h0 psraw mm0, 11 ; sign extended shift right by 3 punpckhbw mm7, mm2 ; a0b0c0d0 psraw mm7, 11 ; sign extended shift right by 3 packsswb mm0, mm7 ; Filter2 >>=3; ; mm0= filter2 mm1 = vp8_filter, mm3 =qs0 mm5=s mm4 =hev mm6=ps0 psubsb mm3, mm0 ; qs0 =qs0 - filter1 paddsb mm6, mm5 ; ps0 =ps0 + Fitler2 ; mm1=vp8_filter, mm3=qs0, mm4 =hev mm6=ps0 ; vp8_filter &= ~hev; ; Filter2 = vp8_filter; pandn mm4, mm1 ; vp8_filter&=~hev ; mm3=qs0, mm4=filter2, mm6=ps0 ; u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7); ; s = vp8_signed_char_clamp(qs0 - u); ; *oq0 = s^0x80; ; s = vp8_signed_char_clamp(ps0 + u); ; *op0 = s^0x80; pxor mm0, mm0 pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s27)] pmulhw mm2, [GLOBAL(s27)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 psubsb mm3, mm1 paddsb mm6, mm1 pxor mm3, [GLOBAL(t80)] pxor mm6, [GLOBAL(t80)] movq [rdx+24], mm6 movq [rdx+32], mm3 ; roughly 2/7th difference across boundary ; u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7); ; s = vp8_signed_char_clamp(qs1 - u); ; *oq1 = s^0x80; ; s = vp8_signed_char_clamp(ps1 + u); ; *op1 = s^0x80; pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s18)] pmulhw mm2, [GLOBAL(s18)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 movq mm3, [rdx + 40] movq mm6, [rdx + 16] ; p1 pxor mm3, [GLOBAL(t80)] pxor mm6, [GLOBAL(t80)] paddsb mm6, mm1 psubsb mm3, mm1 pxor mm6, [GLOBAL(t80)] pxor mm3, [GLOBAL(t80)] movq [rdx + 40], mm3 movq [rdx + 16], mm6 ; roughly 1/7th difference across boundary ; u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7); ; s = vp8_signed_char_clamp(qs2 - u); ; *oq2 = s^0x80; ; s = vp8_signed_char_clamp(ps2 + u); ; *op2 = s^0x80; pxor mm1, mm1 pxor mm2, mm2 punpcklbw mm1, mm4 punpckhbw mm2, mm4 pmulhw mm1, [GLOBAL(s9)] pmulhw mm2, [GLOBAL(s9)] paddw mm1, [GLOBAL(s63)] paddw mm2, [GLOBAL(s63)] psraw mm1, 7 psraw mm2, 7 packsswb mm1, mm2 movq mm6, [rdx+ 8] movq mm3, [rdx+48] pxor mm6, [GLOBAL(t80)] pxor mm3, [GLOBAL(t80)] paddsb mm6, mm1 psubsb mm3, mm1 pxor mm6, [GLOBAL(t80)] ; mm6 = 71 61 51 41 31 21 11 01 pxor mm3, [GLOBAL(t80)] ; mm3 = 76 66 56 46 36 26 15 06 ; tranpose and write back movq mm0, [rdx] ; mm0 = 70 60 50 40 30 20 10 00 movq mm1, mm0 ; mm0 = 70 60 50 40 30 20 10 00 punpcklbw mm0, mm6 ; mm0 = 31 30 21 20 11 10 01 00 punpckhbw mm1, mm6 ; mm3 = 71 70 61 60 51 50 41 40 movq mm2, [rdx+16] ; mm2 = 72 62 52 42 32 22 12 02 movq mm6, mm2 ; mm3 = 72 62 52 42 32 22 12 02 punpcklbw mm2, [rdx+24] ; mm2 = 33 32 23 22 13 12 03 02 punpckhbw mm6, [rdx+24] ; mm3 = 73 72 63 62 53 52 43 42 movq mm5, mm0 ; mm5 = 31 30 21 20 11 10 01 00 punpcklwd mm0, mm2 ; mm0 = 13 12 11 10 03 02 01 00 punpckhwd mm5, mm2 ; mm5 = 33 32 31 30 23 22 21 20 movq mm4, mm1 ; mm4 = 71 70 61 60 51 50 41 40 punpcklwd mm1, mm6 ; mm1 = 53 52 51 50 43 42 41 40 punpckhwd mm4, mm6 ; mm4 = 73 72 71 70 63 62 61 60 movq mm2, [rdx+32] ; mm2 = 74 64 54 44 34 24 14 04 punpcklbw mm2, [rdx+40] ; mm2 = 35 34 25 24 15 14 05 04 movq mm6, mm3 ; mm6 = 76 66 56 46 36 26 15 06 punpcklbw mm6, [rdx+56] ; mm6 = 37 36 27 26 17 16 07 06 movq mm7, mm2 ; mm7 = 35 34 25 24 15 14 05 04 punpcklwd mm2, mm6 ; mm2 = 17 16 15 14 07 06 05 04 punpckhwd mm7, mm6 ; mm7 = 37 36 35 34 27 26 25 24 movq mm6, mm0 ; mm6 = 13 12 11 10 03 02 01 00 punpckldq mm0, mm2 ; mm0 = 07 06 05 04 03 02 01 00 punpckhdq mm6, mm2 ; mm6 = 17 16 15 14 13 12 11 10 movq [rsi+rax*4], mm0 ; write out movq [rdi+rax*4], mm6 ; write out movq mm0, mm5 ; mm0 = 33 32 31 30 23 22 21 20 punpckldq mm0, mm7 ; mm0 = 27 26 25 24 23 22 20 20 punpckhdq mm5, mm7 ; mm5 = 37 36 35 34 33 32 31 30 movq [rsi+rax*2], mm0 ; write out movq [rdi+rax*2], mm5 ; write out movq mm2, [rdx+32] ; mm2 = 74 64 54 44 34 24 14 04 punpckhbw mm2, [rdx+40] ; mm2 = 75 74 65 64 54 54 45 44 punpckhbw mm3, [rdx+56] ; mm3 = 77 76 67 66 57 56 47 46 movq mm5, mm2 ; mm5 = 75 74 65 64 54 54 45 44 punpcklwd mm2, mm3 ; mm2 = 57 56 55 54 47 46 45 44 punpckhwd mm5, mm3 ; mm5 = 77 76 75 74 67 66 65 64 movq mm0, mm1 ; mm0= 53 52 51 50 43 42 41 40 movq mm3, mm4 ; mm4 = 73 72 71 70 63 62 61 60 punpckldq mm0, mm2 ; mm0 = 47 46 45 44 43 42 41 40 punpckhdq mm1, mm2 ; mm1 = 57 56 55 54 53 52 51 50 movq [rsi], mm0 ; write out movq [rdi], mm1 ; write out neg rax punpckldq mm3, mm5 ; mm3 = 67 66 65 64 63 62 61 60 punpckhdq mm4, mm5 ; mm4 = 77 76 75 74 73 72 71 60 movq [rsi+rax*2], mm3 movq [rdi+rax*2], mm4 lea rsi, [rsi+rax*8] dec rcx jnz .next8_mbv add rsp, 96 pop rsp ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void vp8_loop_filter_simple_horizontal_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit ;) global sym(vp8_loop_filter_simple_horizontal_edge_mmx) PRIVATE sym(vp8_loop_filter_simple_horizontal_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 3 GET_GOT rbx push rsi push rdi ; end prolog mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? mov rcx, 2 ; count .nexts8_h: mov rdx, arg(2) ;blimit ; get blimit movq mm3, [rdx] ; mov rdi, rsi ; rdi points to row +1 for indirect addressing add rdi, rax neg rax ; calculate mask movq mm1, [rsi+2*rax] ; p1 movq mm0, [rdi] ; q1 movq mm2, mm1 movq mm7, mm0 movq mm4, mm0 psubusb mm0, mm1 ; q1-=p1 psubusb mm1, mm4 ; p1-=q1 por mm1, mm0 ; abs(p1-q1) pand mm1, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm1, 1 ; abs(p1-q1)/2 movq mm5, [rsi+rax] ; p0 movq mm4, [rsi] ; q0 movq mm0, mm4 ; q0 movq mm6, mm5 ; p0 psubusb mm5, mm4 ; p0-=q0 psubusb mm4, mm6 ; q0-=p0 por mm5, mm4 ; abs(p0 - q0) paddusb mm5, mm5 ; abs(p0-q0)*2 paddusb mm5, mm1 ; abs (p0 - q0) *2 + abs(p1-q1)/2 psubusb mm5, mm3 ; abs(p0 - q0) *2 + abs(p1-q1)/2 > blimit pxor mm3, mm3 pcmpeqb mm5, mm3 ; start work on filters pxor mm2, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm2, mm7 ; p1 - q1 pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm0, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm0 ; q0 psubsb mm0, mm6 ; q0 - p0 paddsb mm2, mm0 ; p1 - q1 + 1 * (q0 - p0) paddsb mm2, mm0 ; p1 - q1 + 2 * (q0 - p0) paddsb mm2, mm0 ; p1 - q1 + 3 * (q0 - p0) pand mm5, mm2 ; mask filter values we don't care about ; do + 4 side paddsb mm5, [GLOBAL(t4)] ; 3* (q0 - p0) + (p1 - q1) + 4 movq mm0, mm5 ; get a copy of filters psllw mm0, 8 ; shift left 8 psraw mm0, 3 ; arithmetic shift right 11 psrlw mm0, 8 movq mm1, mm5 ; get a copy of filters psraw mm1, 11 ; arithmetic shift right 11 psllw mm1, 8 ; shift left 8 to put it back por mm0, mm1 ; put the two together to get result psubsb mm3, mm0 ; q0-= q0 add pxor mm3, [GLOBAL(t80)] ; unoffset movq [rsi], mm3 ; write back ; now do +3 side psubsb mm5, [GLOBAL(t1s)] ; +3 instead of +4 movq mm0, mm5 ; get a copy of filters psllw mm0, 8 ; shift left 8 psraw mm0, 3 ; arithmetic shift right 11 psrlw mm0, 8 psraw mm5, 11 ; arithmetic shift right 11 psllw mm5, 8 ; shift left 8 to put it back por mm0, mm5 ; put the two together to get result paddsb mm6, mm0 ; p0+= p0 add pxor mm6, [GLOBAL(t80)] ; unoffset movq [rsi+rax], mm6 ; write back add rsi,8 neg rax dec rcx jnz .nexts8_h ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void vp8_loop_filter_simple_vertical_edge_mmx ;( ; unsigned char *src_ptr, ; int src_pixel_step, ; const char *blimit ;) global sym(vp8_loop_filter_simple_vertical_edge_mmx) PRIVATE sym(vp8_loop_filter_simple_vertical_edge_mmx): push rbp mov rbp, rsp SHADOW_ARGS_TO_STACK 3 GET_GOT rbx push rsi push rdi ; end prolog ALIGN_STACK 16, rax sub rsp, 32 ; reserve 32 bytes %define t0 [rsp + 0] ;__declspec(align(16)) char t0[8]; %define t1 [rsp + 16] ;__declspec(align(16)) char t1[8]; mov rsi, arg(0) ;src_ptr movsxd rax, dword ptr arg(1) ;src_pixel_step ; destination pitch? lea rsi, [rsi + rax*4- 2]; ; mov rcx, 2 ; count .nexts8_v: lea rdi, [rsi + rax]; movd mm0, [rdi + rax * 2] ; xx xx xx xx 73 72 71 70 movd mm6, [rsi + rax * 2] ; xx xx xx xx 63 62 61 60 punpcklbw mm6, mm0 ; 73 63 72 62 71 61 70 60 movd mm0, [rsi + rax] ; xx xx xx xx 53 52 51 50 movd mm4, [rsi] ; xx xx xx xx 43 42 41 40 punpcklbw mm4, mm0 ; 53 43 52 42 51 41 50 40 movq mm5, mm4 ; 53 43 52 42 51 41 50 40 punpcklwd mm4, mm6 ; 71 61 51 41 70 60 50 40 punpckhwd mm5, mm6 ; 73 63 53 43 72 62 52 42 neg rax movd mm7, [rsi + rax] ; xx xx xx xx 33 32 31 30 movd mm6, [rsi + rax * 2] ; xx xx xx xx 23 22 21 20 punpcklbw mm6, mm7 ; 33 23 32 22 31 21 30 20 movd mm1, [rdi + rax * 4] ; xx xx xx xx 13 12 11 10 movd mm0, [rsi + rax * 4] ; xx xx xx xx 03 02 01 00 punpcklbw mm0, mm1 ; 13 03 12 02 11 01 10 00 movq mm2, mm0 ; 13 03 12 02 11 01 10 00 punpcklwd mm0, mm6 ; 31 21 11 01 30 20 10 00 punpckhwd mm2, mm6 ; 33 23 13 03 32 22 12 02 movq mm1, mm0 ; 13 03 12 02 11 01 10 00 punpckldq mm0, mm4 ; 70 60 50 40 30 20 10 00 = p1 movq mm3, mm2 ; 33 23 13 03 32 22 12 02 punpckhdq mm1, mm4 ; 71 61 51 41 31 21 11 01 = p0 punpckldq mm2, mm5 ; 72 62 52 42 32 22 12 02 = q0 punpckhdq mm3, mm5 ; 73 63 53 43 33 23 13 03 = q1 ; calculate mask movq mm6, mm0 ; p1 movq mm7, mm3 ; q1 psubusb mm7, mm6 ; q1-=p1 psubusb mm6, mm3 ; p1-=q1 por mm6, mm7 ; abs(p1-q1) pand mm6, [GLOBAL(tfe)] ; set lsb of each byte to zero psrlw mm6, 1 ; abs(p1-q1)/2 movq mm5, mm1 ; p0 movq mm4, mm2 ; q0 psubusb mm5, mm2 ; p0-=q0 psubusb mm4, mm1 ; q0-=p0 por mm5, mm4 ; abs(p0 - q0) paddusb mm5, mm5 ; abs(p0-q0)*2 paddusb mm5, mm6 ; abs (p0 - q0) *2 + abs(p1-q1)/2 mov rdx, arg(2) ;blimit ; get blimit movq mm7, [rdx] psubusb mm5, mm7 ; abs(p0 - q0) *2 + abs(p1-q1)/2 > blimit pxor mm7, mm7 pcmpeqb mm5, mm7 ; mm5 = mask ; start work on filters movq t0, mm0 movq t1, mm3 pxor mm0, [GLOBAL(t80)] ; p1 offset to convert to signed values pxor mm3, [GLOBAL(t80)] ; q1 offset to convert to signed values psubsb mm0, mm3 ; p1 - q1 movq mm6, mm1 ; p0 movq mm7, mm2 ; q0 pxor mm6, [GLOBAL(t80)] ; offset to convert to signed values pxor mm7, [GLOBAL(t80)] ; offset to convert to signed values movq mm3, mm7 ; offseted ; q0 psubsb mm7, mm6 ; q0 - p0 paddsb mm0, mm7 ; p1 - q1 + 1 * (q0 - p0) paddsb mm0, mm7 ; p1 - q1 + 2 * (q0 - p0) paddsb mm0, mm7 ; p1 - q1 + 3 * (q0 - p0) pand mm5, mm0 ; mask filter values we don't care about paddsb mm5, [GLOBAL(t4)] ; 3* (q0 - p0) + (p1 - q1) + 4 movq mm0, mm5 ; get a copy of filters psllw mm0, 8 ; shift left 8 psraw mm0, 3 ; arithmetic shift right 11 psrlw mm0, 8 movq mm7, mm5 ; get a copy of filters psraw mm7, 11 ; arithmetic shift right 11 psllw mm7, 8 ; shift left 8 to put it back por mm0, mm7 ; put the two together to get result psubsb mm3, mm0 ; q0-= q0sz add pxor mm3, [GLOBAL(t80)] ; unoffset ; now do +3 side psubsb mm5, [GLOBAL(t1s)] ; +3 instead of +4 movq mm0, mm5 ; get a copy of filters psllw mm0, 8 ; shift left 8 psraw mm0, 3 ; arithmetic shift right 11 psrlw mm0, 8 psraw mm5, 11 ; arithmetic shift right 11 psllw mm5, 8 ; shift left 8 to put it back por mm0, mm5 ; put the two together to get result paddsb mm6, mm0 ; p0+= p0 add pxor mm6, [GLOBAL(t80)] ; unoffset movq mm0, t0 movq mm4, t1 ; mm0 = 70 60 50 40 30 20 10 00 ; mm6 = 71 61 51 41 31 21 11 01 ; mm3 = 72 62 52 42 32 22 12 02 ; mm4 = 73 63 53 43 33 23 13 03 ; transpose back to write out movq mm1, mm0 ; punpcklbw mm0, mm6 ; 31 30 21 20 11 10 01 00 punpckhbw mm1, mm6 ; 71 70 61 60 51 50 41 40 movq mm2, mm3 ; punpcklbw mm2, mm4 ; 33 32 23 22 13 12 03 02 movq mm5, mm1 ; 71 70 61 60 51 50 41 40 punpckhbw mm3, mm4 ; 73 72 63 62 53 52 43 42 movq mm6, mm0 ; 31 30 21 20 11 10 01 00 punpcklwd mm0, mm2 ; 13 12 11 10 03 02 01 00 punpckhwd mm6, mm2 ; 33 32 31 30 23 22 21 20 movd [rsi+rax*4], mm0 ; write 03 02 01 00 punpcklwd mm1, mm3 ; 53 52 51 50 43 42 41 40 psrlq mm0, 32 ; xx xx xx xx 13 12 11 10 punpckhwd mm5, mm3 ; 73 72 71 70 63 62 61 60 movd [rdi+rax*4], mm0 ; write 13 12 11 10 movd [rsi+rax*2], mm6 ; write 23 22 21 20 psrlq mm6, 32 ; 33 32 31 30 movd [rsi], mm1 ; write 43 42 41 40 movd [rsi + rax], mm6 ; write 33 32 31 30 neg rax movd [rsi + rax*2], mm5 ; write 63 62 61 60 psrlq mm1, 32 ; 53 52 51 50 movd [rdi], mm1 ; write out 53 52 51 50 psrlq mm5, 32 ; 73 72 71 70 movd [rdi + rax*2], mm5 ; write 73 72 71 70 lea rsi, [rsi+rax*8] ; next 8 dec rcx jnz .nexts8_v add rsp, 32 pop rsp ; begin epilog pop rdi pop rsi RESTORE_GOT UNSHADOW_ARGS pop rbp ret ;void fast_loop_filter_vertical_edges_mmx(unsigned char *y_ptr, ; int y_stride, ; loop_filter_info *lfi) ;{ ; ; ; vp8_loop_filter_simple_vertical_edge_mmx(y_ptr+4, y_stride, lfi->flim,lfi->lim,lfi->thr,2); ; vp8_loop_filter_simple_vertical_edge_mmx(y_ptr+8, y_stride, lfi->flim,lfi->lim,lfi->thr,2); ; vp8_loop_filter_simple_vertical_edge_mmx(y_ptr+12, y_stride, lfi->flim,lfi->lim,lfi->thr,2); ;} SECTION_RODATA align 16 tfe: times 8 db 0xfe align 16 t80: times 8 db 0x80 align 16 t1s: times 8 db 0x01 align 16 t3: times 8 db 0x03 align 16 t4: times 8 db 0x04 align 16 ones: times 4 dw 0x0001 align 16 s27: times 4 dw 0x1b00 align 16 s18: times 4 dw 0x1200 align 16 s9: times 4 dw 0x0900 align 16 s63: times 4 dw 0x003f