shithub: libvpx

Download patch

ref: a15bebfc0a8628d60e6aa8088f83d8a40428bf14
parent: cdea4a7c6646b572274f84cf0791012966edca89
author: Johann <johann.koenig@duck.com>
date: Fri Jul 12 12:12:58 EDT 2013

vp9_convolve8_[horiz|vert]_avg

Super basic conversion from the other implementations. Any changes to
one should be trivial to copy over keep in sync.

Change-Id: I1720b4128e0aba4b2779e3761f6494f8a09d3ea8

--- a/test/convolve_test.cc
+++ b/test/convolve_test.cc
@@ -549,8 +549,8 @@
 
 #if HAVE_NEON
 const ConvolveFunctions convolve8_neon(
-    vp9_convolve8_horiz_neon, vp9_convolve8_avg_horiz_c,
-    vp9_convolve8_vert_neon, vp9_convolve8_avg_vert_c,
+    vp9_convolve8_horiz_neon, vp9_convolve8_avg_horiz_neon,
+    vp9_convolve8_vert_neon, vp9_convolve8_avg_vert_neon,
     vp9_convolve8_c, vp9_convolve8_avg_c);
 
 INSTANTIATE_TEST_CASE_P(NEON, ConvolveTest, ::testing::Values(
--- /dev/null
+++ b/vp9/common/arm/neon/vp9_convolve8_avg_neon.asm
@@ -1,0 +1,277 @@
+;
+;  Copyright (c) 2013 The WebM project authors. All Rights Reserved.
+;
+;  Use of this source code is governed by a BSD-style license
+;  that can be found in the LICENSE file in the root of the source
+;  tree. An additional intellectual property rights grant can be found
+;  in the file PATENTS.  All contributing project authors may
+;  be found in the AUTHORS file in the root of the source tree.
+;
+
+
+    ; These functions are only valid when:
+    ; x_step_q4 == 16
+    ; w%4 == 0
+    ; h%4 == 0
+    ; taps == 8
+    ; VP9_FILTER_WEIGHT == 128
+    ; VP9_FILTER_SHIFT == 7
+
+    EXPORT  |vp9_convolve8_avg_horiz_neon|
+    EXPORT  |vp9_convolve8_avg_vert_neon|
+    IMPORT  |vp9_convolve8_avg_horiz_c|
+    IMPORT  |vp9_convolve8_avg_vert_c|
+    ARM
+    REQUIRE8
+    PRESERVE8
+
+    AREA ||.text||, CODE, READONLY, ALIGN=2
+
+    ; Multiply and accumulate by q0
+    MACRO
+    MULTIPLY_BY_Q0 $dst, $src0, $src1, $src2, $src3, $src4, $src5, $src6, $src7
+    vmull.s16 $dst, $src0, d0[0]
+    vmlal.s16 $dst, $src1, d0[1]
+    vmlal.s16 $dst, $src2, d0[2]
+    vmlal.s16 $dst, $src3, d0[3]
+    vmlal.s16 $dst, $src4, d1[0]
+    vmlal.s16 $dst, $src5, d1[1]
+    vmlal.s16 $dst, $src6, d1[2]
+    vmlal.s16 $dst, $src7, d1[3]
+    MEND
+
+; r0    const uint8_t *src
+; r1    int src_stride
+; r2    uint8_t *dst
+; r3    int dst_stride
+; sp[]const int16_t *filter_x
+; sp[]int x_step_q4
+; sp[]const int16_t *filter_y ; unused
+; sp[]int y_step_q4           ; unused
+; sp[]int w
+; sp[]int h
+
+|vp9_convolve8_avg_horiz_neon| PROC
+    push            {r4-r10, lr}
+
+    sub             r0, r0, #3              ; adjust for taps
+
+    ldr             r4, [sp, #36]           ; x_step_q4
+    ldr             r5, [sp, #32]           ; filter_x
+    cmp             r4, #16
+    bne             call_horiz_c_convolve   ; x_step_q4 != 16
+
+    ldr             r6, [sp, #48]           ; w
+    ldr             r7, [sp, #52]           ; h
+
+    vld1.s16        {q0}, [r5]              ; filter_x
+
+    add             r8, r1, r1, lsl #1      ; src_stride * 3
+    add             r8, r8, #4              ; src_stride * 3 + 4
+    rsb             r8, r8, #0              ; reset for src
+
+    add             r4, r3, r3, lsl #1      ; dst_stride * 3
+    sub             r4, r4, #4              ; dst_stride * 3 - 4
+    rsb             r4, r4, #0              ; reset for dst
+
+    sub             r9, r1, #8              ; post increment for src load
+
+    rsb             r1, r6, r1, lsl #2      ; reset src for outer loop
+    rsb             r12, r6, r3, lsl #2     ; reset dst for outer loop
+
+    mov             r10, r6                 ; w loop counter
+
+loop_horiz
+    vld4.u8         {d24[0], d25[0], d26[0], d27[0]}, [r0]!
+    vld4.u8         {d24[4], d25[4], d26[4], d27[4]}, [r0]!
+    vld3.u8         {d28[0], d29[0], d30[0]}, [r0], r9
+
+    vld4.u8         {d24[1], d25[1], d26[1], d27[1]}, [r0]!
+    vld4.u8         {d24[5], d25[5], d26[5], d27[5]}, [r0]!
+    vld3.u8         {d28[1], d29[1], d30[1]}, [r0], r9
+
+    vld4.u8         {d24[2], d25[2], d26[2], d27[2]}, [r0]!
+    vld4.u8         {d24[6], d25[6], d26[6], d27[6]}, [r0]!
+    vld3.u8         {d28[2], d29[2], d30[2]}, [r0], r9
+
+    vld4.u8         {d24[3], d25[3], d26[3], d27[3]}, [r0]!
+    vld4.u8         {d24[7], d25[7], d26[7], d27[7]}, [r0]!
+    vld3.u8         {d28[3], d29[3], d30[3]}, [r0], r8
+
+    ; extract to s16
+    vmovl.u8        q8, d24
+    vmovl.u8        q9, d25
+    vmovl.u8        q10, d26
+    vmovl.u8        q11, d27
+    vtrn.32         d28, d29 ; only the first half is populated
+    vmovl.u8        q12, d28
+    vmovl.u8        q13, d30
+
+    ; slightly out of order load to match the existing data
+    vld1.u32        {d6[0]}, [r2], r3
+    vld1.u32        {d7[0]}, [r2], r3
+    vld1.u32        {d6[1]}, [r2], r3
+    vld1.u32        {d7[1]}, [r2], r3
+
+    sub             r2, r2, r3, lsl #2      ; reset for store
+
+    ; src[] * filter_x
+    MULTIPLY_BY_Q0 q1, d16, d18, d20, d22, d17, d19, d21, d23
+    MULTIPLY_BY_Q0 q2, d18, d20, d22, d17, d19, d21, d23, d24
+    MULTIPLY_BY_Q0 q14, d20, d22, d17, d19, d21, d23, d24, d25
+    MULTIPLY_BY_Q0 q15, d22, d17, d19, d21, d23, d24, d25, d26
+
+    ; += 64 >> 7
+    vqrshrun.s32    d2, q1, #7
+    vqrshrun.s32    d3, q2, #7
+    vqrshrun.s32    d4, q14, #7
+    vqrshrun.s32    d5, q15, #7
+
+    ; saturate
+    vqshrn.u16      d2, q1, #0
+    vqshrn.u16      d3, q2, #0
+
+    ; transpose
+    vtrn.16         d2, d3
+    vtrn.32         d2, d3
+    vtrn.8          d2, d3
+    
+    ; average the new value and the dst value
+    vaddl.u8        q8, d2, d6
+    vaddl.u8        q9, d3, d7
+    vqrshrn.u16     d2, q8, #1
+    vqrshrn.u16     d3, q9, #1
+
+    vst1.u32        {d2[0]}, [r2], r3
+    vst1.u32        {d3[0]}, [r2], r3
+    vst1.u32        {d2[1]}, [r2], r3
+    vst1.u32        {d3[1]}, [r2], r4
+
+    subs            r6, r6, #4              ; w -= 4
+    bgt             loop_horiz
+
+    ; outer loop
+    mov             r6, r10                 ; restore w counter
+    add             r0, r0, r1              ; src += src_stride * 4 - w
+    add             r2, r2, r12             ; dst += dst_stride * 4 - w
+    subs            r7, r7, #4              ; h -= 4
+    bgt loop_horiz
+
+    pop             {r4-r10, pc}
+
+call_horiz_c_convolve
+    pop             {r4-r10, lr}
+    add             r0, r0, #3              ; un-adjust for taps
+    b               vp9_convolve8_avg_horiz_c
+
+
+    ENDP
+
+|vp9_convolve8_avg_vert_neon| PROC
+    push            {r4-r10, lr}
+
+    ; adjust for taps
+    sub             r0, r0, r1
+    sub             r0, r0, r1, lsl #1
+
+    ldr             r6, [sp, #44]           ; y_step_q4
+    ldr             r7, [sp, #40]           ; filter_y
+    cmp             r6, #16
+    bne             call_vert_c_convolve    ; y_step_q4 != 16
+
+    ldr             r8, [sp, #48]           ; w
+    ldr             r9, [sp, #52]           ; h
+
+    vld1.s16        {q0}, [r7]              ; filter_y
+
+    mov             r5, r1, lsl #1          ; src_stride * 2
+    add             r5, r5, r1, lsl #3      ; src_stride * 10
+    sub             r5, r5, #4              ; src_stride * 10 + 4
+    rsb             r5, r5, #0              ; reset for src
+
+    add             r6, r3, r3, lsl #1      ; dst_stride * 3
+    sub             r6, r6, #4              ; dst_stride * 3 - 4
+    rsb             r6, r6, #0              ; reset for dst
+
+    rsb             r7, r8, r1, lsl #2      ; reset src for outer loop
+    rsb             r12, r8, r3, lsl #2     ; reset dst for outer loop
+
+    mov             r10, r8                 ; w loop counter
+
+loop_vert
+    ; always process a 4x4 block at a time
+    vld1.u32        {d16[0]}, [r0], r1
+    vld1.u32        {d16[1]}, [r0], r1
+    vld1.u32        {d18[0]}, [r0], r1
+    vld1.u32        {d18[1]}, [r0], r1
+    vld1.u32        {d20[0]}, [r0], r1
+    vld1.u32        {d20[1]}, [r0], r1
+    vld1.u32        {d22[0]}, [r0], r1
+    vld1.u32        {d22[1]}, [r0], r1
+    vld1.u32        {d24[0]}, [r0], r1
+    vld1.u32        {d24[1]}, [r0], r1
+    vld1.u32        {d26[0]}, [r0], r5
+
+    ; extract to s16
+    vmovl.u8        q8, d16
+    vmovl.u8        q9, d18
+    vmovl.u8        q10, d20
+    vmovl.u8        q11, d22
+    vmovl.u8        q12, d24
+    vmovl.u8        q13, d26
+
+    vld1.u32        {d6[0]}, [r2], r3
+    vld1.u32        {d6[1]}, [r2], r3
+    vld1.u32        {d7[0]}, [r2], r3
+    vld1.u32        {d7[1]}, [r2], r3
+
+    sub             r2, r2, r3, lsl #2      ; reset for store
+
+    ; src[] * filter_y
+    MULTIPLY_BY_Q0 q1, d16, d17, d18, d19, d20, d21, d22, d23
+    MULTIPLY_BY_Q0 q2, d17, d18, d19, d20, d21, d22, d23, d24
+    MULTIPLY_BY_Q0 q14, d18, d19, d20, d21, d22, d23, d24, d25
+    MULTIPLY_BY_Q0 q15, d19, d20, d21, d22, d23, d24, d25, d26
+
+    ; += 64 >> 7
+    vqrshrun.s32    d2, q1, #7
+    vqrshrun.s32    d3, q2, #7
+    vqrshrun.s32    d4, q14, #7
+    vqrshrun.s32    d5, q15, #7
+
+    ; saturate
+    vqshrn.u16      d2, q1, #0
+    vqshrn.u16      d3, q2, #0
+
+    ; average the new value and the dst value
+    vaddl.u8        q8, d2, d6
+    vaddl.u8        q9, d3, d7
+    vqrshrn.u16     d2, q8, #1
+    vqrshrn.u16     d3, q9, #1
+
+    vst1.u32        {d2[0]}, [r2], r3
+    vst1.u32        {d2[1]}, [r2], r3
+    vst1.u32        {d3[0]}, [r2], r3
+    vst1.u32        {d3[1]}, [r2], r6
+
+    subs            r8, r8, #4              ; w -= 4
+    bgt             loop_vert
+
+    ; outer loop
+    mov             r8, r10                 ; restore w counter
+    add             r0, r0, r7              ; src += 4 * src_stride - w
+    add             r2, r2, r12             ; dst += 4 * dst_stride - w
+    subs            r9, r9, #4              ; h -= 4
+    bgt             loop_vert
+
+    pop             {r4-r10, pc}
+
+call_vert_c_convolve
+    pop             {r4-r10, lr}
+    ; un-adjust for taps
+    add             r0, r0, r1
+    add             r0, r0, r1, lsl #1
+    b               vp9_convolve8_avg_vert_c
+
+    ENDP
+    END
--- a/vp9/common/vp9_rtcd_defs.sh
+++ b/vp9/common/vp9_rtcd_defs.sh
@@ -283,10 +283,10 @@
 specialize vp9_convolve8_avg ssse3
 
 prototype void vp9_convolve8_avg_horiz "const uint8_t *src, ptrdiff_t src_stride, uint8_t *dst, ptrdiff_t dst_stride, const int16_t *filter_x, int x_step_q4, const int16_t *filter_y, int y_step_q4, int w, int h"
-specialize vp9_convolve8_avg_horiz ssse3
+specialize vp9_convolve8_avg_horiz ssse3 neon
 
 prototype void vp9_convolve8_avg_vert "const uint8_t *src, ptrdiff_t src_stride, uint8_t *dst, ptrdiff_t dst_stride, const int16_t *filter_x, int x_step_q4, const int16_t *filter_y, int y_step_q4, int w, int h"
-specialize vp9_convolve8_avg_vert ssse3
+specialize vp9_convolve8_avg_vert ssse3 neon
 
 #
 # dct
--- a/vp9/vp9_common.mk
+++ b/vp9/vp9_common.mk
@@ -89,6 +89,7 @@
 VP9_COMMON_SRCS-$(HAVE_SSE2) += common/x86/vp9_idct_intrin_sse2.c
 
 VP9_COMMON_SRCS-$(HAVE_NEON) += common/arm/neon/vp9_convolve8_neon$(ASM)
+VP9_COMMON_SRCS-$(HAVE_NEON) += common/arm/neon/vp9_convolve8_avg_neon$(ASM)
 VP9_COMMON_SRCS-$(HAVE_NEON) += common/arm/neon/vp9_loopfilter_neon$(ASM)
 VP9_COMMON_SRCS-$(HAVE_NEON) += common/arm/neon/vp9_dc_only_idct_add_neon$(ASM)