shithub: libvpx

ref: 4d018be950ef8b056a7c797a22ee58012443df26
dir: /vp9/decoder/arm/neon/vp9_add_constant_residual_neon.asm/

View raw version
;
;   Copyright (c) 2013 The WebM project authors. All Rights Reserved.
;
;   Use of this source code is governed by a BSD-style license
;   that can be found in the LICENSE file in the root of the source
;   tree. An additional intellectual property rights grant can be found
;   in the file PATENTS.  All contributing project authors may
;   be found in the AUTHORS file in the root of the source tree.
;

    EXPORT |vp9_add_constant_residual_8x8_neon|
    EXPORT |vp9_add_constant_residual_16x16_neon|
    EXPORT |vp9_add_constant_residual_32x32_neon|
    ARM

    AREA ||.text||, CODE, READONLY, ALIGN=2

    MACRO
    LD_16x8 $src, $stride
    vld1.8              {q8},       [$src],     $stride
    vld1.8              {q9},       [$src],     $stride
    vld1.8              {q10},      [$src],     $stride
    vld1.8              {q11},      [$src],     $stride
    vld1.8              {q12},      [$src],     $stride
    vld1.8              {q13},      [$src],     $stride
    vld1.8              {q14},      [$src],     $stride
    vld1.8              {q15},      [$src],     $stride
    MEND

    MACRO
    ADD_DIFF_16x8 $diff
    vqadd.u8            q8,         q8,         $diff
    vqadd.u8            q9,         q9,         $diff
    vqadd.u8            q10,        q10,        $diff
    vqadd.u8            q11,        q11,        $diff
    vqadd.u8            q12,        q12,        $diff
    vqadd.u8            q13,        q13,        $diff
    vqadd.u8            q14,        q14,        $diff
    vqadd.u8            q15,        q15,        $diff
    MEND

    MACRO
    SUB_DIFF_16x8 $diff
    vqsub.u8            q8,         q8,         $diff
    vqsub.u8            q9,         q9,         $diff
    vqsub.u8            q10,        q10,        $diff
    vqsub.u8            q11,        q11,        $diff
    vqsub.u8            q12,        q12,        $diff
    vqsub.u8            q13,        q13,        $diff
    vqsub.u8            q14,        q14,        $diff
    vqsub.u8            q15,        q15,        $diff
    MEND

    MACRO
    ST_16x8 $dst, $stride
    vst1.8              {q8},       [$dst],     $stride
    vst1.8              {q9},       [$dst],     $stride
    vst1.8              {q10},      [$dst],     $stride
    vst1.8              {q11},      [$dst],     $stride
    vst1.8              {q12},      [$dst],     $stride
    vst1.8              {q13},      [$dst],     $stride
    vst1.8              {q14},      [$dst],     $stride
    vst1.8              {q15},      [$dst],     $stride
    MEND

; void add_constant_residual(const int16_t diff, uint8_t *dest, int stride,
;                             int width, int height) {
;  int r, c;
;
;  for (r = 0; r < height; r++) {
;    for (c = 0; c < width; c++)
;      dest[c] = clip_pixel(diff + dest[c]);
;
;    dest += stride;
;  }
;}
;void vp9_add_constant_residual_8x8_c(const int16_t diff, uint8_t *dest,
;                                     int stride) {
;  add_constant_residual(diff, dest, stride, 8, 8);
;}
;       r0      : const int16_t diff
;       r1      : const uint8_t *dest
;       r2      : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_8x8_neon| PROC
    mov                 r3,         r1                      ; r3: save dest to r3
    vld1.8              {d0},       [r1],       r2
    vld1.8              {d1},       [r1],       r2
    vld1.8              {d2},       [r1],       r2
    vld1.8              {d3},       [r1],       r2
    vld1.8              {d4},       [r1],       r2
    vld1.8              {d5},       [r1],       r2
    vld1.8              {d6},       [r1],       r2
    vld1.8              {d7},       [r1],       r2
    cmp                 r0,         #0
    bge                 DIFF_POSITIVE_8x8

DIFF_NEGATIVE_8x8                                           ; diff < 0
    neg                 r0,         r0
    usat                r0,         #8,         r0
    vdup.u8             q8,         r0

    vqsub.u8            q0,         q0,         q8
    vqsub.u8            q1,         q1,         q8
    vqsub.u8            q2,         q2,         q8
    vqsub.u8            q3,         q3,         q8
    b                   DIFF_SAVE_8x8

DIFF_POSITIVE_8x8                                           ; diff >= 0
    usat                r0,         #8,         r0
    vdup.u8             q8,         r0

    vqadd.u8            q0,         q0,         q8
    vqadd.u8            q1,         q1,         q8
    vqadd.u8            q2,         q2,         q8
    vqadd.u8            q3,         q3,         q8

DIFF_SAVE_8x8
    vst1.8              {d0},       [r3],       r2
    vst1.8              {d1},       [r3],       r2
    vst1.8              {d2},       [r3],       r2
    vst1.8              {d3},       [r3],       r2
    vst1.8              {d4},       [r3],       r2
    vst1.8              {d5},       [r3],       r2
    vst1.8              {d6},       [r3],       r2
    vst1.8              {d7},       [r3],       r2

    bx                  lr
    ENDP

;void vp9_add_constant_residual_16x16_c(const int16_t diff, uint8_t *dest,
;                                       int stride) {
;  add_constant_residual(diff, dest, stride, 16, 16);
;}
;       r0      : const int16_t diff
;       r1      : const uint8_t *dest
;       r2      : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_16x16_neon| PROC
    mov                 r3,         r1
    LD_16x8             r1,         r2
    cmp                 r0,         #0
    bge                 DIFF_POSITIVE_16x16

|DIFF_NEGATIVE_16x16|
    neg                 r0,         r0
    usat                r0,         #8,         r0
    vdup.u8             q0,         r0

    SUB_DIFF_16x8       q0
    ST_16x8             r3,         r2
    LD_16x8             r1,         r2
    SUB_DIFF_16x8       q0
    b                   DIFF_SAVE_16x16

|DIFF_POSITIVE_16x16|
    usat                r0,         #8,         r0
    vdup.u8             q0,         r0

    ADD_DIFF_16x8       q0
    ST_16x8             r3,         r2
    LD_16x8             r1,         r2
    ADD_DIFF_16x8       q0

|DIFF_SAVE_16x16|
    ST_16x8             r3,         r2
    bx                  lr
    ENDP

;void vp9_add_constant_residual_32x32_c(const int16_t diff, uint8_t *dest,
;                                       int stride) {
;  add_constant_residual(diff, dest, stride, 32, 32);
;}
;       r0      : const int16_t diff
;       r1      : const uint8_t *dest
;       r2      : int stride
;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;
|vp9_add_constant_residual_32x32_neon| PROC
    push                {r4,lr}
    pld                 [r1]
    mov                 r3,         r1
    add                 r4,         r1,         #16         ; r4 dest + 16 for second loop
    cmp                 r0,         #0
    bge                 DIFF_POSITIVE_32x32

|DIFF_NEGATIVE_32x32|
    neg                 r0,         r0
    usat                r0,         #8,         r0
    vdup.u8             q0,         r0
    mov                 r0,         #4

|DIFF_NEGATIVE_32x32_LOOP|
    sub                 r0,         #1
    LD_16x8             r1,         r2
    SUB_DIFF_16x8       q0
    ST_16x8             r3,         r2

    LD_16x8             r1,         r2
    SUB_DIFF_16x8       q0
    ST_16x8             r3,         r2
    cmp                 r0,         #2
    moveq               r1,         r4
    moveq               r3,         r4
    cmp                 r0,         #0
    bne                 DIFF_NEGATIVE_32x32_LOOP
    pop                 {r4,pc}

|DIFF_POSITIVE_32x32|
    usat                r0,         #8,         r0
    vdup.u8             q0,         r0
    mov                 r0,         #4

|DIFF_POSITIVE_32x32_LOOP|
    sub                 r0,         #1
    LD_16x8             r1,         r2
    ADD_DIFF_16x8       q0
    ST_16x8             r3,         r2

    LD_16x8             r1,         r2
    ADD_DIFF_16x8       q0
    ST_16x8             r3,         r2
    cmp                 r0,         #2
    moveq               r1,         r4
    moveq               r3,         r4
    cmp                 r0,         #0
    bne                 DIFF_POSITIVE_32x32_LOOP
    pop                 {r4,pc}
    ENDP

    END