DXR is a code search and navigation tool aimed at making sense of large projects. It supports full-text and regex searches as well as structural queries.

Mercurial (d8847129d134)

VCS Links

Line Code
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96

/*
 * Copyright 2012 The Android Open Source Project
 *
 * Use of this source code is governed by a BSD-style license that can be
 * found in the LICENSE file.
 */


#include <arm_neon.h>
#include "SkColorPriv.h"

/*
 * Filter_32_opaque
 *
 * There is no hard-n-fast rule that the filtering must produce
 * exact results for the color components, but if the 4 incoming colors are
 * all opaque, then the output color must also be opaque. Subsequent parts of
 * the drawing pipeline may rely on this (e.g. which blitrow proc to use).
 *
 */
// Chrome on Android uses -Os so we need to force these inline. Otherwise
// calling the function in the inner loops will cause significant overhead on
// some platforms.
static SK_ALWAYS_INLINE void Filter_32_opaque_neon(unsigned x, unsigned y,
                                                   SkPMColor a00, SkPMColor a01,
                                                   SkPMColor a10, SkPMColor a11,
                                                   SkPMColor *dst) {
    uint8x8_t vy, vconst16_8, v16_y, vres;
    uint16x4_t vx, vconst16_16, v16_x, tmp;
    uint32x2_t va0, va1;
    uint16x8_t tmp1, tmp2;

    vy = vdup_n_u8(y);                // duplicate y into vy
    vconst16_8 = vmov_n_u8(16);       // set up constant in vconst16_8
    v16_y = vsub_u8(vconst16_8, vy);  // v16_y = 16-y

    va0 = vdup_n_u32(a00);            // duplicate a00
    va1 = vdup_n_u32(a10);            // duplicate a10
    va0 = vset_lane_u32(a01, va0, 1); // set top to a01
    va1 = vset_lane_u32(a11, va1, 1); // set top to a11

    tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y)
    tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy);    // tmp2 = [a11|a10] * y

    vx = vdup_n_u16(x);                // duplicate x into vx
    vconst16_16 = vmov_n_u16(16);      // set up constant in vconst16_16
    v16_x = vsub_u16(vconst16_16, vx); // v16_x = 16-x

    tmp = vmul_u16(vget_high_u16(tmp1), vx);        // tmp  = a01 * x
    tmp = vmla_u16(tmp, vget_high_u16(tmp2), vx);   // tmp += a11 * x
    tmp = vmla_u16(tmp, vget_low_u16(tmp1), v16_x); // tmp += a00 * (16-x)
    tmp = vmla_u16(tmp, vget_low_u16(tmp2), v16_x); // tmp += a10 * (16-x)

    vres = vshrn_n_u16(vcombine_u16(tmp, vcreate_u16(0)), 8); // shift down result by 8
    vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0);         // store result
}

static SK_ALWAYS_INLINE void Filter_32_alpha_neon(unsigned x, unsigned y,
                                                  SkPMColor a00, SkPMColor a01,
                                                  SkPMColor a10, SkPMColor a11,
                                                  SkPMColor *dst,
                                                  uint16_t scale) {
    uint8x8_t vy, vconst16_8, v16_y, vres;
    uint16x4_t vx, vconst16_16, v16_x, tmp, vscale;
    uint32x2_t va0, va1;
    uint16x8_t tmp1, tmp2;

    vy = vdup_n_u8(y);                // duplicate y into vy
    vconst16_8 = vmov_n_u8(16);       // set up constant in vconst16_8
    v16_y = vsub_u8(vconst16_8, vy);  // v16_y = 16-y

    va0 = vdup_n_u32(a00);            // duplicate a00
    va1 = vdup_n_u32(a10);            // duplicate a10
    va0 = vset_lane_u32(a01, va0, 1); // set top to a01
    va1 = vset_lane_u32(a11, va1, 1); // set top to a11

    tmp1 = vmull_u8(vreinterpret_u8_u32(va0), v16_y); // tmp1 = [a01|a00] * (16-y)
    tmp2 = vmull_u8(vreinterpret_u8_u32(va1), vy);    // tmp2 = [a11|a10] * y

    vx = vdup_n_u16(x);                // duplicate x into vx
    vconst16_16 = vmov_n_u16(16);      // set up constant in vconst16_16
    v16_x = vsub_u16(vconst16_16, vx); // v16_x = 16-x

    tmp = vmul_u16(vget_high_u16(tmp1), vx);        // tmp  = a01 * x
    tmp = vmla_u16(tmp, vget_high_u16(tmp2), vx);   // tmp += a11 * x
    tmp = vmla_u16(tmp, vget_low_u16(tmp1), v16_x); // tmp += a00 * (16-x)
    tmp = vmla_u16(tmp, vget_low_u16(tmp2), v16_x); // tmp += a10 * (16-x)

    vscale = vdup_n_u16(scale);        // duplicate scale
    tmp = vshr_n_u16(tmp, 8);          // shift down result by 8
    tmp = vmul_u16(tmp, vscale);       // multiply result by scale

    vres = vshrn_n_u16(vcombine_u16(tmp, vcreate_u16(0)), 8); // shift down result by 8
    vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0);         // store result
}