| Index: webrtc/modules/video_processing/util/denoiser_filter_neon.cc | 
| diff --git a/webrtc/modules/video_processing/util/denoiser_filter_neon.cc b/webrtc/modules/video_processing/util/denoiser_filter_neon.cc | 
| new file mode 100644 | 
| index 0000000000000000000000000000000000000000..67d420cda723a27bb8bcdc72dbae3ad701bc8a03 | 
| --- /dev/null | 
| +++ b/webrtc/modules/video_processing/util/denoiser_filter_neon.cc | 
| @@ -0,0 +1,284 @@ | 
| +/* | 
| + *  Copyright (c) 2015 The WebRTC project authors. All Rights Reserved. | 
| + * | 
| + *  Use of this source code is governed by a BSD-style license | 
| + *  that can be found in the LICENSE file in the root of the source | 
| + *  tree. An additional intellectual property rights grant can be found | 
| + *  in the file PATENTS.  All contributing project authors may | 
| + *  be found in the AUTHORS file in the root of the source tree. | 
| + */ | 
| + | 
| +#include <arm_neon.h> | 
| + | 
| +#include "webrtc/modules/video_processing/util/denoiser_filter_neon.h" | 
| + | 
| +namespace webrtc { | 
| + | 
| +static int HorizontalAddS16x8(const int16x8_t v_16x8) { | 
| +  const int32x4_t a = vpaddlq_s16(v_16x8); | 
| +  const int64x2_t b = vpaddlq_s32(a); | 
| +  const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), | 
| +                               vreinterpret_s32_s64(vget_high_s64(b))); | 
| +  return vget_lane_s32(c, 0); | 
| +} | 
| + | 
| +static int HorizontalAddS32x4(const int32x4_t v_32x4) { | 
| +  const int64x2_t b = vpaddlq_s32(v_32x4); | 
| +  const int32x2_t c = vadd_s32(vreinterpret_s32_s64(vget_low_s64(b)), | 
| +                               vreinterpret_s32_s64(vget_high_s64(b))); | 
| +  return vget_lane_s32(c, 0); | 
| +} | 
| + | 
| +static void VarianceNeonW8(const uint8_t* a, | 
| +                           int a_stride, | 
| +                           const uint8_t* b, | 
| +                           int b_stride, | 
| +                           int w, | 
| +                           int h, | 
| +                           uint32_t* sse, | 
| +                           int64_t* sum) { | 
| +  int16x8_t v_sum = vdupq_n_s16(0); | 
| +  int32x4_t v_sse_lo = vdupq_n_s32(0); | 
| +  int32x4_t v_sse_hi = vdupq_n_s32(0); | 
| + | 
| +  for (int i = 0; i < h; ++i) { | 
| +    for (int j = 0; j < w; j += 8) { | 
| +      const uint8x8_t v_a = vld1_u8(&a[j]); | 
| +      const uint8x8_t v_b = vld1_u8(&b[j]); | 
| +      const uint16x8_t v_diff = vsubl_u8(v_a, v_b); | 
| +      const int16x8_t sv_diff = vreinterpretq_s16_u16(v_diff); | 
| +      v_sum = vaddq_s16(v_sum, sv_diff); | 
| +      v_sse_lo = | 
| +          vmlal_s16(v_sse_lo, vget_low_s16(sv_diff), vget_low_s16(sv_diff)); | 
| +      v_sse_hi = | 
| +          vmlal_s16(v_sse_hi, vget_high_s16(sv_diff), vget_high_s16(sv_diff)); | 
| +    } | 
| +    a += a_stride; | 
| +    b += b_stride; | 
| +  } | 
| + | 
| +  *sum = HorizontalAddS16x8(v_sum); | 
| +  *sse = static_cast<uint32_t>( | 
| +      HorizontalAddS32x4(vaddq_s32(v_sse_lo, v_sse_hi))); | 
| +} | 
| + | 
| +void DenoiserFilterNEON::CopyMem16x16(const uint8_t* src, | 
| +                                      int src_stride, | 
| +                                      uint8_t* dst, | 
| +                                      int dst_stride) { | 
| +  uint8x16_t qtmp; | 
| +  for (int r = 0; r < 16; r++) { | 
| +    qtmp = vld1q_u8(src); | 
| +    vst1q_u8(dst, qtmp); | 
| +    src += src_stride; | 
| +    dst += dst_stride; | 
| +  } | 
| +} | 
| + | 
| +void DenoiserFilterNEON::CopyMem8x8(const uint8_t* src, | 
| +                                    int src_stride, | 
| +                                    uint8_t* dst, | 
| +                                    int dst_stride) { | 
| +  uint8x8_t vtmp; | 
| + | 
| +  for (int r = 0; r < 8; r++) { | 
| +    vtmp = vld1_u8(src); | 
| +    vst1_u8(dst, vtmp); | 
| +    src += src_stride; | 
| +    dst += dst_stride; | 
| +  } | 
| +} | 
| + | 
| +uint32_t DenoiserFilterNEON::Variance16x8(const uint8_t* a, | 
| +                                          int a_stride, | 
| +                                          const uint8_t* b, | 
| +                                          int b_stride, | 
| +                                          uint32_t* sse) { | 
| +  int64_t sum = 0; | 
| +  VarianceNeonW8(a, a_stride << 1, b, b_stride << 1, 16, 8, sse, &sum); | 
| +  return *sse - ((sum * sum) >> 7); | 
| +} | 
| + | 
| +DenoiserDecision DenoiserFilterNEON::MbDenoise(uint8_t* mc_running_avg_y, | 
| +                                               int mc_running_avg_y_stride, | 
| +                                               uint8_t* running_avg_y, | 
| +                                               int running_avg_y_stride, | 
| +                                               const uint8_t* sig, | 
| +                                               int sig_stride, | 
| +                                               uint8_t motion_magnitude, | 
| +                                               int increase_denoising) { | 
| +  // If motion_magnitude is small, making the denoiser more aggressive by | 
| +  // increasing the adjustment for each level, level1 adjustment is | 
| +  // increased, the deltas stay the same. | 
| +  int shift_inc = | 
| +      (increase_denoising && motion_magnitude <= kMotionMagnitudeThreshold) | 
| +          ? 1 | 
| +          : 0; | 
| +  const uint8x16_t v_level1_adjustment = vmovq_n_u8( | 
| +      (motion_magnitude <= kMotionMagnitudeThreshold) ? 4 + shift_inc : 3); | 
| +  const uint8x16_t v_delta_level_1_and_2 = vdupq_n_u8(1); | 
| +  const uint8x16_t v_delta_level_2_and_3 = vdupq_n_u8(2); | 
| +  const uint8x16_t v_level1_threshold = vmovq_n_u8(4 + shift_inc); | 
| +  const uint8x16_t v_level2_threshold = vdupq_n_u8(8); | 
| +  const uint8x16_t v_level3_threshold = vdupq_n_u8(16); | 
| +  int64x2_t v_sum_diff_total = vdupq_n_s64(0); | 
| + | 
| +  // Go over lines. | 
| +  for (int r = 0; r < 16; ++r) { | 
| +    // Load inputs. | 
| +    const uint8x16_t v_sig = vld1q_u8(sig); | 
| +    const uint8x16_t v_mc_running_avg_y = vld1q_u8(mc_running_avg_y); | 
| + | 
| +    // Calculate absolute difference and sign masks. | 
| +    const uint8x16_t v_abs_diff = vabdq_u8(v_sig, v_mc_running_avg_y); | 
| +    const uint8x16_t v_diff_pos_mask = vcltq_u8(v_sig, v_mc_running_avg_y); | 
| +    const uint8x16_t v_diff_neg_mask = vcgtq_u8(v_sig, v_mc_running_avg_y); | 
| + | 
| +    // Figure out which level that put us in. | 
| +    const uint8x16_t v_level1_mask = vcleq_u8(v_level1_threshold, v_abs_diff); | 
| +    const uint8x16_t v_level2_mask = vcleq_u8(v_level2_threshold, v_abs_diff); | 
| +    const uint8x16_t v_level3_mask = vcleq_u8(v_level3_threshold, v_abs_diff); | 
| + | 
| +    // Calculate absolute adjustments for level 1, 2 and 3. | 
| +    const uint8x16_t v_level2_adjustment = | 
| +        vandq_u8(v_level2_mask, v_delta_level_1_and_2); | 
| +    const uint8x16_t v_level3_adjustment = | 
| +        vandq_u8(v_level3_mask, v_delta_level_2_and_3); | 
| +    const uint8x16_t v_level1and2_adjustment = | 
| +        vaddq_u8(v_level1_adjustment, v_level2_adjustment); | 
| +    const uint8x16_t v_level1and2and3_adjustment = | 
| +        vaddq_u8(v_level1and2_adjustment, v_level3_adjustment); | 
| + | 
| +    // Figure adjustment absolute value by selecting between the absolute | 
| +    // difference if in level0 or the value for level 1, 2 and 3. | 
| +    const uint8x16_t v_abs_adjustment = | 
| +        vbslq_u8(v_level1_mask, v_level1and2and3_adjustment, v_abs_diff); | 
| + | 
| +    // Calculate positive and negative adjustments. Apply them to the signal | 
| +    // and accumulate them. Adjustments are less than eight and the maximum | 
| +    // sum of them (7 * 16) can fit in a signed char. | 
| +    const uint8x16_t v_pos_adjustment = | 
| +        vandq_u8(v_diff_pos_mask, v_abs_adjustment); | 
| +    const uint8x16_t v_neg_adjustment = | 
| +        vandq_u8(v_diff_neg_mask, v_abs_adjustment); | 
| + | 
| +    uint8x16_t v_running_avg_y = vqaddq_u8(v_sig, v_pos_adjustment); | 
| +    v_running_avg_y = vqsubq_u8(v_running_avg_y, v_neg_adjustment); | 
| + | 
| +    // Store results. | 
| +    vst1q_u8(running_avg_y, v_running_avg_y); | 
| + | 
| +    // Sum all the accumulators to have the sum of all pixel differences | 
| +    // for this macroblock. | 
| +    { | 
| +      const int8x16_t v_sum_diff = | 
| +          vqsubq_s8(vreinterpretq_s8_u8(v_pos_adjustment), | 
| +                    vreinterpretq_s8_u8(v_neg_adjustment)); | 
| +      const int16x8_t fe_dc_ba_98_76_54_32_10 = vpaddlq_s8(v_sum_diff); | 
| +      const int32x4_t fedc_ba98_7654_3210 = | 
| +          vpaddlq_s16(fe_dc_ba_98_76_54_32_10); | 
| +      const int64x2_t fedcba98_76543210 = vpaddlq_s32(fedc_ba98_7654_3210); | 
| + | 
| +      v_sum_diff_total = vqaddq_s64(v_sum_diff_total, fedcba98_76543210); | 
| +    } | 
| + | 
| +    // Update pointers for next iteration. | 
| +    sig += sig_stride; | 
| +    mc_running_avg_y += mc_running_avg_y_stride; | 
| +    running_avg_y += running_avg_y_stride; | 
| +  } | 
| + | 
| +  // Too much adjustments => copy block. | 
| +  { | 
| +    int64x1_t x = vqadd_s64(vget_high_s64(v_sum_diff_total), | 
| +                            vget_low_s64(v_sum_diff_total)); | 
| +    int sum_diff = vget_lane_s32(vabs_s32(vreinterpret_s32_s64(x)), 0); | 
| +    int sum_diff_thresh = kSumDiffThreshold; | 
| + | 
| +    if (increase_denoising) | 
| +      sum_diff_thresh = kSumDiffThresholdHigh; | 
| +    if (sum_diff > sum_diff_thresh) { | 
| +      // Before returning to copy the block (i.e., apply no denoising), | 
| +      // checK if we can still apply some (weaker) temporal filtering to | 
| +      // this block, that would otherwise not be denoised at all. Simplest | 
| +      // is to apply an additional adjustment to running_avg_y to bring it | 
| +      // closer to sig. The adjustment is capped by a maximum delta, and | 
| +      // chosen such that in most cases the resulting sum_diff will be | 
| +      // within the accceptable range given by sum_diff_thresh. | 
| + | 
| +      // The delta is set by the excess of absolute pixel diff over the | 
| +      // threshold. | 
| +      int delta = ((sum_diff - sum_diff_thresh) >> 8) + 1; | 
| +      // Only apply the adjustment for max delta up to 3. | 
| +      if (delta < 4) { | 
| +        const uint8x16_t k_delta = vmovq_n_u8(delta); | 
| +        sig -= sig_stride * 16; | 
| +        mc_running_avg_y -= mc_running_avg_y_stride * 16; | 
| +        running_avg_y -= running_avg_y_stride * 16; | 
| +        for (int r = 0; r < 16; ++r) { | 
| +          uint8x16_t v_running_avg_y = vld1q_u8(running_avg_y); | 
| +          const uint8x16_t v_sig = vld1q_u8(sig); | 
| +          const uint8x16_t v_mc_running_avg_y = vld1q_u8(mc_running_avg_y); | 
| + | 
| +          // Calculate absolute difference and sign masks. | 
| +          const uint8x16_t v_abs_diff = vabdq_u8(v_sig, v_mc_running_avg_y); | 
| +          const uint8x16_t v_diff_pos_mask = | 
| +              vcltq_u8(v_sig, v_mc_running_avg_y); | 
| +          const uint8x16_t v_diff_neg_mask = | 
| +              vcgtq_u8(v_sig, v_mc_running_avg_y); | 
| +          // Clamp absolute difference to delta to get the adjustment. | 
| +          const uint8x16_t v_abs_adjustment = vminq_u8(v_abs_diff, (k_delta)); | 
| + | 
| +          const uint8x16_t v_pos_adjustment = | 
| +              vandq_u8(v_diff_pos_mask, v_abs_adjustment); | 
| +          const uint8x16_t v_neg_adjustment = | 
| +              vandq_u8(v_diff_neg_mask, v_abs_adjustment); | 
| + | 
| +          v_running_avg_y = vqsubq_u8(v_running_avg_y, v_pos_adjustment); | 
| +          v_running_avg_y = vqaddq_u8(v_running_avg_y, v_neg_adjustment); | 
| + | 
| +          // Store results. | 
| +          vst1q_u8(running_avg_y, v_running_avg_y); | 
| + | 
| +          { | 
| +            const int8x16_t v_sum_diff = | 
| +                vqsubq_s8(vreinterpretq_s8_u8(v_neg_adjustment), | 
| +                          vreinterpretq_s8_u8(v_pos_adjustment)); | 
| + | 
| +            const int16x8_t fe_dc_ba_98_76_54_32_10 = vpaddlq_s8(v_sum_diff); | 
| +            const int32x4_t fedc_ba98_7654_3210 = | 
| +                vpaddlq_s16(fe_dc_ba_98_76_54_32_10); | 
| +            const int64x2_t fedcba98_76543210 = | 
| +                vpaddlq_s32(fedc_ba98_7654_3210); | 
| + | 
| +            v_sum_diff_total = vqaddq_s64(v_sum_diff_total, fedcba98_76543210); | 
| +          } | 
| +          // Update pointers for next iteration. | 
| +          sig += sig_stride; | 
| +          mc_running_avg_y += mc_running_avg_y_stride; | 
| +          running_avg_y += running_avg_y_stride; | 
| +        } | 
| +        { | 
| +          // Update the sum of all pixel differences of this MB. | 
| +          x = vqadd_s64(vget_high_s64(v_sum_diff_total), | 
| +                        vget_low_s64(v_sum_diff_total)); | 
| +          sum_diff = vget_lane_s32(vabs_s32(vreinterpret_s32_s64(x)), 0); | 
| + | 
| +          if (sum_diff > sum_diff_thresh) { | 
| +            return COPY_BLOCK; | 
| +          } | 
| +        } | 
| +      } else { | 
| +        return COPY_BLOCK; | 
| +      } | 
| +    } | 
| +  } | 
| + | 
| +  // Tell above level that block was filtered. | 
| +  running_avg_y -= running_avg_y_stride * 16; | 
| +  sig -= sig_stride * 16; | 
| + | 
| +  return FILTER_BLOCK; | 
| +} | 
| + | 
| +}  // namespace webrtc | 
|  |