ref: 6965bde4e0b27e6db83f243361d11791fb520b02
dir: /vpx_dsp/sad.c/
/* * Copyright (c) 2015 The WebM project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include <stdlib.h> #include "./vpx_config.h" #include "./vpx_dsp_rtcd.h" #include "vpx/vpx_integer.h" #include "vpx_ports/mem.h" /* Sum the difference between every corresponding element of the buffers. */ static INLINE unsigned int sad(const uint8_t *a, int a_stride, const uint8_t *b, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } // TODO(johannkoenig): this moved to vpx_dsp, should be able to clean this up. /* Remove dependency on vp9 variance function by duplicating vp9_comp_avg_pred. * The function averages every corresponding element of the buffers and stores * the value in a third buffer, comp_pred. * pred and comp_pred are assumed to have stride = width * In the usage below comp_pred is a local array. */ static INLINE void avg_pred(uint8_t *comp_pred, const uint8_t *pred, int width, int height, const uint8_t *ref, int ref_stride) { int i, j; for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { const int tmp = pred[j] + ref[j]; comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1); } comp_pred += width; pred += width; ref += ref_stride; } } #if CONFIG_VP9_HIGHBITDEPTH static INLINE void highbd_avg_pred(uint16_t *comp_pred, const uint8_t *pred8, int width, int height, const uint8_t *ref8, int ref_stride) { int i, j; uint16_t *pred = CONVERT_TO_SHORTPTR(pred8); uint16_t *ref = CONVERT_TO_SHORTPTR(ref8); for (i = 0; i < height; i++) { for (j = 0; j < width; j++) { const int tmp = pred[j] + ref[j]; comp_pred[j] = ROUND_POWER_OF_TWO(tmp, 1); } comp_pred += width; pred += width; ref += ref_stride; } } #endif // CONFIG_VP9_HIGHBITDEPTH #define sadMxN(m, n) \ unsigned int vpx_sad##m##x##n##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, int ref_stride) { \ return sad(src, src_stride, ref, ref_stride, m, n); \ } \ unsigned int vpx_sad##m##x##n##_avg_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, int ref_stride, \ const uint8_t *second_pred) { \ uint8_t comp_pred[m * n]; \ avg_pred(comp_pred, second_pred, m, n, ref, ref_stride); \ return sad(src, src_stride, comp_pred, m, m, n); \ } // depending on call sites, pass **ref_array to avoid & in subsequent call and // de-dup with 4D below. #define sadMxNxK(m, n, k) \ void vpx_sad##m##x##n##x##k##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref_array, int ref_stride, \ uint32_t *sad_array) { \ int i; \ for (i = 0; i < k; ++i) \ sad_array[i] = vpx_sad##m##x##n##_c(src, src_stride, &ref_array[i], ref_stride); \ } // This appears to be equivalent to the above when k == 4 and refs is const #define sadMxNx4D(m, n) \ void vpx_sad##m##x##n##x4d_c(const uint8_t *src, int src_stride, \ const uint8_t *const ref_array[], int ref_stride, \ uint32_t *sad_array) { \ int i; \ for (i = 0; i < 4; ++i) \ sad_array[i] = vpx_sad##m##x##n##_c(src, src_stride, ref_array[i], ref_stride); \ } // 64x64 sadMxN(64, 64) sadMxNxK(64, 64, 3) sadMxNxK(64, 64, 8) sadMxNx4D(64, 64) // 64x32 sadMxN(64, 32) sadMxNx4D(64, 32) // 32x64 sadMxN(32, 64) sadMxNx4D(32, 64) // 32x32 sadMxN(32, 32) sadMxNxK(32, 32, 3) sadMxNxK(32, 32, 8) sadMxNx4D(32, 32) // 32x16 sadMxN(32, 16) sadMxNx4D(32, 16) // 16x32 sadMxN(16, 32) sadMxNx4D(16, 32) // 16x16 sadMxN(16, 16) sadMxNxK(16, 16, 3) sadMxNxK(16, 16, 8) sadMxNx4D(16, 16) // 16x8 sadMxN(16, 8) sadMxNxK(16, 8, 3) sadMxNxK(16, 8, 8) sadMxNx4D(16, 8) // 8x16 sadMxN(8, 16) sadMxNxK(8, 16, 3) sadMxNxK(8, 16, 8) sadMxNx4D(8, 16) // 8x8 sadMxN(8, 8) sadMxNxK(8, 8, 3) sadMxNxK(8, 8, 8) sadMxNx4D(8, 8) // 8x4 sadMxN(8, 4) sadMxNxK(8, 4, 8) sadMxNx4D(8, 4) // 4x8 sadMxN(4, 8) sadMxNxK(4, 8, 8) sadMxNx4D(4, 8) // 4x4 sadMxN(4, 4) sadMxNxK(4, 4, 3) sadMxNxK(4, 4, 8) sadMxNx4D(4, 4) #if CONFIG_VP9_HIGHBITDEPTH static INLINE unsigned int highbd_sad(const uint8_t *a8, int a_stride, const uint8_t *b8, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *a = CONVERT_TO_SHORTPTR(a8); const uint16_t *b = CONVERT_TO_SHORTPTR(b8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } static INLINE unsigned int highbd_sadb(const uint8_t *a8, int a_stride, const uint16_t *b, int b_stride, int width, int height) { int y, x; unsigned int sad = 0; const uint16_t *a = CONVERT_TO_SHORTPTR(a8); for (y = 0; y < height; y++) { for (x = 0; x < width; x++) sad += abs(a[x] - b[x]); a += a_stride; b += b_stride; } return sad; } #define highbd_sadMxN(m, n) \ unsigned int vpx_highbd_sad##m##x##n##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref, int ref_stride) { \ return highbd_sad(src, src_stride, ref, ref_stride, m, n); \ } \ unsigned int vpx_highbd_sad##m##x##n##_avg_c(const uint8_t *src, \ int src_stride, \ const uint8_t *ref, \ int ref_stride, \ const uint8_t *second_pred) { \ uint16_t comp_pred[m * n]; \ highbd_avg_pred(comp_pred, second_pred, m, n, ref, ref_stride); \ return highbd_sadb(src, src_stride, comp_pred, m, m, n); \ } #define highbd_sadMxNxK(m, n, k) \ void vpx_highbd_sad##m##x##n##x##k##_c(const uint8_t *src, int src_stride, \ const uint8_t *ref_array, int ref_stride, \ uint32_t *sad_array) { \ int i; \ for (i = 0; i < k; ++i) { \ sad_array[i] = vpx_highbd_sad##m##x##n##_c(src, src_stride, &ref_array[i], \ ref_stride); \ } \ } #define highbd_sadMxNx4D(m, n) \ void vpx_highbd_sad##m##x##n##x4d_c(const uint8_t *src, int src_stride, \ const uint8_t *const ref_array[], \ int ref_stride, uint32_t *sad_array) { \ int i; \ for (i = 0; i < 4; ++i) { \ sad_array[i] = vpx_highbd_sad##m##x##n##_c(src, src_stride, ref_array[i], \ ref_stride); \ } \ } // 64x64 highbd_sadMxN(64, 64) highbd_sadMxNxK(64, 64, 3) highbd_sadMxNxK(64, 64, 8) highbd_sadMxNx4D(64, 64) // 64x32 highbd_sadMxN(64, 32) highbd_sadMxNx4D(64, 32) // 32x64 highbd_sadMxN(32, 64) highbd_sadMxNx4D(32, 64) // 32x32 highbd_sadMxN(32, 32) highbd_sadMxNxK(32, 32, 3) highbd_sadMxNxK(32, 32, 8) highbd_sadMxNx4D(32, 32) // 32x16 highbd_sadMxN(32, 16) highbd_sadMxNx4D(32, 16) // 16x32 highbd_sadMxN(16, 32) highbd_sadMxNx4D(16, 32) // 16x16 highbd_sadMxN(16, 16) highbd_sadMxNxK(16, 16, 3) highbd_sadMxNxK(16, 16, 8) highbd_sadMxNx4D(16, 16) // 16x8 highbd_sadMxN(16, 8) highbd_sadMxNxK(16, 8, 3) highbd_sadMxNxK(16, 8, 8) highbd_sadMxNx4D(16, 8) // 8x16 highbd_sadMxN(8, 16) highbd_sadMxNxK(8, 16, 3) highbd_sadMxNxK(8, 16, 8) highbd_sadMxNx4D(8, 16) // 8x8 highbd_sadMxN(8, 8) highbd_sadMxNxK(8, 8, 3) highbd_sadMxNxK(8, 8, 8) highbd_sadMxNx4D(8, 8) // 8x4 highbd_sadMxN(8, 4) highbd_sadMxNxK(8, 4, 8) highbd_sadMxNx4D(8, 4) // 4x8 highbd_sadMxN(4, 8) highbd_sadMxNxK(4, 8, 8) highbd_sadMxNx4D(4, 8) // 4x4 highbd_sadMxN(4, 4) highbd_sadMxNxK(4, 4, 3) highbd_sadMxNxK(4, 4, 8) highbd_sadMxNx4D(4, 4) #endif // CONFIG_VP9_HIGHBITDEPTH