ref: 3dd20456ab0e5a9daeea0cdcc82e1342c586b607
parent: 3486abd54abad4ba2f67f83a0b5b11ee833b3412
parent: 201dcefafe92d5ffc274a462708b8c11dd2d4212
author: Linfeng Zhang <[email protected]>
date: Wed Dec 14 12:05:34 EST 2016
Merge "Update idct test code to test 8-bit & high bitdepth simultaneously"
--- a/test/partial_idct_test.cc
+++ b/test/partial_idct_test.cc
@@ -32,18 +32,25 @@
namespace {
typedef void (*FwdTxfmFunc)(const int16_t *in, tran_low_t *out, int stride);
+typedef void (*InvTxfmFunc)(const tran_low_t *in, uint8_t *out, int stride);
+typedef void (*InvTxfmWithBdFunc)(const tran_low_t *in, uint8_t *out,
+ int stride, int bd);
+template <InvTxfmFunc fn>
+void wrapper(const tran_low_t *in, uint8_t *out, int stride, int bd) {
+ (void)bd;
+ fn(in, out, stride);
+}
+
#if CONFIG_VP9_HIGHBITDEPTH
-typedef uint16_t Pixel;
-typedef void (*InvTxfmFunc)(const tran_low_t *in, uint8_t *out, int stride,
- int bd);
-#else // !CONFIG_VP9_HIGHBITDEPTH
-typedef uint8_t Pixel;
-typedef void (*InvTxfmFunc)(const tran_low_t *in, uint8_t *out, int stride);
-#endif // CONFIG_VP9_HIGHBITDEPTH
+template <InvTxfmWithBdFunc fn>
+void highbd_wrapper(const tran_low_t *in, uint8_t *out, int stride, int bd) {
+ fn(in, CONVERT_TO_BYTEPTR(out), stride, bd);
+}
+#endif
-typedef std::tr1::tuple<FwdTxfmFunc, InvTxfmFunc, InvTxfmFunc, TX_SIZE, int,
- int>
+typedef std::tr1::tuple<FwdTxfmFunc, InvTxfmWithBdFunc, InvTxfmWithBdFunc,
+ TX_SIZE, int, int, int>
PartialInvTxfmParam;
const int kMaxNumCoeffs = 1024;
const int kCountTestBlock = 1000;
@@ -51,10 +58,10 @@
// https://bugs.chromium.org/p/webm/issues/detail?id=1332
// The functions specified do not pass with INT16_MIN/MAX. They fail at the
// value specified, but pass when 1 is added/subtracted.
-int16_t MaxSupportedCoeff(InvTxfmFunc a) {
-#if HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_VP9_HIGHBITDEPTH && \
- !CONFIG_EMULATE_HARDWARE
- if (a == vpx_idct8x8_64_add_ssse3 || a == vpx_idct8x8_12_add_ssse3) {
+int16_t MaxSupportedCoeff(InvTxfmWithBdFunc a) {
+#if HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_EMULATE_HARDWARE
+ if (a == &wrapper<vpx_idct8x8_64_add_ssse3> ||
+ a == &wrapper<vpx_idct8x8_12_add_ssse3>) {
return 23625 - 1;
}
#else
@@ -63,14 +70,15 @@
return std::numeric_limits<int16_t>::max();
}
-int16_t MinSupportedCoeff(InvTxfmFunc a) {
- (void)a;
-#if HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_VP9_HIGHBITDEPTH && \
- !CONFIG_EMULATE_HARDWARE
- if (a == vpx_idct8x8_64_add_ssse3 || a == vpx_idct8x8_12_add_ssse3) {
+int16_t MinSupportedCoeff(InvTxfmWithBdFunc a) {
+#if HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_EMULATE_HARDWARE
+ if (a == &wrapper<vpx_idct8x8_64_add_ssse3> ||
+ a == &wrapper<vpx_idct8x8_12_add_ssse3>) {
return -23625 + 1;
}
-#endif // !CONFIG_EMULATE_HARDWARE
+#else
+ (void)a;
+#endif
return std::numeric_limits<int16_t>::min();
}
@@ -85,6 +93,7 @@
tx_size_ = GET_PARAM(3);
last_nonzero_ = GET_PARAM(4);
bit_depth_ = GET_PARAM(5);
+ pixel_size_ = GET_PARAM(6);
mask_ = (1 << bit_depth_) - 1;
switch (tx_size_) {
@@ -110,10 +119,10 @@
input_block_ = reinterpret_cast<tran_low_t *>(
vpx_memalign(16, sizeof(*input_block_) * input_block_size_));
- output_block_ = reinterpret_cast<Pixel *>(
- vpx_memalign(16, sizeof(*output_block_) * output_block_size_));
- output_block_ref_ = reinterpret_cast<Pixel *>(
- vpx_memalign(16, sizeof(*output_block_ref_) * output_block_size_));
+ output_block_ = reinterpret_cast<uint8_t *>(
+ vpx_memalign(16, pixel_size_ * output_block_size_));
+ output_block_ref_ = reinterpret_cast<uint8_t *>(
+ vpx_memalign(16, pixel_size_ * output_block_size_));
}
virtual void TearDown() {
@@ -128,8 +137,18 @@
void InitMem() {
memset(input_block_, 0, sizeof(*input_block_) * input_block_size_);
- for (int j = 0; j < output_block_size_; ++j) {
- output_block_[j] = output_block_ref_[j] = rnd_.Rand16() & mask_;
+ if (pixel_size_ == 1) {
+ for (int j = 0; j < output_block_size_; ++j) {
+ output_block_[j] = output_block_ref_[j] = rnd_.Rand16() & mask_;
+ }
+ } else {
+ ASSERT_EQ(2, pixel_size_);
+ uint16_t *const output = reinterpret_cast<uint16_t *>(output_block_);
+ uint16_t *const output_ref =
+ reinterpret_cast<uint16_t *>(output_block_ref_);
+ for (int j = 0; j < output_block_size_; ++j) {
+ output[j] = output_ref[j] = rnd_.Rand16() & mask_;
+ }
}
}
@@ -148,29 +167,22 @@
}
}
- void Exec(InvTxfmFunc func, void *out) {
-#if CONFIG_VP9_HIGHBITDEPTH
- func(input_block_, CONVERT_TO_BYTEPTR(out), stride_, bit_depth_);
-#else
- func(input_block_, reinterpret_cast<uint8_t *>(out), stride_);
-#endif
- }
-
protected:
int last_nonzero_;
TX_SIZE tx_size_;
tran_low_t *input_block_;
- Pixel *output_block_;
- Pixel *output_block_ref_;
+ uint8_t *output_block_;
+ uint8_t *output_block_ref_;
int size_;
int stride_;
+ int pixel_size_;
int input_block_size_;
int output_block_size_;
int bit_depth_;
int mask_;
FwdTxfmFunc ftxfm_;
- InvTxfmFunc full_itxfm_;
- InvTxfmFunc partial_itxfm_;
+ InvTxfmWithBdFunc full_itxfm_;
+ InvTxfmWithBdFunc partial_itxfm_;
ACMRandom rnd_;
};
@@ -204,10 +216,12 @@
input_block_[pos] = (output_ref_block[pos] / 4) * 4;
}
- ASM_REGISTER_STATE_CHECK(Exec(full_itxfm_, output_block_ref_));
- ASM_REGISTER_STATE_CHECK(Exec(partial_itxfm_, output_block_));
+ ASM_REGISTER_STATE_CHECK(
+ full_itxfm_(input_block_, output_block_ref_, stride_, bit_depth_));
+ ASM_REGISTER_STATE_CHECK(
+ partial_itxfm_(input_block_, output_block_, stride_, bit_depth_));
ASSERT_EQ(0, memcmp(output_block_ref_, output_block_,
- sizeof(*output_block_) * output_block_size_))
+ pixel_size_ * output_block_size_))
<< "Error: partial inverse transform produces different results";
}
}
@@ -217,11 +231,12 @@
InitMem();
InitInput();
- ASM_REGISTER_STATE_CHECK(Exec(full_itxfm_, output_block_ref_));
- ASM_REGISTER_STATE_CHECK(Exec(partial_itxfm_, output_block_));
-
+ ASM_REGISTER_STATE_CHECK(
+ full_itxfm_(input_block_, output_block_ref_, stride_, bit_depth_));
+ ASM_REGISTER_STATE_CHECK(
+ partial_itxfm_(input_block_, output_block_, stride_, bit_depth_));
ASSERT_EQ(0, memcmp(output_block_ref_, output_block_,
- sizeof(*output_block_) * output_block_size_))
+ pixel_size_ * output_block_size_))
<< "Error: partial inverse transform produces different results";
}
}
@@ -233,11 +248,12 @@
input_block_[vp9_default_scan_orders[tx_size_].scan[j]] = 10;
}
- ASM_REGISTER_STATE_CHECK(Exec(full_itxfm_, output_block_ref_));
- ASM_REGISTER_STATE_CHECK(Exec(partial_itxfm_, output_block_));
-
+ ASM_REGISTER_STATE_CHECK(
+ full_itxfm_(input_block_, output_block_ref_, stride_, bit_depth_));
+ ASM_REGISTER_STATE_CHECK(
+ partial_itxfm_(input_block_, output_block_, stride_, bit_depth_));
ASSERT_EQ(0, memcmp(output_block_ref_, output_block_,
- sizeof(*output_block_) * output_block_size_))
+ pixel_size_ * output_block_size_))
<< "Error: Transform results are not correctly added to output.";
}
}
@@ -251,16 +267,16 @@
for (int j = 0; j < 2; ++j) {
const int coeff = j ? min_coeff : max_coeff;
- memset(output_block_, 0, sizeof(*output_block_) * output_block_size_);
- memset(output_block_ref_, 0,
- sizeof(*output_block_ref_) * output_block_size_);
+ memset(output_block_, 0, pixel_size_ * output_block_size_);
+ memset(output_block_ref_, 0, pixel_size_ * output_block_size_);
input_block_[vp9_default_scan_orders[tx_size_].scan[i]] = coeff;
- ASM_REGISTER_STATE_CHECK(Exec(full_itxfm_, output_block_ref_));
- ASM_REGISTER_STATE_CHECK(Exec(partial_itxfm_, output_block_));
-
+ ASM_REGISTER_STATE_CHECK(
+ full_itxfm_(input_block_, output_block_ref_, stride_, bit_depth_));
+ ASM_REGISTER_STATE_CHECK(
+ partial_itxfm_(input_block_, output_block_, stride_, bit_depth_));
ASSERT_EQ(0, memcmp(output_block_ref_, output_block_,
- sizeof(*output_block_) * output_block_size_))
+ pixel_size_ * output_block_size_))
<< "Error: Fails with single coeff of " << coeff << " at " << i
<< ".";
}
@@ -274,12 +290,13 @@
InitInput();
for (int i = 0; i < kCountSpeedTestBlock; ++i) {
- ASM_REGISTER_STATE_CHECK(Exec(full_itxfm_, output_block_ref_));
+ ASM_REGISTER_STATE_CHECK(
+ full_itxfm_(input_block_, output_block_ref_, stride_, bit_depth_));
}
vpx_usec_timer timer;
vpx_usec_timer_start(&timer);
for (int i = 0; i < kCountSpeedTestBlock; ++i) {
- Exec(partial_itxfm_, output_block_);
+ partial_itxfm_(input_block_, output_block_, stride_, bit_depth_);
}
libvpx_test::ClearSystemState();
vpx_usec_timer_mark(&timer);
@@ -289,283 +306,331 @@
last_nonzero_, bit_depth_, elapsed_time);
ASSERT_EQ(0, memcmp(output_block_ref_, output_block_,
- sizeof(*output_block_) * output_block_size_))
+ pixel_size_ * output_block_size_))
<< "Error: partial inverse transform produces different results";
}
using std::tr1::make_tuple;
+const PartialInvTxfmParam c_partial_idct_tests[] = {
#if CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>, TX_32X32, 1024, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>, TX_32X32, 1024, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>, TX_32X32, 1024, 12, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_34_add_c>, TX_32X32, 34, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_34_add_c>, TX_32X32, 34, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_34_add_c>, TX_32X32, 34, 12, 2),
+ make_tuple(&vpx_highbd_fdct32x32_c,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_c>, TX_32X32, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct32x32_c,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_c>, TX_32X32, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct32x32_c,
+ &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_c>, TX_32X32, 1, 12, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>, TX_16X16, 256, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>, TX_16X16, 256, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>, TX_16X16, 256, 12, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_c>, TX_16X16, 10, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_c>, TX_16X16, 10, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_c>, TX_16X16, 10, 12, 2),
+ make_tuple(&vpx_highbd_fdct16x16_c,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_1_add_c>, TX_16X16, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct16x16_c,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_1_add_c>, TX_16X16, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct16x16_c,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_1_add_c>, TX_16X16, 1, 12, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>, TX_8X8, 64, 8, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>, TX_8X8, 64, 10, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>, TX_8X8, 64, 12, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_c>, TX_8X8, 12, 8, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_c>, TX_8X8, 12, 10, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_c>, TX_8X8, 12, 12, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_1_add_c>, TX_8X8, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_1_add_c>, TX_8X8, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_1_add_c>, TX_8X8, 1, 12, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>, TX_4X4, 16, 8, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>, TX_4X4, 16, 10, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>, TX_4X4, 16, 12, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>, TX_4X4, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>, TX_4X4, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>, TX_4X4, 1, 12, 2),
+#endif // CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_c>, TX_32X32, 1024, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_135_add_c>, TX_32X32, 135, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_34_add_c>, TX_32X32, 34, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1_add_c>, TX_32X32, 1, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_256_add_c>, TX_16X16, 256, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_10_add_c>, TX_16X16, 10, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_1_add_c>, TX_16X16, 1, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_64_add_c>, TX_8X8, 64, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_12_add_c>, TX_8X8, 12, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_1_add_c>, TX_8X8, 1, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_16_add_c>, TX_4X4, 16, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_1_add_c>, TX_4X4, 1, 8, 1)
+};
-INSTANTIATE_TEST_CASE_P(
- C, PartialIDctTest,
- ::testing::Values(
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1024_add_c, TX_32X32, 1024, 8),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1024_add_c, TX_32X32, 1024, 10),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1024_add_c, TX_32X32, 1024, 12),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_34_add_c, TX_32X32, 34, 8),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_34_add_c, TX_32X32, 34, 10),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_34_add_c, TX_32X32, 34, 12),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_c, TX_32X32, 1, 8),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_c, TX_32X32, 1, 10),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_c, TX_32X32, 1, 12),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_c, TX_16X16, 256, 8),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_c, TX_16X16, 256, 10),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_c, TX_16X16, 256, 12),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_c, TX_16X16, 10, 8),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_c, TX_16X16, 10, 10),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_c, TX_16X16, 10, 12),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_1_add_c, TX_16X16, 1, 8),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_1_add_c, TX_16X16, 1, 10),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_1_add_c, TX_16X16, 1, 12),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_c, TX_8X8, 64, 8),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_c, TX_8X8, 64, 10),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_c, TX_8X8, 64, 12),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_c, TX_8X8, 12, 8),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_c, TX_8X8, 12, 10),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_c, TX_8X8, 12, 12),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_1_add_c, TX_8X8, 1, 8),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_1_add_c, TX_8X8, 1, 10),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_1_add_c, TX_8X8, 1, 12),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_c, TX_4X4, 16, 8),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_c, TX_4X4, 16, 10),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_c, TX_4X4, 16, 12),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_1_add_c, TX_4X4, 1, 8),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_1_add_c, TX_4X4, 1, 10),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_1_add_c, TX_4X4, 1, 12)));
+INSTANTIATE_TEST_CASE_P(C, PartialIDctTest,
+ ::testing::ValuesIn(c_partial_idct_tests));
-#if HAVE_SSE2 && !CONFIG_EMULATE_HARDWARE
-INSTANTIATE_TEST_CASE_P(
- SSE2, PartialIDctTest,
- ::testing::Values(
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_sse2, TX_32X32, 1, 8),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_sse2, TX_32X32, 1, 10),
- make_tuple(&vpx_highbd_fdct32x32_c, &vpx_highbd_idct32x32_1024_add_c,
- &vpx_highbd_idct32x32_1_add_sse2, TX_32X32, 1, 12),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_sse2, TX_16X16, 256, 8),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_sse2, TX_16X16, 256, 10),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_256_add_sse2, TX_16X16, 256, 12),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_sse2, TX_16X16, 10, 8),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_sse2, TX_16X16, 10, 10),
- make_tuple(&vpx_highbd_fdct16x16_c, &vpx_highbd_idct16x16_256_add_c,
- &vpx_highbd_idct16x16_10_add_sse2, TX_16X16, 10, 12),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_sse2, TX_8X8, 64, 8),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_sse2, TX_8X8, 64, 10),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_64_add_sse2, TX_8X8, 64, 12),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_sse2, TX_8X8, 12, 8),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_sse2, TX_8X8, 12, 10),
- make_tuple(&vpx_highbd_fdct8x8_c, &vpx_highbd_idct8x8_64_add_c,
- &vpx_highbd_idct8x8_12_add_sse2, TX_8X8, 12, 12),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_sse2, TX_4X4, 1, 8),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_sse2, TX_4X4, 1, 10),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_sse2, TX_4X4, 1, 12)));
-#endif // HAVE_SSE2 && !CONFIG_EMULATE_HARDWARE
-
#if HAVE_NEON && !CONFIG_EMULATE_HARDWARE
-INSTANTIATE_TEST_CASE_P(
- NEON, PartialIDctTest,
- ::testing::Values(
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_neon, TX_4X4, 16, 8),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_neon, TX_4X4, 16, 10),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_16_add_c,
- &vpx_highbd_idct4x4_16_add_neon, TX_4X4, 16, 12),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_1_add_c,
- &vpx_highbd_idct4x4_1_add_neon, TX_4X4, 1, 8),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_1_add_c,
- &vpx_highbd_idct4x4_1_add_neon, TX_4X4, 1, 10),
- make_tuple(&vpx_highbd_fdct4x4_c, &vpx_highbd_idct4x4_1_add_c,
- &vpx_highbd_idct4x4_1_add_neon, TX_4X4, 1, 12)));
-#endif // HAVE_NEON && !CONFIG_EMULATE_HARDWARE
+const PartialInvTxfmParam neon_partial_idct_tests[] = {
+#if CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_neon>, TX_4X4, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_neon>, TX_4X4, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_neon>, TX_4X4, 1, 12, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c, &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_neon>, TX_4X4, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c, &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_neon>, TX_4X4, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c, &highbd_wrapper<vpx_highbd_idct4x4_1_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_1_add_neon>, TX_4X4, 1, 12, 2),
+#endif // CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_neon>, TX_32X32, 1024, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_135_add_neon>, TX_32X32, 135, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_34_add_neon>, TX_32X32, 34, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1_add_neon>, TX_32X32, 1, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_256_add_neon>, TX_16X16, 256, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_10_add_neon>, TX_16X16, 10, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_1_add_neon>, TX_16X16, 1, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_64_add_neon>, TX_8X8, 64, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_12_add_neon>, TX_8X8, 12, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_1_add_neon>, TX_8X8, 1, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_16_add_neon>, TX_4X4, 16, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_1_add_neon>, TX_4X4, 1, 8, 1)
+};
-#else // !CONFIG_VP9_HIGHBITDEPTH
-
-INSTANTIATE_TEST_CASE_P(
- C, PartialIDctTest,
- ::testing::Values(make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_c, TX_32X32, 1024, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_135_add_c, TX_32X32, 135, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_34_add_c, TX_32X32, 34, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1_add_c, TX_32X32, 1, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_256_add_c, TX_16X16, 256, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_10_add_c, TX_16X16, 10, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_1_add_c, TX_16X16, 1, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_64_add_c, TX_8X8, 64, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_12_add_c, TX_8X8, 12, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_1_add_c, TX_8X8, 1, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_16_add_c, TX_4X4, 16, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_1_add_c, TX_4X4, 1, 8)));
-
-#if HAVE_NEON && !CONFIG_EMULATE_HARDWARE
-INSTANTIATE_TEST_CASE_P(
- NEON, PartialIDctTest,
- ::testing::Values(make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_neon, TX_32X32, 1024,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_135_add_neon, TX_32X32, 135, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_34_add_neon, TX_32X32, 34, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1_add_neon, TX_32X32, 1, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_256_add_neon, TX_16X16, 256, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_10_add_neon, TX_16X16, 10, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_1_add_neon, TX_16X16, 1, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_64_add_neon, TX_8X8, 64, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_12_add_neon, TX_8X8, 12, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_1_add_neon, TX_8X8, 1, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_16_add_neon, TX_4X4, 16, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_1_add_neon, TX_4X4, 1, 8)));
+INSTANTIATE_TEST_CASE_P(NEON, PartialIDctTest,
+ ::testing::ValuesIn(neon_partial_idct_tests));
#endif // HAVE_NEON && !CONFIG_EMULATE_HARDWARE
#if HAVE_SSE2 && !CONFIG_EMULATE_HARDWARE
// 32x32_135_ is implemented using the 1024 version.
-INSTANTIATE_TEST_CASE_P(
- SSE2, PartialIDctTest,
- ::testing::Values(make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_sse2, TX_32X32, 1024,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_sse2, TX_32X32, 135,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_34_add_sse2, TX_32X32, 34, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1_add_sse2, TX_32X32, 1, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_256_add_sse2, TX_16X16, 256, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_10_add_sse2, TX_16X16, 10, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_1_add_sse2, TX_16X16, 1, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_64_add_sse2, TX_8X8, 64, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_12_add_sse2, TX_8X8, 12, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_1_add_sse2, TX_8X8, 1, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_16_add_sse2, TX_4X4, 16, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_1_add_sse2, TX_4X4, 1, 8)));
+const PartialInvTxfmParam sse2_partial_idct_tests[] = {
+#if CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_sse2>, TX_32X32, 1, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_sse2>, TX_32X32, 1, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct32x32_c, &highbd_wrapper<vpx_highbd_idct32x32_1024_add_c>,
+ &highbd_wrapper<vpx_highbd_idct32x32_1_add_sse2>, TX_32X32, 1, 12, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_sse2>, TX_16X16, 256, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_sse2>, TX_16X16, 256, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_256_add_sse2>, TX_16X16, 256, 12, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_sse2>, TX_16X16, 10, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_sse2>, TX_16X16, 10, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct16x16_c, &highbd_wrapper<vpx_highbd_idct16x16_256_add_c>,
+ &highbd_wrapper<vpx_highbd_idct16x16_10_add_sse2>, TX_16X16, 10, 12, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_sse2>, TX_8X8, 64, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct8x8_c, &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_sse2>, TX_8X8, 64, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct8x8_c, &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_sse2>, TX_8X8, 64, 12, 2),
+ make_tuple(&vpx_highbd_fdct8x8_c,
+ &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_sse2>, TX_8X8, 12, 8, 2),
+ make_tuple(
+ &vpx_highbd_fdct8x8_c, &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_sse2>, TX_8X8, 12, 10, 2),
+ make_tuple(
+ &vpx_highbd_fdct8x8_c, &highbd_wrapper<vpx_highbd_idct8x8_64_add_c>,
+ &highbd_wrapper<vpx_highbd_idct8x8_12_add_sse2>, TX_8X8, 12, 12, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_sse2>, TX_4X4, 1, 8, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_sse2>, TX_4X4, 1, 10, 2),
+ make_tuple(&vpx_highbd_fdct4x4_c,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_c>,
+ &highbd_wrapper<vpx_highbd_idct4x4_16_add_sse2>, TX_4X4, 1, 12, 2),
+#endif // CONFIG_VP9_HIGHBITDEPTH
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_sse2>, TX_32X32, 1024, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_sse2>, TX_32X32, 135, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_34_add_sse2>, TX_32X32, 34, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1_add_sse2>, TX_32X32, 1, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_256_add_sse2>, TX_16X16, 256, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_10_add_sse2>, TX_16X16, 10, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_1_add_sse2>, TX_16X16, 1, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_64_add_sse2>, TX_8X8, 64, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_12_add_sse2>, TX_8X8, 12, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_1_add_sse2>, TX_8X8, 1, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_16_add_sse2>, TX_4X4, 16, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_1_add_sse2>, TX_4X4, 1, 8, 1)
+};
+
+INSTANTIATE_TEST_CASE_P(SSE2, PartialIDctTest,
+ ::testing::ValuesIn(sse2_partial_idct_tests));
+
#endif // HAVE_SSE2 && !CONFIG_EMULATE_HARDWARE
#if HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_EMULATE_HARDWARE
-INSTANTIATE_TEST_CASE_P(
- SSSE3_64, PartialIDctTest,
- ::testing::Values(make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_ssse3, TX_32X32, 1024,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_135_add_ssse3, TX_32X32, 135,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_34_add_ssse3, TX_32X32, 34, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_64_add_ssse3, TX_8X8, 64, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_12_add_ssse3, TX_8X8, 12, 8)));
+const PartialInvTxfmParam ssse3_partial_idct_tests[] = {
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_ssse3>, TX_32X32, 1024, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_135_add_ssse3>, TX_32X32, 135, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_34_add_ssse3>, TX_32X32, 34, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_64_add_ssse3>, TX_8X8, 64, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_12_add_ssse3>, TX_8X8, 12, 8, 1)
+};
+
+INSTANTIATE_TEST_CASE_P(SSSE3, PartialIDctTest,
+ ::testing::ValuesIn(ssse3_partial_idct_tests));
#endif // HAVE_SSSE3 && ARCH_X86_64 && !CONFIG_EMULATE_HARDWARE
-#if HAVE_MSA && !CONFIG_EMULATE_HARDWARE
+#if HAVE_MSA && !CONFIG_EMULATE_HARDWARE && !CONFIG_VP9_HIGHBITDEPTH
// 32x32_135_ is implemented using the 1024 version.
-INSTANTIATE_TEST_CASE_P(
- MSA, PartialIDctTest,
- ::testing::Values(make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_msa, TX_32X32, 1024,
- 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1024_add_msa, TX_32X32, 135, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_34_add_msa, TX_32X32, 34, 8),
- make_tuple(&vpx_fdct32x32_c, &vpx_idct32x32_1024_add_c,
- &vpx_idct32x32_1_add_msa, TX_32X32, 1, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_256_add_msa, TX_16X16, 256, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_10_add_msa, TX_16X16, 10, 8),
- make_tuple(&vpx_fdct16x16_c, &vpx_idct16x16_256_add_c,
- &vpx_idct16x16_1_add_msa, TX_16X16, 1, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_64_add_msa, TX_8X8, 64, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_12_add_msa, TX_8X8, 10, 8),
- make_tuple(&vpx_fdct8x8_c, &vpx_idct8x8_64_add_c,
- &vpx_idct8x8_1_add_msa, TX_8X8, 1, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_16_add_msa, TX_4X4, 16, 8),
- make_tuple(&vpx_fdct4x4_c, &vpx_idct4x4_16_add_c,
- &vpx_idct4x4_1_add_msa, TX_4X4, 1, 8)));
-#endif // HAVE_MSA && !CONFIG_EMULATE_HARDWARE
+const PartialInvTxfmParam msa_partial_idct_tests[] = {
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_msa>, TX_32X32, 1024, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1024_add_msa>, TX_32X32, 135, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_34_add_msa>, TX_32X32, 34, 8, 1),
+ make_tuple(&vpx_fdct32x32_c, &wrapper<vpx_idct32x32_1024_add_c>,
+ &wrapper<vpx_idct32x32_1_add_msa>, TX_32X32, 1, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_256_add_msa>, TX_16X16, 256, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_10_add_msa>, TX_16X16, 10, 8, 1),
+ make_tuple(&vpx_fdct16x16_c, &wrapper<vpx_idct16x16_256_add_c>,
+ &wrapper<vpx_idct16x16_1_add_msa>, TX_16X16, 1, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_64_add_msa>, TX_8X8, 64, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_12_add_msa>, TX_8X8, 10, 8, 1),
+ make_tuple(&vpx_fdct8x8_c, &wrapper<vpx_idct8x8_64_add_c>,
+ &wrapper<vpx_idct8x8_1_add_msa>, TX_8X8, 1, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_16_add_msa>, TX_4X4, 16, 8, 1),
+ make_tuple(&vpx_fdct4x4_c, &wrapper<vpx_idct4x4_16_add_c>,
+ &wrapper<vpx_idct4x4_1_add_msa>, TX_4X4, 1, 8, 1)
+};
-#endif // CONFIG_VP9_HIGHBITDEPTH
+INSTANTIATE_TEST_CASE_P(MSA, PartialIDctTest,
+ ::testing::ValuesIn(msa_partial_idct_tests));
+#endif // HAVE_MSA && !CONFIG_EMULATE_HARDWARE && !CONFIG_VP9_HIGHBITDEPTH
} // namespace