blob: 0c623dfd3b82e1ed4cf5aba0852c1dc31f4bccb7 [file] [log] [blame]
/*
* Copyright (c) 2016 The WebM project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include <assert.h>
#include <smmintrin.h> /* SSE4.1 */
#include "./vp10_rtcd.h"
#include "./vpx_config.h"
#include "vp10/common/vp10_inv_txfm2d_cfg.h"
static INLINE void load_buffer_4x4(const int32_t *coeff, __m128i *in) {
in[0] = _mm_loadu_si128((const __m128i *)(coeff + 0));
in[1] = _mm_loadu_si128((const __m128i *)(coeff + 4));
in[2] = _mm_loadu_si128((const __m128i *)(coeff + 8));
in[3] = _mm_loadu_si128((const __m128i *)(coeff + 12));
}
static void idct4x4_sse4_1(__m128i *in, int bit) {
const int32_t *cospi = cospi_arr[bit - cos_bit_min];
const __m128i cospi32 = _mm_set1_epi32(cospi[32]);
const __m128i cospi48 = _mm_set1_epi32(cospi[48]);
const __m128i cospi16 = _mm_set1_epi32(cospi[16]);
const __m128i cospim16 = _mm_set1_epi32(-cospi[16]);
const __m128i rnding = _mm_set1_epi32(1 << (bit - 1));
__m128i u0, u1, u2, u3;
__m128i v0, v1, v2, v3, x, y;
v0 = _mm_unpacklo_epi32(in[0], in[1]);
v1 = _mm_unpackhi_epi32(in[0], in[1]);
v2 = _mm_unpacklo_epi32(in[2], in[3]);
v3 = _mm_unpackhi_epi32(in[2], in[3]);
u0 = _mm_unpacklo_epi64(v0, v2);
u1 = _mm_unpackhi_epi64(v0, v2);
u2 = _mm_unpacklo_epi64(v1, v3);
u3 = _mm_unpackhi_epi64(v1, v3);
x = _mm_mullo_epi32(u0, cospi32);
y = _mm_mullo_epi32(u2, cospi32);
v0 = _mm_add_epi32(x, y);
v0 = _mm_add_epi32(v0, rnding);
v0 = _mm_srai_epi32(v0, bit);
v1 = _mm_sub_epi32(x, y);
v1 = _mm_add_epi32(v1, rnding);
v1 = _mm_srai_epi32(v1, bit);
x = _mm_mullo_epi32(u1, cospi48);
y = _mm_mullo_epi32(u3, cospim16);
v2 = _mm_add_epi32(x, y);
v2 = _mm_add_epi32(v2, rnding);
v2 = _mm_srai_epi32(v2, bit);
x = _mm_mullo_epi32(u1, cospi16);
y = _mm_mullo_epi32(u3, cospi48);
v3 = _mm_add_epi32(x, y);
v3 = _mm_add_epi32(v3, rnding);
v3 = _mm_srai_epi32(v3, bit);
in[0] = _mm_add_epi32(v0, v3);
in[1] = _mm_add_epi32(v1, v2);
in[2] = _mm_sub_epi32(v1, v2);
in[3] = _mm_sub_epi32(v0, v3);
}
static void iadst4x4_sse4_1(__m128i *in, int bit) {
const int32_t *cospi = cospi_arr[bit - cos_bit_min];
const __m128i cospi32 = _mm_set1_epi32(cospi[32]);
const __m128i cospi8 = _mm_set1_epi32(cospi[8]);
const __m128i cospim8 = _mm_set1_epi32(-cospi[8]);
const __m128i cospi40 = _mm_set1_epi32(cospi[40]);
const __m128i cospim40 = _mm_set1_epi32(-cospi[40]);
const __m128i cospi24 = _mm_set1_epi32(cospi[24]);
const __m128i cospi56 = _mm_set1_epi32(cospi[56]);
const __m128i rnding = _mm_set1_epi32(1 << (bit - 1));
const __m128i zero = _mm_setzero_si128();
__m128i u0, u1, u2, u3;
__m128i v0, v1, v2, v3, x, y;
v0 = _mm_unpacklo_epi32(in[0], in[1]);
v1 = _mm_unpackhi_epi32(in[0], in[1]);
v2 = _mm_unpacklo_epi32(in[2], in[3]);
v3 = _mm_unpackhi_epi32(in[2], in[3]);
u0 = _mm_unpacklo_epi64(v0, v2);
u1 = _mm_unpackhi_epi64(v0, v2);
u2 = _mm_unpacklo_epi64(v1, v3);
u3 = _mm_unpackhi_epi64(v1, v3);
// stage 0
// stage 1
u1 = _mm_sub_epi32(zero, u1);
u3 = _mm_sub_epi32(zero, u3);
// stage 2
v0 = u0;
v1 = u3;
x = _mm_mullo_epi32(u1, cospi32);
y = _mm_mullo_epi32(u2, cospi32);
v2 = _mm_add_epi32(x, y);
v2 = _mm_add_epi32(v2, rnding);
v2 = _mm_srai_epi32(v2, bit);
v3 = _mm_sub_epi32(x, y);
v3 = _mm_add_epi32(v3, rnding);
v3 = _mm_srai_epi32(v3, bit);
// stage 3
u0 = _mm_add_epi32(v0, v2);
u1 = _mm_add_epi32(v1, v3);
u2 = _mm_sub_epi32(v0, v2);
u3 = _mm_sub_epi32(v1, v3);
// stage 4
x = _mm_mullo_epi32(u0, cospi8);
y = _mm_mullo_epi32(u1, cospi56);
in[3] = _mm_add_epi32(x, y);
in[3] = _mm_add_epi32(in[3], rnding);
in[3] = _mm_srai_epi32(in[3], bit);
x = _mm_mullo_epi32(u0, cospi56);
y = _mm_mullo_epi32(u1, cospim8);
in[0] = _mm_add_epi32(x, y);
in[0] = _mm_add_epi32(in[0], rnding);
in[0] = _mm_srai_epi32(in[0], bit);
x = _mm_mullo_epi32(u2, cospi40);
y = _mm_mullo_epi32(u3, cospi24);
in[1] = _mm_add_epi32(x, y);
in[1] = _mm_add_epi32(in[1], rnding);
in[1] = _mm_srai_epi32(in[1], bit);
x = _mm_mullo_epi32(u2, cospi24);
y = _mm_mullo_epi32(u3, cospim40);
in[2] = _mm_add_epi32(x, y);
in[2] = _mm_add_epi32(in[2], rnding);
in[2] = _mm_srai_epi32(in[2], bit);
}
static INLINE void round_shift_4x4(__m128i *in, int shift) {
__m128i rnding = _mm_set1_epi32(1 << (shift - 1));
in[0] = _mm_add_epi32(in[0], rnding);
in[1] = _mm_add_epi32(in[1], rnding);
in[2] = _mm_add_epi32(in[2], rnding);
in[3] = _mm_add_epi32(in[3], rnding);
in[0] = _mm_srai_epi32(in[0], shift);
in[1] = _mm_srai_epi32(in[1], shift);
in[2] = _mm_srai_epi32(in[2], shift);
in[3] = _mm_srai_epi32(in[3], shift);
}
static INLINE __m128i highbd_clamp_epi16(__m128i u, int bd) {
const __m128i zero = _mm_setzero_si128();
const __m128i one = _mm_set1_epi16(1);
const __m128i max = _mm_sub_epi16(_mm_slli_epi16(one, bd), one);
__m128i clamped, mask;
mask = _mm_cmpgt_epi16(u, max);
clamped = _mm_andnot_si128(mask, u);
mask = _mm_and_si128(mask, max);
clamped = _mm_or_si128(mask, clamped);
mask = _mm_cmpgt_epi16(clamped, zero);
clamped = _mm_and_si128(clamped, mask);
return clamped;
}
static void write_buffer_4x4(__m128i *in, uint16_t *output, int stride,
int flipud, int fliplr, int shift, int bd) {
const __m128i zero = _mm_setzero_si128();
__m128i u0, u1, u2, u3;
__m128i v0, v1, v2, v3;
round_shift_4x4(in, shift);
v0 = _mm_loadl_epi64((__m128i const *)(output + 0 * stride));
v1 = _mm_loadl_epi64((__m128i const *)(output + 1 * stride));
v2 = _mm_loadl_epi64((__m128i const *)(output + 2 * stride));
v3 = _mm_loadl_epi64((__m128i const *)(output + 3 * stride));
v0 = _mm_unpacklo_epi16(v0, zero);
v1 = _mm_unpacklo_epi16(v1, zero);
v2 = _mm_unpacklo_epi16(v2, zero);
v3 = _mm_unpacklo_epi16(v3, zero);
u0 = _mm_add_epi32(in[0], v0);
u1 = _mm_add_epi32(in[1], v1);
u2 = _mm_add_epi32(in[2], v2);
u3 = _mm_add_epi32(in[3], v3);
v0 = _mm_packus_epi32(u0, u1);
v2 = _mm_packus_epi32(u2, u3);
u0 = highbd_clamp_epi16(v0, bd);
u2 = highbd_clamp_epi16(v2, bd);
v0 = _mm_unpacklo_epi64(u0, u0);
v1 = _mm_unpackhi_epi64(u0, u0);
v2 = _mm_unpacklo_epi64(u2, u2);
v3 = _mm_unpackhi_epi64(u2, u2);
_mm_storel_epi64((__m128i *)(output + 0 * stride), v0);
_mm_storel_epi64((__m128i *)(output + 1 * stride), v1);
_mm_storel_epi64((__m128i *)(output + 2 * stride), v2);
_mm_storel_epi64((__m128i *)(output + 3 * stride), v3);
(void) flipud;
(void) fliplr;
}
void vp10_inv_txfm2d_add_4x4_sse4_1(const int32_t *coeff, uint16_t *output,
int stride, int tx_type, int bd) {
__m128i in[4];
const TXFM_2D_CFG *cfg = NULL;
switch (tx_type) {
case DCT_DCT:
cfg = &inv_txfm_2d_cfg_dct_dct_4;
load_buffer_4x4(coeff, in);
idct4x4_sse4_1(in, cfg->cos_bit_row[2]);
idct4x4_sse4_1(in, cfg->cos_bit_row[2]);
write_buffer_4x4(in, output, stride, 0, 0, -cfg->shift[1], bd);
break;
case ADST_DCT:
cfg = &inv_txfm_2d_cfg_adst_dct_4;
load_buffer_4x4(coeff, in);
idct4x4_sse4_1(in, cfg->cos_bit_row[2]);
iadst4x4_sse4_1(in, cfg->cos_bit_row[2]);
write_buffer_4x4(in, output, stride, 0, 0, -cfg->shift[1], bd);
break;
case DCT_ADST:
cfg = &inv_txfm_2d_cfg_dct_adst_4;
load_buffer_4x4(coeff, in);
iadst4x4_sse4_1(in, cfg->cos_bit_row[2]);
idct4x4_sse4_1(in, cfg->cos_bit_row[2]);
write_buffer_4x4(in, output, stride, 0, 0, -cfg->shift[1], bd);
break;
case ADST_ADST:
cfg = &inv_txfm_2d_cfg_adst_adst_4;
load_buffer_4x4(coeff, in);
iadst4x4_sse4_1(in, cfg->cos_bit_row[2]);
iadst4x4_sse4_1(in, cfg->cos_bit_row[2]);
write_buffer_4x4(in, output, stride, 0, 0, -cfg->shift[1], bd);
break;
default:
assert(0);
}
}