| /* |
| * Copyright (c) 2012 The WebM project authors. All Rights Reserved. |
| * |
| * Use of this source code is governed by a BSD-style license |
| * that can be found in the LICENSE file in the root of the source |
| * tree. An additional intellectual property rights grant can be found |
| * in the file PATENTS. All contributing project authors may |
| * be found in the AUTHORS file in the root of the source tree. |
| */ |
| #include "./vpx_config.h" |
| |
| #include "vp9/encoder/vp9_variance.h" |
| #include "vpx_ports/mem.h" |
| |
| typedef void (*get_var_avx2) ( |
| const unsigned char *src_ptr, |
| int source_stride, |
| const unsigned char *ref_ptr, |
| int recon_stride, |
| unsigned int *SSE, |
| int *Sum |
| ); |
| |
| void vp9_get16x16var_avx2 |
| ( |
| const unsigned char *src_ptr, |
| int source_stride, |
| const unsigned char *ref_ptr, |
| int recon_stride, |
| unsigned int *SSE, |
| int *Sum |
| ); |
| |
| void vp9_get32x32var_avx2 |
| ( |
| const unsigned char *src_ptr, |
| int source_stride, |
| const unsigned char *ref_ptr, |
| int recon_stride, |
| unsigned int *SSE, |
| int *Sum |
| ); |
| |
| unsigned int vp9_sub_pixel_variance32xh_avx2 |
| ( |
| const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| int height, |
| unsigned int *sse |
| ); |
| |
| unsigned int vp9_sub_pixel_avg_variance32xh_avx2 |
| ( |
| const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| const uint8_t *sec, |
| int sec_stride, |
| int height, |
| unsigned int *sseptr |
| ); |
| |
| static void variance_avx2(const unsigned char *src_ptr, int source_stride, |
| const unsigned char *ref_ptr, int recon_stride, |
| int w, int h, unsigned int *sse, int *sum, |
| get_var_avx2 var_fn, int block_size) { |
| unsigned int sse0; |
| int sum0; |
| int i, j; |
| |
| *sse = 0; |
| *sum = 0; |
| |
| for (i = 0; i < h; i += 16) { |
| for (j = 0; j < w; j += block_size) { |
| // processing 16 rows horizontally each call |
| var_fn(src_ptr + source_stride * i + j, source_stride, |
| ref_ptr + recon_stride * i + j, recon_stride, &sse0, &sum0); |
| *sse += sse0; |
| *sum += sum0; |
| } |
| } |
| } |
| |
| unsigned int vp9_variance16x16_avx2 |
| ( |
| const unsigned char *src_ptr, |
| int source_stride, |
| const unsigned char *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int var; |
| int avg; |
| |
| variance_avx2(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, |
| &var, &avg, vp9_get16x16var_avx2, 16); |
| *sse = var; |
| return (var - (((unsigned int)avg * avg) >> 8)); |
| } |
| |
| unsigned int vp9_mse16x16_avx2( |
| const unsigned char *src_ptr, |
| int source_stride, |
| const unsigned char *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int sse0; |
| int sum0; |
| vp9_get16x16var_avx2(src_ptr, source_stride, ref_ptr, recon_stride, &sse0, |
| &sum0); |
| *sse = sse0; |
| return sse0; |
| } |
| |
| unsigned int vp9_variance32x32_avx2(const uint8_t *src_ptr, |
| int source_stride, |
| const uint8_t *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int var; |
| int avg; |
| |
| // processing 32 elements vertically in parallel |
| variance_avx2(src_ptr, source_stride, ref_ptr, recon_stride, 32, 32, |
| &var, &avg, vp9_get32x32var_avx2, 32); |
| *sse = var; |
| return (var - (((int64_t)avg * avg) >> 10)); |
| } |
| |
| unsigned int vp9_variance32x16_avx2(const uint8_t *src_ptr, |
| int source_stride, |
| const uint8_t *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int var; |
| int avg; |
| |
| // processing 32 elements vertically in parallel |
| variance_avx2(src_ptr, source_stride, ref_ptr, recon_stride, 32, 16, |
| &var, &avg, vp9_get32x32var_avx2, 32); |
| *sse = var; |
| return (var - (((int64_t)avg * avg) >> 9)); |
| } |
| |
| |
| unsigned int vp9_variance64x64_avx2(const uint8_t *src_ptr, |
| int source_stride, |
| const uint8_t *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int var; |
| int avg; |
| |
| // processing 32 elements vertically in parallel |
| variance_avx2(src_ptr, source_stride, ref_ptr, recon_stride, 64, 64, |
| &var, &avg, vp9_get32x32var_avx2, 32); |
| *sse = var; |
| return (var - (((int64_t)avg * avg) >> 12)); |
| } |
| |
| unsigned int vp9_variance64x32_avx2(const uint8_t *src_ptr, |
| int source_stride, |
| const uint8_t *ref_ptr, |
| int recon_stride, |
| unsigned int *sse) { |
| unsigned int var; |
| int avg; |
| |
| // processing 32 elements vertically in parallel |
| variance_avx2(src_ptr, source_stride, ref_ptr, recon_stride, 64, 32, |
| &var, &avg, vp9_get32x32var_avx2, 32); |
| |
| *sse = var; |
| return (var - (((int64_t)avg * avg) >> 11)); |
| } |
| |
| unsigned int vp9_sub_pixel_variance64x64_avx2(const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| unsigned int *sse_ptr) { |
| // processing 32 elements in parallel |
| unsigned int sse; |
| int se = vp9_sub_pixel_variance32xh_avx2(src, src_stride, x_offset, |
| y_offset, dst, dst_stride, |
| 64, &sse); |
| // processing the next 32 elements in parallel |
| unsigned int sse2; |
| int se2 = vp9_sub_pixel_variance32xh_avx2(src + 32, src_stride, |
| x_offset, y_offset, |
| dst + 32, dst_stride, |
| 64, &sse2); |
| se += se2; |
| sse += sse2; |
| *sse_ptr = sse; |
| return sse - (((int64_t)se * se) >> 12); |
| } |
| |
| unsigned int vp9_sub_pixel_variance32x32_avx2(const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| unsigned int *sse_ptr) { |
| // processing 32 element in parallel |
| unsigned int sse; |
| int se = vp9_sub_pixel_variance32xh_avx2(src, src_stride, x_offset, |
| y_offset, dst, dst_stride, |
| 32, &sse); |
| *sse_ptr = sse; |
| return sse - (((int64_t)se * se) >> 10); |
| } |
| |
| unsigned int vp9_sub_pixel_avg_variance64x64_avx2(const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| unsigned int *sseptr, |
| const uint8_t *sec) { |
| // processing 32 elements in parallel |
| unsigned int sse; |
| |
| int se = vp9_sub_pixel_avg_variance32xh_avx2(src, src_stride, x_offset, |
| y_offset, dst, dst_stride, |
| sec, 64, 64, &sse); |
| unsigned int sse2; |
| // processing the next 32 elements in parallel |
| int se2 = vp9_sub_pixel_avg_variance32xh_avx2(src + 32, src_stride, x_offset, |
| y_offset, dst + 32, dst_stride, |
| sec + 32, 64, 64, &sse2); |
| se += se2; |
| sse += sse2; |
| *sseptr = sse; |
| |
| return sse - (((int64_t)se * se) >> 12); |
| } |
| |
| unsigned int vp9_sub_pixel_avg_variance32x32_avx2(const uint8_t *src, |
| int src_stride, |
| int x_offset, |
| int y_offset, |
| const uint8_t *dst, |
| int dst_stride, |
| unsigned int *sseptr, |
| const uint8_t *sec) { |
| // processing 32 element in parallel |
| unsigned int sse; |
| int se = vp9_sub_pixel_avg_variance32xh_avx2(src, src_stride, x_offset, |
| y_offset, dst, dst_stride, |
| sec, 32, 32, &sse); |
| *sseptr = sse; |
| return sse - (((int64_t)se * se) >> 10); |
| } |
| |
| |