/src/mozilla-central/third_party/aom/av1/common/cfl.h
Line | Count | Source (jump to first uncovered line) |
1 | | /* |
2 | | * Copyright (c) 2016, Alliance for Open Media. All rights reserved |
3 | | * |
4 | | * This source code is subject to the terms of the BSD 2 Clause License and |
5 | | * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
6 | | * was not distributed with this source code in the LICENSE file, you can |
7 | | * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
8 | | * Media Patent License 1.0 was not distributed with this source code in the |
9 | | * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
10 | | */ |
11 | | |
12 | | #ifndef AOM_AV1_COMMON_CFL_H_ |
13 | | #define AOM_AV1_COMMON_CFL_H_ |
14 | | |
15 | | #include "av1/common/blockd.h" |
16 | | #include "av1/common/onyxc_int.h" |
17 | | |
18 | | // Can we use CfL for the current block? |
19 | 0 | static INLINE CFL_ALLOWED_TYPE is_cfl_allowed(const MACROBLOCKD *xd) { |
20 | 0 | const MB_MODE_INFO *mbmi = xd->mi[0]; |
21 | 0 | const BLOCK_SIZE bsize = mbmi->sb_type; |
22 | 0 | assert(bsize < BLOCK_SIZES_ALL); |
23 | 0 | if (xd->lossless[mbmi->segment_id]) { |
24 | 0 | // In lossless, CfL is available when the partition size is equal to the |
25 | 0 | // transform size. |
26 | 0 | const int ssx = xd->plane[AOM_PLANE_U].subsampling_x; |
27 | 0 | const int ssy = xd->plane[AOM_PLANE_U].subsampling_y; |
28 | 0 | const int plane_bsize = get_plane_block_size(bsize, ssx, ssy); |
29 | 0 | return (CFL_ALLOWED_TYPE)(plane_bsize == BLOCK_4X4); |
30 | 0 | } |
31 | 0 | // Spec: CfL is available to luma partitions lesser than or equal to 32x32 |
32 | 0 | return (CFL_ALLOWED_TYPE)(block_size_wide[bsize] <= 32 && |
33 | 0 | block_size_high[bsize] <= 32); |
34 | 0 | } Unexecuted instantiation: cfl.c:is_cfl_allowed Unexecuted instantiation: reconintra.c:is_cfl_allowed Unexecuted instantiation: cfl_avx2.c:is_cfl_allowed Unexecuted instantiation: cfl_sse2.c:is_cfl_allowed Unexecuted instantiation: cfl_ssse3.c:is_cfl_allowed Unexecuted instantiation: decodeframe.c:is_cfl_allowed Unexecuted instantiation: decodemv.c:is_cfl_allowed |
35 | | |
36 | | // Do we need to save the luma pixels from the current block, |
37 | | // for a possible future CfL prediction? |
38 | | static INLINE CFL_ALLOWED_TYPE store_cfl_required(const AV1_COMMON *cm, |
39 | 0 | const MACROBLOCKD *xd) { |
40 | 0 | const MB_MODE_INFO *mbmi = xd->mi[0]; |
41 | 0 |
|
42 | 0 | if (cm->seq_params.monochrome) return CFL_DISALLOWED; |
43 | 0 | |
44 | 0 | if (!xd->cfl.is_chroma_reference) { |
45 | 0 | // For non-chroma-reference blocks, we should always store the luma pixels, |
46 | 0 | // in case the corresponding chroma-reference block uses CfL. |
47 | 0 | // Note that this can only happen for block sizes which are <8 on |
48 | 0 | // their shortest side, as otherwise they would be chroma reference |
49 | 0 | // blocks. |
50 | 0 | return CFL_ALLOWED; |
51 | 0 | } |
52 | 0 | |
53 | 0 | // If this block has chroma information, we know whether we're |
54 | 0 | // actually going to perform a CfL prediction |
55 | 0 | return (CFL_ALLOWED_TYPE)(!is_inter_block(mbmi) && |
56 | 0 | mbmi->uv_mode == UV_CFL_PRED); |
57 | 0 | } Unexecuted instantiation: cfl.c:store_cfl_required Unexecuted instantiation: reconintra.c:store_cfl_required Unexecuted instantiation: cfl_avx2.c:store_cfl_required Unexecuted instantiation: cfl_sse2.c:store_cfl_required Unexecuted instantiation: cfl_ssse3.c:store_cfl_required Unexecuted instantiation: decodeframe.c:store_cfl_required Unexecuted instantiation: decodemv.c:store_cfl_required |
58 | | |
59 | 0 | static INLINE int get_scaled_luma_q0(int alpha_q3, int16_t pred_buf_q3) { |
60 | 0 | int scaled_luma_q6 = alpha_q3 * pred_buf_q3; |
61 | 0 | return ROUND_POWER_OF_TWO_SIGNED(scaled_luma_q6, 6); |
62 | 0 | } Unexecuted instantiation: cfl.c:get_scaled_luma_q0 Unexecuted instantiation: reconintra.c:get_scaled_luma_q0 Unexecuted instantiation: cfl_avx2.c:get_scaled_luma_q0 Unexecuted instantiation: cfl_sse2.c:get_scaled_luma_q0 Unexecuted instantiation: cfl_ssse3.c:get_scaled_luma_q0 Unexecuted instantiation: decodeframe.c:get_scaled_luma_q0 Unexecuted instantiation: decodemv.c:get_scaled_luma_q0 |
63 | | |
64 | 0 | static INLINE CFL_PRED_TYPE get_cfl_pred_type(PLANE_TYPE plane) { |
65 | 0 | assert(plane > 0); |
66 | 0 | return (CFL_PRED_TYPE)(plane - 1); |
67 | 0 | } Unexecuted instantiation: cfl.c:get_cfl_pred_type Unexecuted instantiation: reconintra.c:get_cfl_pred_type Unexecuted instantiation: cfl_avx2.c:get_cfl_pred_type Unexecuted instantiation: cfl_sse2.c:get_cfl_pred_type Unexecuted instantiation: cfl_ssse3.c:get_cfl_pred_type Unexecuted instantiation: decodeframe.c:get_cfl_pred_type Unexecuted instantiation: decodemv.c:get_cfl_pred_type |
68 | | |
69 | | void cfl_predict_block(MACROBLOCKD *const xd, uint8_t *dst, int dst_stride, |
70 | | TX_SIZE tx_size, int plane); |
71 | | |
72 | | void cfl_store_block(MACROBLOCKD *const xd, BLOCK_SIZE bsize, TX_SIZE tx_size); |
73 | | |
74 | | void cfl_store_tx(MACROBLOCKD *const xd, int row, int col, TX_SIZE tx_size, |
75 | | BLOCK_SIZE bsize); |
76 | | |
77 | | void cfl_store_dc_pred(MACROBLOCKD *const xd, const uint8_t *input, |
78 | | CFL_PRED_TYPE pred_plane, int width); |
79 | | |
80 | | void cfl_load_dc_pred(MACROBLOCKD *const xd, uint8_t *dst, int dst_stride, |
81 | | TX_SIZE tx_size, CFL_PRED_TYPE pred_plane); |
82 | | |
83 | | // Null function used for invalid tx_sizes |
84 | | void cfl_subsample_lbd_null(const uint8_t *input, int input_stride, |
85 | | uint16_t *output_q3); |
86 | | |
87 | | // Null function used for invalid tx_sizes |
88 | | void cfl_subsample_hbd_null(const uint16_t *input, int input_stride, |
89 | | uint16_t *output_q3); |
90 | | |
91 | | // Allows the CFL_SUBSAMPLE function to switch types depending on the bitdepth. |
92 | | #define CFL_lbd_TYPE uint8_t *cfl_type |
93 | | #define CFL_hbd_TYPE uint16_t *cfl_type |
94 | | |
95 | | // Declare a size-specific wrapper for the size-generic function. The compiler |
96 | | // will inline the size generic function in here, the advantage is that the size |
97 | | // will be constant allowing for loop unrolling and other constant propagated |
98 | | // goodness. |
99 | | #define CFL_SUBSAMPLE(arch, sub, bd, width, height) \ |
100 | | void subsample_##bd##_##sub##_##width##x##height##_##arch( \ |
101 | 0 | const CFL_##bd##_TYPE, int input_stride, uint16_t *output_q3) { \ |
102 | 0 | cfl_luma_subsampling_##sub##_##bd##_##arch(cfl_type, input_stride, \ |
103 | 0 | output_q3, width, height); \ |
104 | 0 | } Unexecuted instantiation: subsample_lbd_420_4x4_c Unexecuted instantiation: subsample_lbd_420_8x8_c Unexecuted instantiation: subsample_lbd_420_16x16_c Unexecuted instantiation: subsample_lbd_420_32x32_c Unexecuted instantiation: subsample_lbd_420_4x8_c Unexecuted instantiation: subsample_lbd_420_8x4_c Unexecuted instantiation: subsample_lbd_420_8x16_c Unexecuted instantiation: subsample_lbd_420_16x8_c Unexecuted instantiation: subsample_lbd_420_16x32_c Unexecuted instantiation: subsample_lbd_420_32x16_c Unexecuted instantiation: subsample_lbd_420_4x16_c Unexecuted instantiation: subsample_lbd_420_16x4_c Unexecuted instantiation: subsample_lbd_420_8x32_c Unexecuted instantiation: subsample_lbd_420_32x8_c Unexecuted instantiation: subsample_lbd_422_4x4_c Unexecuted instantiation: subsample_lbd_422_8x8_c Unexecuted instantiation: subsample_lbd_422_16x16_c Unexecuted instantiation: subsample_lbd_422_32x32_c Unexecuted instantiation: subsample_lbd_422_4x8_c Unexecuted instantiation: subsample_lbd_422_8x4_c Unexecuted instantiation: subsample_lbd_422_8x16_c Unexecuted instantiation: subsample_lbd_422_16x8_c Unexecuted instantiation: subsample_lbd_422_16x32_c Unexecuted instantiation: subsample_lbd_422_32x16_c Unexecuted instantiation: subsample_lbd_422_4x16_c Unexecuted instantiation: subsample_lbd_422_16x4_c Unexecuted instantiation: subsample_lbd_422_8x32_c Unexecuted instantiation: subsample_lbd_422_32x8_c Unexecuted instantiation: subsample_lbd_444_4x4_c Unexecuted instantiation: subsample_lbd_444_8x8_c Unexecuted instantiation: subsample_lbd_444_16x16_c Unexecuted instantiation: subsample_lbd_444_32x32_c Unexecuted instantiation: subsample_lbd_444_4x8_c Unexecuted instantiation: subsample_lbd_444_8x4_c Unexecuted instantiation: subsample_lbd_444_8x16_c Unexecuted instantiation: subsample_lbd_444_16x8_c Unexecuted instantiation: subsample_lbd_444_16x32_c Unexecuted instantiation: subsample_lbd_444_32x16_c Unexecuted instantiation: subsample_lbd_444_4x16_c Unexecuted instantiation: subsample_lbd_444_16x4_c Unexecuted instantiation: subsample_lbd_444_8x32_c Unexecuted instantiation: subsample_lbd_444_32x8_c Unexecuted instantiation: subsample_hbd_420_4x4_c Unexecuted instantiation: subsample_hbd_420_8x8_c Unexecuted instantiation: subsample_hbd_420_16x16_c Unexecuted instantiation: subsample_hbd_420_32x32_c Unexecuted instantiation: subsample_hbd_420_4x8_c Unexecuted instantiation: subsample_hbd_420_8x4_c Unexecuted instantiation: subsample_hbd_420_8x16_c Unexecuted instantiation: subsample_hbd_420_16x8_c Unexecuted instantiation: subsample_hbd_420_16x32_c Unexecuted instantiation: subsample_hbd_420_32x16_c Unexecuted instantiation: subsample_hbd_420_4x16_c Unexecuted instantiation: subsample_hbd_420_16x4_c Unexecuted instantiation: subsample_hbd_420_8x32_c Unexecuted instantiation: subsample_hbd_420_32x8_c Unexecuted instantiation: subsample_hbd_422_4x4_c Unexecuted instantiation: subsample_hbd_422_8x8_c Unexecuted instantiation: subsample_hbd_422_16x16_c Unexecuted instantiation: subsample_hbd_422_32x32_c Unexecuted instantiation: subsample_hbd_422_4x8_c Unexecuted instantiation: subsample_hbd_422_8x4_c Unexecuted instantiation: subsample_hbd_422_8x16_c Unexecuted instantiation: subsample_hbd_422_16x8_c Unexecuted instantiation: subsample_hbd_422_16x32_c Unexecuted instantiation: subsample_hbd_422_32x16_c Unexecuted instantiation: subsample_hbd_422_4x16_c Unexecuted instantiation: subsample_hbd_422_16x4_c Unexecuted instantiation: subsample_hbd_422_8x32_c Unexecuted instantiation: subsample_hbd_422_32x8_c Unexecuted instantiation: subsample_hbd_444_4x4_c Unexecuted instantiation: subsample_hbd_444_8x8_c Unexecuted instantiation: subsample_hbd_444_16x16_c Unexecuted instantiation: subsample_hbd_444_32x32_c Unexecuted instantiation: subsample_hbd_444_4x8_c Unexecuted instantiation: subsample_hbd_444_8x4_c Unexecuted instantiation: subsample_hbd_444_8x16_c Unexecuted instantiation: subsample_hbd_444_16x8_c Unexecuted instantiation: subsample_hbd_444_16x32_c Unexecuted instantiation: subsample_hbd_444_32x16_c Unexecuted instantiation: subsample_hbd_444_4x16_c Unexecuted instantiation: subsample_hbd_444_16x4_c Unexecuted instantiation: subsample_hbd_444_8x32_c Unexecuted instantiation: subsample_hbd_444_32x8_c Unexecuted instantiation: subsample_lbd_420_32x32_avx2 Unexecuted instantiation: subsample_lbd_420_32x16_avx2 Unexecuted instantiation: subsample_lbd_420_32x8_avx2 Unexecuted instantiation: subsample_lbd_422_32x32_avx2 Unexecuted instantiation: subsample_lbd_422_32x16_avx2 Unexecuted instantiation: subsample_lbd_422_32x8_avx2 Unexecuted instantiation: subsample_lbd_444_32x32_avx2 Unexecuted instantiation: subsample_lbd_444_32x16_avx2 Unexecuted instantiation: subsample_lbd_444_32x8_avx2 Unexecuted instantiation: subsample_hbd_420_32x32_avx2 Unexecuted instantiation: subsample_hbd_420_32x16_avx2 Unexecuted instantiation: subsample_hbd_420_32x8_avx2 Unexecuted instantiation: subsample_hbd_422_32x32_avx2 Unexecuted instantiation: subsample_hbd_422_32x16_avx2 Unexecuted instantiation: subsample_hbd_422_32x8_avx2 Unexecuted instantiation: subsample_hbd_444_32x32_avx2 Unexecuted instantiation: subsample_hbd_444_32x16_avx2 Unexecuted instantiation: subsample_hbd_444_32x8_avx2 Unexecuted instantiation: subsample_lbd_420_4x4_ssse3 Unexecuted instantiation: subsample_lbd_420_8x8_ssse3 Unexecuted instantiation: subsample_lbd_420_16x16_ssse3 Unexecuted instantiation: subsample_lbd_420_32x32_ssse3 Unexecuted instantiation: subsample_lbd_420_4x8_ssse3 Unexecuted instantiation: subsample_lbd_420_8x4_ssse3 Unexecuted instantiation: subsample_lbd_420_8x16_ssse3 Unexecuted instantiation: subsample_lbd_420_16x8_ssse3 Unexecuted instantiation: subsample_lbd_420_16x32_ssse3 Unexecuted instantiation: subsample_lbd_420_32x16_ssse3 Unexecuted instantiation: subsample_lbd_420_4x16_ssse3 Unexecuted instantiation: subsample_lbd_420_16x4_ssse3 Unexecuted instantiation: subsample_lbd_420_8x32_ssse3 Unexecuted instantiation: subsample_lbd_420_32x8_ssse3 Unexecuted instantiation: subsample_lbd_422_4x4_ssse3 Unexecuted instantiation: subsample_lbd_422_8x8_ssse3 Unexecuted instantiation: subsample_lbd_422_16x16_ssse3 Unexecuted instantiation: subsample_lbd_422_32x32_ssse3 Unexecuted instantiation: subsample_lbd_422_4x8_ssse3 Unexecuted instantiation: subsample_lbd_422_8x4_ssse3 Unexecuted instantiation: subsample_lbd_422_8x16_ssse3 Unexecuted instantiation: subsample_lbd_422_16x8_ssse3 Unexecuted instantiation: subsample_lbd_422_16x32_ssse3 Unexecuted instantiation: subsample_lbd_422_32x16_ssse3 Unexecuted instantiation: subsample_lbd_422_4x16_ssse3 Unexecuted instantiation: subsample_lbd_422_16x4_ssse3 Unexecuted instantiation: subsample_lbd_422_8x32_ssse3 Unexecuted instantiation: subsample_lbd_422_32x8_ssse3 Unexecuted instantiation: subsample_lbd_444_4x4_ssse3 Unexecuted instantiation: subsample_lbd_444_8x8_ssse3 Unexecuted instantiation: subsample_lbd_444_16x16_ssse3 Unexecuted instantiation: subsample_lbd_444_32x32_ssse3 Unexecuted instantiation: subsample_lbd_444_4x8_ssse3 Unexecuted instantiation: subsample_lbd_444_8x4_ssse3 Unexecuted instantiation: subsample_lbd_444_8x16_ssse3 Unexecuted instantiation: subsample_lbd_444_16x8_ssse3 Unexecuted instantiation: subsample_lbd_444_16x32_ssse3 Unexecuted instantiation: subsample_lbd_444_32x16_ssse3 Unexecuted instantiation: subsample_lbd_444_4x16_ssse3 Unexecuted instantiation: subsample_lbd_444_16x4_ssse3 Unexecuted instantiation: subsample_lbd_444_8x32_ssse3 Unexecuted instantiation: subsample_lbd_444_32x8_ssse3 Unexecuted instantiation: subsample_hbd_420_4x4_ssse3 Unexecuted instantiation: subsample_hbd_420_8x8_ssse3 Unexecuted instantiation: subsample_hbd_420_16x16_ssse3 Unexecuted instantiation: subsample_hbd_420_32x32_ssse3 Unexecuted instantiation: subsample_hbd_420_4x8_ssse3 Unexecuted instantiation: subsample_hbd_420_8x4_ssse3 Unexecuted instantiation: subsample_hbd_420_8x16_ssse3 Unexecuted instantiation: subsample_hbd_420_16x8_ssse3 Unexecuted instantiation: subsample_hbd_420_16x32_ssse3 Unexecuted instantiation: subsample_hbd_420_32x16_ssse3 Unexecuted instantiation: subsample_hbd_420_4x16_ssse3 Unexecuted instantiation: subsample_hbd_420_16x4_ssse3 Unexecuted instantiation: subsample_hbd_420_8x32_ssse3 Unexecuted instantiation: subsample_hbd_420_32x8_ssse3 Unexecuted instantiation: subsample_hbd_422_4x4_ssse3 Unexecuted instantiation: subsample_hbd_422_8x8_ssse3 Unexecuted instantiation: subsample_hbd_422_16x16_ssse3 Unexecuted instantiation: subsample_hbd_422_32x32_ssse3 Unexecuted instantiation: subsample_hbd_422_4x8_ssse3 Unexecuted instantiation: subsample_hbd_422_8x4_ssse3 Unexecuted instantiation: subsample_hbd_422_8x16_ssse3 Unexecuted instantiation: subsample_hbd_422_16x8_ssse3 Unexecuted instantiation: subsample_hbd_422_16x32_ssse3 Unexecuted instantiation: subsample_hbd_422_32x16_ssse3 Unexecuted instantiation: subsample_hbd_422_4x16_ssse3 Unexecuted instantiation: subsample_hbd_422_16x4_ssse3 Unexecuted instantiation: subsample_hbd_422_8x32_ssse3 Unexecuted instantiation: subsample_hbd_422_32x8_ssse3 Unexecuted instantiation: subsample_hbd_444_4x4_ssse3 Unexecuted instantiation: subsample_hbd_444_8x8_ssse3 Unexecuted instantiation: subsample_hbd_444_16x16_ssse3 Unexecuted instantiation: subsample_hbd_444_32x32_ssse3 Unexecuted instantiation: subsample_hbd_444_4x8_ssse3 Unexecuted instantiation: subsample_hbd_444_8x4_ssse3 Unexecuted instantiation: subsample_hbd_444_8x16_ssse3 Unexecuted instantiation: subsample_hbd_444_16x8_ssse3 Unexecuted instantiation: subsample_hbd_444_16x32_ssse3 Unexecuted instantiation: subsample_hbd_444_32x16_ssse3 Unexecuted instantiation: subsample_hbd_444_4x16_ssse3 Unexecuted instantiation: subsample_hbd_444_16x4_ssse3 Unexecuted instantiation: subsample_hbd_444_8x32_ssse3 Unexecuted instantiation: subsample_hbd_444_32x8_ssse3 |
105 | | |
106 | | // Declare size-specific wrappers for all valid CfL sizes. |
107 | | #define CFL_SUBSAMPLE_FUNCTIONS(arch, sub, bd) \ |
108 | | CFL_SUBSAMPLE(arch, sub, bd, 4, 4) \ |
109 | | CFL_SUBSAMPLE(arch, sub, bd, 8, 8) \ |
110 | | CFL_SUBSAMPLE(arch, sub, bd, 16, 16) \ |
111 | | CFL_SUBSAMPLE(arch, sub, bd, 32, 32) \ |
112 | | CFL_SUBSAMPLE(arch, sub, bd, 4, 8) \ |
113 | | CFL_SUBSAMPLE(arch, sub, bd, 8, 4) \ |
114 | | CFL_SUBSAMPLE(arch, sub, bd, 8, 16) \ |
115 | | CFL_SUBSAMPLE(arch, sub, bd, 16, 8) \ |
116 | | CFL_SUBSAMPLE(arch, sub, bd, 16, 32) \ |
117 | | CFL_SUBSAMPLE(arch, sub, bd, 32, 16) \ |
118 | | CFL_SUBSAMPLE(arch, sub, bd, 4, 16) \ |
119 | | CFL_SUBSAMPLE(arch, sub, bd, 16, 4) \ |
120 | | CFL_SUBSAMPLE(arch, sub, bd, 8, 32) \ |
121 | | CFL_SUBSAMPLE(arch, sub, bd, 32, 8) \ |
122 | | cfl_subsample_##bd##_fn cfl_get_luma_subsampling_##sub##_##bd##_##arch( \ |
123 | 0 | TX_SIZE tx_size) { \ |
124 | 0 | CFL_SUBSAMPLE_FUNCTION_ARRAY(arch, sub, bd) \ |
125 | 0 | return subfn_##sub[tx_size]; \ |
126 | 0 | } Unexecuted instantiation: cfl_get_luma_subsampling_420_lbd_c Unexecuted instantiation: cfl_get_luma_subsampling_422_lbd_c Unexecuted instantiation: cfl_get_luma_subsampling_444_lbd_c Unexecuted instantiation: cfl_get_luma_subsampling_420_hbd_c Unexecuted instantiation: cfl_get_luma_subsampling_422_hbd_c Unexecuted instantiation: cfl_get_luma_subsampling_444_hbd_c Unexecuted instantiation: cfl_get_luma_subsampling_420_lbd_ssse3 Unexecuted instantiation: cfl_get_luma_subsampling_422_lbd_ssse3 Unexecuted instantiation: cfl_get_luma_subsampling_444_lbd_ssse3 Unexecuted instantiation: cfl_get_luma_subsampling_420_hbd_ssse3 Unexecuted instantiation: cfl_get_luma_subsampling_422_hbd_ssse3 Unexecuted instantiation: cfl_get_luma_subsampling_444_hbd_ssse3 |
127 | | |
128 | | // Declare an architecture-specific array of function pointers for size-specific |
129 | | // wrappers. |
130 | | #define CFL_SUBSAMPLE_FUNCTION_ARRAY(arch, sub, bd) \ |
131 | 0 | static const cfl_subsample_##bd##_fn subfn_##sub[TX_SIZES_ALL] = { \ |
132 | 0 | subsample_##bd##_##sub##_4x4_##arch, /* 4x4 */ \ |
133 | 0 | subsample_##bd##_##sub##_8x8_##arch, /* 8x8 */ \ |
134 | 0 | subsample_##bd##_##sub##_16x16_##arch, /* 16x16 */ \ |
135 | 0 | subsample_##bd##_##sub##_32x32_##arch, /* 32x32 */ \ |
136 | 0 | cfl_subsample_##bd##_null, /* 64x64 (invalid CFL size) */ \ |
137 | 0 | subsample_##bd##_##sub##_4x8_##arch, /* 4x8 */ \ |
138 | 0 | subsample_##bd##_##sub##_8x4_##arch, /* 8x4 */ \ |
139 | 0 | subsample_##bd##_##sub##_8x16_##arch, /* 8x16 */ \ |
140 | 0 | subsample_##bd##_##sub##_16x8_##arch, /* 16x8 */ \ |
141 | 0 | subsample_##bd##_##sub##_16x32_##arch, /* 16x32 */ \ |
142 | 0 | subsample_##bd##_##sub##_32x16_##arch, /* 32x16 */ \ |
143 | 0 | cfl_subsample_##bd##_null, /* 32x64 (invalid CFL size) */ \ |
144 | 0 | cfl_subsample_##bd##_null, /* 64x32 (invalid CFL size) */ \ |
145 | 0 | subsample_##bd##_##sub##_4x16_##arch, /* 4x16 */ \ |
146 | 0 | subsample_##bd##_##sub##_16x4_##arch, /* 16x4 */ \ |
147 | 0 | subsample_##bd##_##sub##_8x32_##arch, /* 8x32 */ \ |
148 | 0 | subsample_##bd##_##sub##_32x8_##arch, /* 32x8 */ \ |
149 | 0 | cfl_subsample_##bd##_null, /* 16x64 (invalid CFL size) */ \ |
150 | 0 | cfl_subsample_##bd##_null, /* 64x16 (invalid CFL size) */ \ |
151 | 0 | }; |
152 | | |
153 | | // The RTCD script does not support passing in an array, so we wrap it in this |
154 | | // function. |
155 | | #define CFL_GET_SUBSAMPLE_FUNCTION(arch) \ |
156 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 420, lbd) \ |
157 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 422, lbd) \ |
158 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 444, lbd) \ |
159 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 420, hbd) \ |
160 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 422, hbd) \ |
161 | | CFL_SUBSAMPLE_FUNCTIONS(arch, 444, hbd) |
162 | | |
163 | | // Null function used for invalid tx_sizes |
164 | | static INLINE void cfl_subtract_average_null(const uint16_t *src, |
165 | 0 | int16_t *dst) { |
166 | 0 | (void)dst; |
167 | 0 | (void)src; |
168 | 0 | assert(0); |
169 | 0 | } Unexecuted instantiation: cfl.c:cfl_subtract_average_null Unexecuted instantiation: reconintra.c:cfl_subtract_average_null Unexecuted instantiation: cfl_avx2.c:cfl_subtract_average_null Unexecuted instantiation: cfl_sse2.c:cfl_subtract_average_null Unexecuted instantiation: cfl_ssse3.c:cfl_subtract_average_null Unexecuted instantiation: decodeframe.c:cfl_subtract_average_null Unexecuted instantiation: decodemv.c:cfl_subtract_average_null |
170 | | |
171 | | // Declare a size-specific wrapper for the size-generic function. The compiler |
172 | | // will inline the size generic function in here, the advantage is that the size |
173 | | // will be constant allowing for loop unrolling and other constant propagated |
174 | | // goodness. |
175 | | #define CFL_SUB_AVG_X(arch, width, height, round_offset, num_pel_log2) \ |
176 | | void subtract_average_##width##x##height##_##arch(const uint16_t *src, \ |
177 | 0 | int16_t *dst) { \ |
178 | 0 | subtract_average_##arch(src, dst, width, height, round_offset, \ |
179 | 0 | num_pel_log2); \ |
180 | 0 | } Unexecuted instantiation: subtract_average_4x4_c Unexecuted instantiation: subtract_average_4x8_c Unexecuted instantiation: subtract_average_4x16_c Unexecuted instantiation: subtract_average_8x4_c Unexecuted instantiation: subtract_average_8x8_c Unexecuted instantiation: subtract_average_8x16_c Unexecuted instantiation: subtract_average_8x32_c Unexecuted instantiation: subtract_average_16x4_c Unexecuted instantiation: subtract_average_16x8_c Unexecuted instantiation: subtract_average_16x16_c Unexecuted instantiation: subtract_average_16x32_c Unexecuted instantiation: subtract_average_32x8_c Unexecuted instantiation: subtract_average_32x16_c Unexecuted instantiation: subtract_average_32x32_c Unexecuted instantiation: subtract_average_16x4_avx2 Unexecuted instantiation: subtract_average_16x8_avx2 Unexecuted instantiation: subtract_average_16x16_avx2 Unexecuted instantiation: subtract_average_16x32_avx2 Unexecuted instantiation: subtract_average_32x8_avx2 Unexecuted instantiation: subtract_average_32x16_avx2 Unexecuted instantiation: subtract_average_32x32_avx2 Unexecuted instantiation: subtract_average_4x4_sse2 Unexecuted instantiation: subtract_average_4x8_sse2 Unexecuted instantiation: subtract_average_4x16_sse2 Unexecuted instantiation: subtract_average_8x4_sse2 Unexecuted instantiation: subtract_average_8x8_sse2 Unexecuted instantiation: subtract_average_8x16_sse2 Unexecuted instantiation: subtract_average_8x32_sse2 Unexecuted instantiation: subtract_average_16x4_sse2 Unexecuted instantiation: subtract_average_16x8_sse2 Unexecuted instantiation: subtract_average_16x16_sse2 Unexecuted instantiation: subtract_average_16x32_sse2 Unexecuted instantiation: subtract_average_32x8_sse2 Unexecuted instantiation: subtract_average_32x16_sse2 Unexecuted instantiation: subtract_average_32x32_sse2 |
181 | | |
182 | | // Declare size-specific wrappers for all valid CfL sizes. |
183 | | #define CFL_SUB_AVG_FN(arch) \ |
184 | | CFL_SUB_AVG_X(arch, 4, 4, 8, 4) \ |
185 | | CFL_SUB_AVG_X(arch, 4, 8, 16, 5) \ |
186 | | CFL_SUB_AVG_X(arch, 4, 16, 32, 6) \ |
187 | | CFL_SUB_AVG_X(arch, 8, 4, 16, 5) \ |
188 | | CFL_SUB_AVG_X(arch, 8, 8, 32, 6) \ |
189 | | CFL_SUB_AVG_X(arch, 8, 16, 64, 7) \ |
190 | | CFL_SUB_AVG_X(arch, 8, 32, 128, 8) \ |
191 | | CFL_SUB_AVG_X(arch, 16, 4, 32, 6) \ |
192 | | CFL_SUB_AVG_X(arch, 16, 8, 64, 7) \ |
193 | | CFL_SUB_AVG_X(arch, 16, 16, 128, 8) \ |
194 | | CFL_SUB_AVG_X(arch, 16, 32, 256, 9) \ |
195 | | CFL_SUB_AVG_X(arch, 32, 8, 128, 8) \ |
196 | | CFL_SUB_AVG_X(arch, 32, 16, 256, 9) \ |
197 | | CFL_SUB_AVG_X(arch, 32, 32, 512, 10) \ |
198 | 0 | cfl_subtract_average_fn get_subtract_average_fn_##arch(TX_SIZE tx_size) { \ |
199 | 0 | static const cfl_subtract_average_fn sub_avg[TX_SIZES_ALL] = { \ |
200 | 0 | subtract_average_4x4_##arch, /* 4x4 */ \ |
201 | 0 | subtract_average_8x8_##arch, /* 8x8 */ \ |
202 | 0 | subtract_average_16x16_##arch, /* 16x16 */ \ |
203 | 0 | subtract_average_32x32_##arch, /* 32x32 */ \ |
204 | 0 | cfl_subtract_average_null, /* 64x64 (invalid CFL size) */ \ |
205 | 0 | subtract_average_4x8_##arch, /* 4x8 */ \ |
206 | 0 | subtract_average_8x4_##arch, /* 8x4 */ \ |
207 | 0 | subtract_average_8x16_##arch, /* 8x16 */ \ |
208 | 0 | subtract_average_16x8_##arch, /* 16x8 */ \ |
209 | 0 | subtract_average_16x32_##arch, /* 16x32 */ \ |
210 | 0 | subtract_average_32x16_##arch, /* 32x16 */ \ |
211 | 0 | cfl_subtract_average_null, /* 32x64 (invalid CFL size) */ \ |
212 | 0 | cfl_subtract_average_null, /* 64x32 (invalid CFL size) */ \ |
213 | 0 | subtract_average_4x16_##arch, /* 4x16 (invalid CFL size) */ \ |
214 | 0 | subtract_average_16x4_##arch, /* 16x4 (invalid CFL size) */ \ |
215 | 0 | subtract_average_8x32_##arch, /* 8x32 (invalid CFL size) */ \ |
216 | 0 | subtract_average_32x8_##arch, /* 32x8 (invalid CFL size) */ \ |
217 | 0 | cfl_subtract_average_null, /* 16x64 (invalid CFL size) */ \ |
218 | 0 | cfl_subtract_average_null, /* 64x16 (invalid CFL size) */ \ |
219 | 0 | }; \ |
220 | 0 | /* Modulo TX_SIZES_ALL to ensure that an attacker won't be able to */ \ |
221 | 0 | /* index the function pointer array out of bounds. */ \ |
222 | 0 | return sub_avg[tx_size % TX_SIZES_ALL]; \ |
223 | 0 | } Unexecuted instantiation: get_subtract_average_fn_c Unexecuted instantiation: get_subtract_average_fn_sse2 |
224 | | |
225 | | // For VSX SIMD optimization, the C versions of width == 4 subtract are |
226 | | // faster than the VSX. As such, the VSX code calls the C versions. |
227 | | void subtract_average_4x4_c(const uint16_t *src, int16_t *dst); |
228 | | void subtract_average_4x8_c(const uint16_t *src, int16_t *dst); |
229 | | void subtract_average_4x16_c(const uint16_t *src, int16_t *dst); |
230 | | |
231 | | #define CFL_PREDICT_lbd(arch, width, height) \ |
232 | | void predict_lbd_##width##x##height##_##arch(const int16_t *pred_buf_q3, \ |
233 | | uint8_t *dst, int dst_stride, \ |
234 | 0 | int alpha_q3) { \ |
235 | 0 | cfl_predict_lbd_##arch(pred_buf_q3, dst, dst_stride, alpha_q3, width, \ |
236 | 0 | height); \ |
237 | 0 | } Unexecuted instantiation: predict_lbd_4x4_c Unexecuted instantiation: predict_lbd_4x8_c Unexecuted instantiation: predict_lbd_4x16_c Unexecuted instantiation: predict_lbd_8x4_c Unexecuted instantiation: predict_lbd_8x8_c Unexecuted instantiation: predict_lbd_8x16_c Unexecuted instantiation: predict_lbd_8x32_c Unexecuted instantiation: predict_lbd_16x4_c Unexecuted instantiation: predict_lbd_16x8_c Unexecuted instantiation: predict_lbd_16x16_c Unexecuted instantiation: predict_lbd_16x32_c Unexecuted instantiation: predict_lbd_32x8_c Unexecuted instantiation: predict_lbd_32x16_c Unexecuted instantiation: predict_lbd_32x32_c Unexecuted instantiation: predict_lbd_32x8_avx2 Unexecuted instantiation: predict_lbd_32x16_avx2 Unexecuted instantiation: predict_lbd_32x32_avx2 Unexecuted instantiation: predict_lbd_4x4_ssse3 Unexecuted instantiation: predict_lbd_4x8_ssse3 Unexecuted instantiation: predict_lbd_4x16_ssse3 Unexecuted instantiation: predict_lbd_8x4_ssse3 Unexecuted instantiation: predict_lbd_8x8_ssse3 Unexecuted instantiation: predict_lbd_8x16_ssse3 Unexecuted instantiation: predict_lbd_8x32_ssse3 Unexecuted instantiation: predict_lbd_16x4_ssse3 Unexecuted instantiation: predict_lbd_16x8_ssse3 Unexecuted instantiation: predict_lbd_16x16_ssse3 Unexecuted instantiation: predict_lbd_16x32_ssse3 Unexecuted instantiation: predict_lbd_32x8_ssse3 Unexecuted instantiation: predict_lbd_32x16_ssse3 Unexecuted instantiation: predict_lbd_32x32_ssse3 |
238 | | |
239 | | #define CFL_PREDICT_hbd(arch, width, height) \ |
240 | | void predict_hbd_##width##x##height##_##arch(const int16_t *pred_buf_q3, \ |
241 | | uint16_t *dst, int dst_stride, \ |
242 | 0 | int alpha_q3, int bd) { \ |
243 | 0 | cfl_predict_hbd_##arch(pred_buf_q3, dst, dst_stride, alpha_q3, bd, width, \ |
244 | 0 | height); \ |
245 | 0 | } Unexecuted instantiation: predict_hbd_4x4_c Unexecuted instantiation: predict_hbd_4x8_c Unexecuted instantiation: predict_hbd_4x16_c Unexecuted instantiation: predict_hbd_8x4_c Unexecuted instantiation: predict_hbd_8x8_c Unexecuted instantiation: predict_hbd_8x16_c Unexecuted instantiation: predict_hbd_8x32_c Unexecuted instantiation: predict_hbd_16x4_c Unexecuted instantiation: predict_hbd_16x8_c Unexecuted instantiation: predict_hbd_16x16_c Unexecuted instantiation: predict_hbd_16x32_c Unexecuted instantiation: predict_hbd_32x8_c Unexecuted instantiation: predict_hbd_32x16_c Unexecuted instantiation: predict_hbd_32x32_c Unexecuted instantiation: predict_hbd_16x4_avx2 Unexecuted instantiation: predict_hbd_16x8_avx2 Unexecuted instantiation: predict_hbd_16x16_avx2 Unexecuted instantiation: predict_hbd_16x32_avx2 Unexecuted instantiation: predict_hbd_32x8_avx2 Unexecuted instantiation: predict_hbd_32x16_avx2 Unexecuted instantiation: predict_hbd_32x32_avx2 Unexecuted instantiation: predict_hbd_4x4_ssse3 Unexecuted instantiation: predict_hbd_4x8_ssse3 Unexecuted instantiation: predict_hbd_4x16_ssse3 Unexecuted instantiation: predict_hbd_8x4_ssse3 Unexecuted instantiation: predict_hbd_8x8_ssse3 Unexecuted instantiation: predict_hbd_8x16_ssse3 Unexecuted instantiation: predict_hbd_8x32_ssse3 Unexecuted instantiation: predict_hbd_16x4_ssse3 Unexecuted instantiation: predict_hbd_16x8_ssse3 Unexecuted instantiation: predict_hbd_16x16_ssse3 Unexecuted instantiation: predict_hbd_16x32_ssse3 Unexecuted instantiation: predict_hbd_32x8_ssse3 Unexecuted instantiation: predict_hbd_32x16_ssse3 Unexecuted instantiation: predict_hbd_32x32_ssse3 |
246 | | |
247 | | // This wrapper exists because clang format does not like calling macros with |
248 | | // lowercase letters. |
249 | | #define CFL_PREDICT_X(arch, width, height, bd) \ |
250 | | CFL_PREDICT_##bd(arch, width, height) |
251 | | |
252 | | // Null function used for invalid tx_sizes |
253 | | void cfl_predict_lbd_null(const int16_t *pred_buf_q3, uint8_t *dst, |
254 | | int dst_stride, int alpha_q3); |
255 | | |
256 | | // Null function used for invalid tx_sizes |
257 | | void cfl_predict_hbd_null(const int16_t *pred_buf_q3, uint16_t *dst, |
258 | | int dst_stride, int alpha_q3, int bd); |
259 | | |
260 | | #define CFL_PREDICT_FN(arch, bd) \ |
261 | | CFL_PREDICT_X(arch, 4, 4, bd) \ |
262 | | CFL_PREDICT_X(arch, 4, 8, bd) \ |
263 | | CFL_PREDICT_X(arch, 4, 16, bd) \ |
264 | | CFL_PREDICT_X(arch, 8, 4, bd) \ |
265 | | CFL_PREDICT_X(arch, 8, 8, bd) \ |
266 | | CFL_PREDICT_X(arch, 8, 16, bd) \ |
267 | | CFL_PREDICT_X(arch, 8, 32, bd) \ |
268 | | CFL_PREDICT_X(arch, 16, 4, bd) \ |
269 | | CFL_PREDICT_X(arch, 16, 8, bd) \ |
270 | | CFL_PREDICT_X(arch, 16, 16, bd) \ |
271 | | CFL_PREDICT_X(arch, 16, 32, bd) \ |
272 | | CFL_PREDICT_X(arch, 32, 8, bd) \ |
273 | | CFL_PREDICT_X(arch, 32, 16, bd) \ |
274 | | CFL_PREDICT_X(arch, 32, 32, bd) \ |
275 | 0 | cfl_predict_##bd##_fn get_predict_##bd##_fn_##arch(TX_SIZE tx_size) { \ |
276 | 0 | static const cfl_predict_##bd##_fn pred[TX_SIZES_ALL] = { \ |
277 | 0 | predict_##bd##_4x4_##arch, /* 4x4 */ \ |
278 | 0 | predict_##bd##_8x8_##arch, /* 8x8 */ \ |
279 | 0 | predict_##bd##_16x16_##arch, /* 16x16 */ \ |
280 | 0 | predict_##bd##_32x32_##arch, /* 32x32 */ \ |
281 | 0 | cfl_predict_##bd##_null, /* 64x64 (invalid CFL size) */ \ |
282 | 0 | predict_##bd##_4x8_##arch, /* 4x8 */ \ |
283 | 0 | predict_##bd##_8x4_##arch, /* 8x4 */ \ |
284 | 0 | predict_##bd##_8x16_##arch, /* 8x16 */ \ |
285 | 0 | predict_##bd##_16x8_##arch, /* 16x8 */ \ |
286 | 0 | predict_##bd##_16x32_##arch, /* 16x32 */ \ |
287 | 0 | predict_##bd##_32x16_##arch, /* 32x16 */ \ |
288 | 0 | cfl_predict_##bd##_null, /* 32x64 (invalid CFL size) */ \ |
289 | 0 | cfl_predict_##bd##_null, /* 64x32 (invalid CFL size) */ \ |
290 | 0 | predict_##bd##_4x16_##arch, /* 4x16 */ \ |
291 | 0 | predict_##bd##_16x4_##arch, /* 16x4 */ \ |
292 | 0 | predict_##bd##_8x32_##arch, /* 8x32 */ \ |
293 | 0 | predict_##bd##_32x8_##arch, /* 32x8 */ \ |
294 | 0 | cfl_predict_##bd##_null, /* 16x64 (invalid CFL size) */ \ |
295 | 0 | cfl_predict_##bd##_null, /* 64x16 (invalid CFL size) */ \ |
296 | 0 | }; \ |
297 | 0 | /* Modulo TX_SIZES_ALL to ensure that an attacker won't be able to */ \ |
298 | 0 | /* index the function pointer array out of bounds. */ \ |
299 | 0 | return pred[tx_size % TX_SIZES_ALL]; \ |
300 | 0 | } Unexecuted instantiation: get_predict_lbd_fn_c Unexecuted instantiation: get_predict_hbd_fn_c Unexecuted instantiation: get_predict_lbd_fn_ssse3 Unexecuted instantiation: get_predict_hbd_fn_ssse3 |
301 | | |
302 | | #endif // AOM_AV1_COMMON_CFL_H_ |