/src/libjxl/lib/jxl/enc_entropy_coder.cc
Line | Count | Source (jump to first uncovered line) |
1 | | // Copyright (c) the JPEG XL Project Authors. All rights reserved. |
2 | | // |
3 | | // Use of this source code is governed by a BSD-style |
4 | | // license that can be found in the LICENSE file. |
5 | | |
6 | | #include "lib/jxl/enc_entropy_coder.h" |
7 | | |
8 | | #include <cstddef> |
9 | | #include <cstdint> |
10 | | #include <vector> |
11 | | |
12 | | #include "lib/jxl/base/compiler_specific.h" // ssize_t |
13 | | #include "lib/jxl/base/rect.h" |
14 | | #include "lib/jxl/enc_ans.h" |
15 | | #include "lib/jxl/frame_dimensions.h" |
16 | | #include "lib/jxl/frame_header.h" |
17 | | |
18 | | #undef HWY_TARGET_INCLUDE |
19 | | #define HWY_TARGET_INCLUDE "lib/jxl/enc_entropy_coder.cc" |
20 | | #include <hwy/foreach_target.h> |
21 | | #include <hwy/highway.h> |
22 | | |
23 | | #include "lib/jxl/ac_context.h" |
24 | | #include "lib/jxl/ac_strategy.h" |
25 | | #include "lib/jxl/base/bits.h" |
26 | | #include "lib/jxl/base/compiler_specific.h" |
27 | | #include "lib/jxl/base/status.h" |
28 | | #include "lib/jxl/coeff_order.h" |
29 | | #include "lib/jxl/coeff_order_fwd.h" |
30 | | #include "lib/jxl/entropy_coder.h" |
31 | | #include "lib/jxl/image.h" |
32 | | #include "lib/jxl/pack_signed.h" |
33 | | |
34 | | HWY_BEFORE_NAMESPACE(); |
35 | | namespace jxl { |
36 | | namespace HWY_NAMESPACE { |
37 | | |
38 | | // These templates are not found via ADL. |
39 | | using hwy::HWY_NAMESPACE::Add; |
40 | | using hwy::HWY_NAMESPACE::AndNot; |
41 | | using hwy::HWY_NAMESPACE::Eq; |
42 | | using hwy::HWY_NAMESPACE::GetLane; |
43 | | |
44 | | // Returns number of non-zero coefficients (but skip LLF). |
45 | | // We cannot rely on block[] being all-zero bits, so first truncate to integer. |
46 | | // Also writes the per-8x8 block nzeros starting at nzeros_pos. |
47 | | int32_t NumNonZeroExceptLLF(const size_t cx, const size_t cy, |
48 | | const AcStrategy acs, const size_t covered_blocks, |
49 | | const size_t log2_covered_blocks, |
50 | | const int32_t* JXL_RESTRICT block, |
51 | | const size_t nzeros_stride, |
52 | 97.9k | int32_t* JXL_RESTRICT nzeros_pos) { |
53 | 97.9k | const HWY_CAPPED(int32_t, kBlockDim) di; |
54 | | |
55 | 97.9k | const auto zero = Zero(di); |
56 | | // Add FF..FF for every zero coefficient, negate to get #zeros. |
57 | 97.9k | auto neg_sum_zero = zero; |
58 | | |
59 | 97.9k | { |
60 | | // Mask sufficient for one row of coefficients. |
61 | 97.9k | HWY_ALIGN const int32_t |
62 | 97.9k | llf_mask_lanes[AcStrategy::kMaxCoeffBlocks * (1 + kBlockDim)] = { |
63 | 97.9k | -1, -1, -1, -1}; |
64 | | // First cx=1,2,4 elements are FF..FF, others 0. |
65 | 97.9k | const int32_t* llf_mask_pos = |
66 | 97.9k | llf_mask_lanes + AcStrategy::kMaxCoeffBlocks - cx; |
67 | | |
68 | | // Rows with LLF: mask out the LLF |
69 | 274k | for (size_t y = 0; y < cy; y++) { |
70 | 743k | for (size_t x = 0; x < cx * kBlockDim; x += Lanes(di)) { |
71 | 566k | const auto llf_mask = LoadU(di, llf_mask_pos + x); |
72 | | |
73 | | // LLF counts as zero so we don't include it in nzeros. |
74 | 566k | const auto coef = |
75 | 566k | AndNot(llf_mask, Load(di, &block[y * cx * kBlockDim + x])); |
76 | | |
77 | 566k | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); |
78 | 566k | } |
79 | 176k | } |
80 | 97.9k | } |
81 | | |
82 | | // Remaining rows: no mask |
83 | 1.33M | for (size_t y = cy; y < cy * kBlockDim; y++) { |
84 | 5.20M | for (size_t x = 0; x < cx * kBlockDim; x += Lanes(di)) { |
85 | 3.96M | const auto coef = Load(di, &block[y * cx * kBlockDim + x]); |
86 | 3.96M | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); |
87 | 3.96M | } |
88 | 1.23M | } |
89 | | |
90 | | // We want area - sum_zero, add because neg_sum_zero is already negated. |
91 | 97.9k | const int32_t nzeros = static_cast<int32_t>(cx * cy * kDCTBlockSize) + |
92 | 97.9k | GetLane(SumOfLanes(di, neg_sum_zero)); |
93 | | |
94 | 97.9k | const int32_t shifted_nzeros = static_cast<int32_t>( |
95 | 97.9k | (nzeros + covered_blocks - 1) >> log2_covered_blocks); |
96 | | // Need non-canonicalized dimensions! |
97 | 310k | for (size_t y = 0; y < acs.covered_blocks_y(); y++) { |
98 | 779k | for (size_t x = 0; x < acs.covered_blocks_x(); x++) { |
99 | 566k | nzeros_pos[x + y * nzeros_stride] = shifted_nzeros; |
100 | 566k | } |
101 | 212k | } |
102 | | |
103 | 97.9k | return nzeros; |
104 | 97.9k | } Unexecuted instantiation: jxl::N_SSE4::NumNonZeroExceptLLF(unsigned long, unsigned long, jxl::AcStrategy, unsigned long, unsigned long, int const*, unsigned long, int*) jxl::N_AVX2::NumNonZeroExceptLLF(unsigned long, unsigned long, jxl::AcStrategy, unsigned long, unsigned long, int const*, unsigned long, int*) Line | Count | Source | 52 | 97.9k | int32_t* JXL_RESTRICT nzeros_pos) { | 53 | 97.9k | const HWY_CAPPED(int32_t, kBlockDim) di; | 54 | | | 55 | 97.9k | const auto zero = Zero(di); | 56 | | // Add FF..FF for every zero coefficient, negate to get #zeros. | 57 | 97.9k | auto neg_sum_zero = zero; | 58 | | | 59 | 97.9k | { | 60 | | // Mask sufficient for one row of coefficients. | 61 | 97.9k | HWY_ALIGN const int32_t | 62 | 97.9k | llf_mask_lanes[AcStrategy::kMaxCoeffBlocks * (1 + kBlockDim)] = { | 63 | 97.9k | -1, -1, -1, -1}; | 64 | | // First cx=1,2,4 elements are FF..FF, others 0. | 65 | 97.9k | const int32_t* llf_mask_pos = | 66 | 97.9k | llf_mask_lanes + AcStrategy::kMaxCoeffBlocks - cx; | 67 | | | 68 | | // Rows with LLF: mask out the LLF | 69 | 274k | for (size_t y = 0; y < cy; y++) { | 70 | 743k | for (size_t x = 0; x < cx * kBlockDim; x += Lanes(di)) { | 71 | 566k | const auto llf_mask = LoadU(di, llf_mask_pos + x); | 72 | | | 73 | | // LLF counts as zero so we don't include it in nzeros. | 74 | 566k | const auto coef = | 75 | 566k | AndNot(llf_mask, Load(di, &block[y * cx * kBlockDim + x])); | 76 | | | 77 | 566k | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); | 78 | 566k | } | 79 | 176k | } | 80 | 97.9k | } | 81 | | | 82 | | // Remaining rows: no mask | 83 | 1.33M | for (size_t y = cy; y < cy * kBlockDim; y++) { | 84 | 5.20M | for (size_t x = 0; x < cx * kBlockDim; x += Lanes(di)) { | 85 | 3.96M | const auto coef = Load(di, &block[y * cx * kBlockDim + x]); | 86 | 3.96M | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); | 87 | 3.96M | } | 88 | 1.23M | } | 89 | | | 90 | | // We want area - sum_zero, add because neg_sum_zero is already negated. | 91 | 97.9k | const int32_t nzeros = static_cast<int32_t>(cx * cy * kDCTBlockSize) + | 92 | 97.9k | GetLane(SumOfLanes(di, neg_sum_zero)); | 93 | | | 94 | 97.9k | const int32_t shifted_nzeros = static_cast<int32_t>( | 95 | 97.9k | (nzeros + covered_blocks - 1) >> log2_covered_blocks); | 96 | | // Need non-canonicalized dimensions! | 97 | 310k | for (size_t y = 0; y < acs.covered_blocks_y(); y++) { | 98 | 779k | for (size_t x = 0; x < acs.covered_blocks_x(); x++) { | 99 | 566k | nzeros_pos[x + y * nzeros_stride] = shifted_nzeros; | 100 | 566k | } | 101 | 212k | } | 102 | | | 103 | 97.9k | return nzeros; | 104 | 97.9k | } |
Unexecuted instantiation: jxl::N_SSE2::NumNonZeroExceptLLF(unsigned long, unsigned long, jxl::AcStrategy, unsigned long, unsigned long, int const*, unsigned long, int*) |
105 | | |
106 | | // Specialization for 8x8, where only top-left is LLF/DC. |
107 | | // About 1% overall speedup vs. NumNonZeroExceptLLF. |
108 | | int32_t NumNonZero8x8ExceptDC(const int32_t* JXL_RESTRICT block, |
109 | 627k | int32_t* JXL_RESTRICT nzeros_pos) { |
110 | 627k | const HWY_CAPPED(int32_t, kBlockDim) di; |
111 | | |
112 | 627k | const auto zero = Zero(di); |
113 | | // Add FF..FF for every zero coefficient, negate to get #zeros. |
114 | 627k | auto neg_sum_zero = zero; |
115 | | |
116 | 627k | { |
117 | | // First row has DC, so mask |
118 | 627k | const size_t y = 0; |
119 | 627k | HWY_ALIGN const int32_t dc_mask_lanes[kBlockDim] = {-1}; |
120 | | |
121 | 1.25M | for (size_t x = 0; x < kBlockDim; x += Lanes(di)) { |
122 | 627k | const auto dc_mask = Load(di, dc_mask_lanes + x); |
123 | | |
124 | | // DC counts as zero so we don't include it in nzeros. |
125 | 627k | const auto coef = AndNot(dc_mask, Load(di, &block[y * kBlockDim + x])); |
126 | | |
127 | 627k | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); |
128 | 627k | } |
129 | 627k | } |
130 | | |
131 | | // Remaining rows: no mask |
132 | 5.01M | for (size_t y = 1; y < kBlockDim; y++) { |
133 | 8.78M | for (size_t x = 0; x < kBlockDim; x += Lanes(di)) { |
134 | 4.39M | const auto coef = Load(di, &block[y * kBlockDim + x]); |
135 | 4.39M | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); |
136 | 4.39M | } |
137 | 4.39M | } |
138 | | |
139 | | // We want 64 - sum_zero, add because neg_sum_zero is already negated. |
140 | 627k | const int32_t nzeros = static_cast<int32_t>(kDCTBlockSize) + |
141 | 627k | GetLane(SumOfLanes(di, neg_sum_zero)); |
142 | | |
143 | 627k | *nzeros_pos = nzeros; |
144 | | |
145 | 627k | return nzeros; |
146 | 627k | } Unexecuted instantiation: jxl::N_SSE4::NumNonZero8x8ExceptDC(int const*, int*) jxl::N_AVX2::NumNonZero8x8ExceptDC(int const*, int*) Line | Count | Source | 109 | 627k | int32_t* JXL_RESTRICT nzeros_pos) { | 110 | 627k | const HWY_CAPPED(int32_t, kBlockDim) di; | 111 | | | 112 | 627k | const auto zero = Zero(di); | 113 | | // Add FF..FF for every zero coefficient, negate to get #zeros. | 114 | 627k | auto neg_sum_zero = zero; | 115 | | | 116 | 627k | { | 117 | | // First row has DC, so mask | 118 | 627k | const size_t y = 0; | 119 | 627k | HWY_ALIGN const int32_t dc_mask_lanes[kBlockDim] = {-1}; | 120 | | | 121 | 1.25M | for (size_t x = 0; x < kBlockDim; x += Lanes(di)) { | 122 | 627k | const auto dc_mask = Load(di, dc_mask_lanes + x); | 123 | | | 124 | | // DC counts as zero so we don't include it in nzeros. | 125 | 627k | const auto coef = AndNot(dc_mask, Load(di, &block[y * kBlockDim + x])); | 126 | | | 127 | 627k | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); | 128 | 627k | } | 129 | 627k | } | 130 | | | 131 | | // Remaining rows: no mask | 132 | 5.01M | for (size_t y = 1; y < kBlockDim; y++) { | 133 | 8.78M | for (size_t x = 0; x < kBlockDim; x += Lanes(di)) { | 134 | 4.39M | const auto coef = Load(di, &block[y * kBlockDim + x]); | 135 | 4.39M | neg_sum_zero = Add(neg_sum_zero, VecFromMask(di, Eq(coef, zero))); | 136 | 4.39M | } | 137 | 4.39M | } | 138 | | | 139 | | // We want 64 - sum_zero, add because neg_sum_zero is already negated. | 140 | 627k | const int32_t nzeros = static_cast<int32_t>(kDCTBlockSize) + | 141 | 627k | GetLane(SumOfLanes(di, neg_sum_zero)); | 142 | | | 143 | 627k | *nzeros_pos = nzeros; | 144 | | | 145 | 627k | return nzeros; | 146 | 627k | } |
Unexecuted instantiation: jxl::N_SSE2::NumNonZero8x8ExceptDC(int const*, int*) |
147 | | |
148 | | // The number of nonzeros of each block is predicted from the top and the left |
149 | | // blocks, with opportune scaling to take into account the number of blocks of |
150 | | // each strategy. The predicted number of nonzeros divided by two is used as a |
151 | | // context; if this number is above 63, a specific context is used. If the |
152 | | // number of nonzeros of a strategy is above 63, it is written directly using a |
153 | | // fixed number of bits (that depends on the size of the strategy). |
154 | | Status TokenizeCoefficients(const coeff_order_t* JXL_RESTRICT orders, |
155 | | const Rect& rect, |
156 | | const int32_t* JXL_RESTRICT* JXL_RESTRICT ac_rows, |
157 | | const AcStrategyImage& ac_strategy, |
158 | | const YCbCrChromaSubsampling& cs, |
159 | | Image3I* JXL_RESTRICT tmp_num_nzeroes, |
160 | | std::vector<Token>* JXL_RESTRICT output, |
161 | | const ImageB& qdc, const ImageI& qf, |
162 | 595 | const BlockCtxMap& block_ctx_map) { |
163 | 595 | const size_t xsize_blocks = rect.xsize(); |
164 | 595 | const size_t ysize_blocks = rect.ysize(); |
165 | 595 | output->clear(); |
166 | | // TODO(user): update the estimate: usually less coefficients are used. |
167 | 595 | output->reserve(3 * xsize_blocks * ysize_blocks * kDCTBlockSize); |
168 | | |
169 | 595 | size_t offset[3] = {}; |
170 | 595 | const size_t nzeros_stride = tmp_num_nzeroes->PixelsPerRow(); |
171 | 15.0k | for (size_t by = 0; by < ysize_blocks; ++by) { |
172 | 14.4k | size_t sby[3] = {by >> cs.VShift(0), by >> cs.VShift(1), |
173 | 14.4k | by >> cs.VShift(2)}; |
174 | 14.4k | int32_t* JXL_RESTRICT row_nzeros[3] = { |
175 | 14.4k | tmp_num_nzeroes->PlaneRow(0, sby[0]), |
176 | 14.4k | tmp_num_nzeroes->PlaneRow(1, sby[1]), |
177 | 14.4k | tmp_num_nzeroes->PlaneRow(2, sby[2]), |
178 | 14.4k | }; |
179 | 14.4k | const int32_t* JXL_RESTRICT row_nzeros_top[3] = { |
180 | 14.4k | sby[0] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(0, sby[0] - 1), |
181 | 14.4k | sby[1] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(1, sby[1] - 1), |
182 | 14.4k | sby[2] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(2, sby[2] - 1), |
183 | 14.4k | }; |
184 | 14.4k | const uint8_t* JXL_RESTRICT row_qdc = |
185 | 14.4k | qdc.ConstRow(rect.y0() + by) + rect.x0(); |
186 | 14.4k | const int32_t* JXL_RESTRICT row_qf = rect.ConstRow(qf, by); |
187 | 14.4k | AcStrategyRow acs_row = ac_strategy.ConstRow(rect, by); |
188 | 412k | for (size_t bx = 0; bx < xsize_blocks; ++bx) { |
189 | 398k | AcStrategy acs = acs_row[bx]; |
190 | 398k | if (!acs.IsFirstBlock()) continue; |
191 | 241k | size_t sbx[3] = {bx >> cs.HShift(0), bx >> cs.HShift(1), |
192 | 241k | bx >> cs.HShift(2)}; |
193 | 241k | size_t cx = acs.covered_blocks_x(); |
194 | 241k | size_t cy = acs.covered_blocks_y(); |
195 | 241k | const size_t covered_blocks = cx * cy; // = #LLF coefficients |
196 | 241k | const size_t log2_covered_blocks = |
197 | 241k | Num0BitsBelowLS1Bit_Nonzero(covered_blocks); |
198 | 241k | const size_t size = covered_blocks * kDCTBlockSize; |
199 | | |
200 | 241k | CoefficientLayout(&cy, &cx); // swap cx/cy to canonical order |
201 | | |
202 | 725k | for (int c : {1, 0, 2}) { |
203 | 725k | if (sbx[c] << cs.HShift(c) != bx) continue; |
204 | 725k | if (sby[c] << cs.VShift(c) != by) continue; |
205 | 725k | const int32_t* JXL_RESTRICT block = ac_rows[c] + offset[c]; |
206 | | |
207 | 725k | int32_t nzeros = |
208 | 725k | (covered_blocks == 1) |
209 | 725k | ? NumNonZero8x8ExceptDC(block, row_nzeros[c] + sbx[c]) |
210 | 725k | : NumNonZeroExceptLLF(cx, cy, acs, covered_blocks, |
211 | 97.9k | log2_covered_blocks, block, nzeros_stride, |
212 | 97.9k | row_nzeros[c] + sbx[c]); |
213 | | |
214 | 725k | int ord = kStrategyOrder[acs.RawStrategy()]; |
215 | 725k | const coeff_order_t* JXL_RESTRICT order = |
216 | 725k | &orders[CoeffOrderOffset(ord, c)]; |
217 | | |
218 | 725k | int32_t predicted_nzeros = |
219 | 725k | PredictFromTopAndLeft(row_nzeros_top[c], row_nzeros[c], sbx[c], 32); |
220 | 725k | size_t block_ctx = |
221 | 725k | block_ctx_map.Context(row_qdc[bx], row_qf[sbx[c]], ord, c); |
222 | 725k | const int32_t nzero_ctx = |
223 | 725k | block_ctx_map.NonZeroContext(predicted_nzeros, block_ctx); |
224 | | |
225 | 725k | output->emplace_back(nzero_ctx, nzeros); |
226 | 725k | const size_t histo_offset = |
227 | 725k | block_ctx_map.ZeroDensityContextsOffset(block_ctx); |
228 | | // Skip LLF. |
229 | 725k | size_t prev = (nzeros > static_cast<ssize_t>(size / 16) ? 0 : 1); |
230 | 26.8M | for (size_t k = covered_blocks; k < size && nzeros != 0; ++k) { |
231 | 26.0M | int32_t coeff = block[order[k]]; |
232 | 26.0M | size_t ctx = |
233 | 26.0M | histo_offset + ZeroDensityContext(nzeros, k, covered_blocks, |
234 | 26.0M | log2_covered_blocks, prev); |
235 | 26.0M | uint32_t u_coeff = PackSigned(coeff); |
236 | 26.0M | output->emplace_back(ctx, u_coeff); |
237 | 26.0M | prev = (coeff != 0) ? 1 : 0; |
238 | 26.0M | nzeros -= prev; |
239 | 26.0M | } |
240 | 725k | JXL_ENSURE(nzeros == 0); |
241 | 725k | offset[c] += size; |
242 | 725k | } |
243 | 241k | } |
244 | 14.4k | } |
245 | 595 | return true; |
246 | 595 | } Unexecuted instantiation: jxl::N_SSE4::TokenizeCoefficients(unsigned int const*, jxl::RectT<unsigned long> const&, int const* restrict*, jxl::AcStrategyImage const&, jxl::YCbCrChromaSubsampling const&, jxl::Image3<int>*, std::__1::vector<jxl::Token, std::__1::allocator<jxl::Token> >*, jxl::Plane<unsigned char> const&, jxl::Plane<int> const&, jxl::BlockCtxMap const&) jxl::N_AVX2::TokenizeCoefficients(unsigned int const*, jxl::RectT<unsigned long> const&, int const* restrict*, jxl::AcStrategyImage const&, jxl::YCbCrChromaSubsampling const&, jxl::Image3<int>*, std::__1::vector<jxl::Token, std::__1::allocator<jxl::Token> >*, jxl::Plane<unsigned char> const&, jxl::Plane<int> const&, jxl::BlockCtxMap const&) Line | Count | Source | 162 | 595 | const BlockCtxMap& block_ctx_map) { | 163 | 595 | const size_t xsize_blocks = rect.xsize(); | 164 | 595 | const size_t ysize_blocks = rect.ysize(); | 165 | 595 | output->clear(); | 166 | | // TODO(user): update the estimate: usually less coefficients are used. | 167 | 595 | output->reserve(3 * xsize_blocks * ysize_blocks * kDCTBlockSize); | 168 | | | 169 | 595 | size_t offset[3] = {}; | 170 | 595 | const size_t nzeros_stride = tmp_num_nzeroes->PixelsPerRow(); | 171 | 15.0k | for (size_t by = 0; by < ysize_blocks; ++by) { | 172 | 14.4k | size_t sby[3] = {by >> cs.VShift(0), by >> cs.VShift(1), | 173 | 14.4k | by >> cs.VShift(2)}; | 174 | 14.4k | int32_t* JXL_RESTRICT row_nzeros[3] = { | 175 | 14.4k | tmp_num_nzeroes->PlaneRow(0, sby[0]), | 176 | 14.4k | tmp_num_nzeroes->PlaneRow(1, sby[1]), | 177 | 14.4k | tmp_num_nzeroes->PlaneRow(2, sby[2]), | 178 | 14.4k | }; | 179 | 14.4k | const int32_t* JXL_RESTRICT row_nzeros_top[3] = { | 180 | 14.4k | sby[0] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(0, sby[0] - 1), | 181 | 14.4k | sby[1] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(1, sby[1] - 1), | 182 | 14.4k | sby[2] == 0 ? nullptr : tmp_num_nzeroes->ConstPlaneRow(2, sby[2] - 1), | 183 | 14.4k | }; | 184 | 14.4k | const uint8_t* JXL_RESTRICT row_qdc = | 185 | 14.4k | qdc.ConstRow(rect.y0() + by) + rect.x0(); | 186 | 14.4k | const int32_t* JXL_RESTRICT row_qf = rect.ConstRow(qf, by); | 187 | 14.4k | AcStrategyRow acs_row = ac_strategy.ConstRow(rect, by); | 188 | 412k | for (size_t bx = 0; bx < xsize_blocks; ++bx) { | 189 | 398k | AcStrategy acs = acs_row[bx]; | 190 | 398k | if (!acs.IsFirstBlock()) continue; | 191 | 241k | size_t sbx[3] = {bx >> cs.HShift(0), bx >> cs.HShift(1), | 192 | 241k | bx >> cs.HShift(2)}; | 193 | 241k | size_t cx = acs.covered_blocks_x(); | 194 | 241k | size_t cy = acs.covered_blocks_y(); | 195 | 241k | const size_t covered_blocks = cx * cy; // = #LLF coefficients | 196 | 241k | const size_t log2_covered_blocks = | 197 | 241k | Num0BitsBelowLS1Bit_Nonzero(covered_blocks); | 198 | 241k | const size_t size = covered_blocks * kDCTBlockSize; | 199 | | | 200 | 241k | CoefficientLayout(&cy, &cx); // swap cx/cy to canonical order | 201 | | | 202 | 725k | for (int c : {1, 0, 2}) { | 203 | 725k | if (sbx[c] << cs.HShift(c) != bx) continue; | 204 | 725k | if (sby[c] << cs.VShift(c) != by) continue; | 205 | 725k | const int32_t* JXL_RESTRICT block = ac_rows[c] + offset[c]; | 206 | | | 207 | 725k | int32_t nzeros = | 208 | 725k | (covered_blocks == 1) | 209 | 725k | ? NumNonZero8x8ExceptDC(block, row_nzeros[c] + sbx[c]) | 210 | 725k | : NumNonZeroExceptLLF(cx, cy, acs, covered_blocks, | 211 | 97.9k | log2_covered_blocks, block, nzeros_stride, | 212 | 97.9k | row_nzeros[c] + sbx[c]); | 213 | | | 214 | 725k | int ord = kStrategyOrder[acs.RawStrategy()]; | 215 | 725k | const coeff_order_t* JXL_RESTRICT order = | 216 | 725k | &orders[CoeffOrderOffset(ord, c)]; | 217 | | | 218 | 725k | int32_t predicted_nzeros = | 219 | 725k | PredictFromTopAndLeft(row_nzeros_top[c], row_nzeros[c], sbx[c], 32); | 220 | 725k | size_t block_ctx = | 221 | 725k | block_ctx_map.Context(row_qdc[bx], row_qf[sbx[c]], ord, c); | 222 | 725k | const int32_t nzero_ctx = | 223 | 725k | block_ctx_map.NonZeroContext(predicted_nzeros, block_ctx); | 224 | | | 225 | 725k | output->emplace_back(nzero_ctx, nzeros); | 226 | 725k | const size_t histo_offset = | 227 | 725k | block_ctx_map.ZeroDensityContextsOffset(block_ctx); | 228 | | // Skip LLF. | 229 | 725k | size_t prev = (nzeros > static_cast<ssize_t>(size / 16) ? 0 : 1); | 230 | 26.8M | for (size_t k = covered_blocks; k < size && nzeros != 0; ++k) { | 231 | 26.0M | int32_t coeff = block[order[k]]; | 232 | 26.0M | size_t ctx = | 233 | 26.0M | histo_offset + ZeroDensityContext(nzeros, k, covered_blocks, | 234 | 26.0M | log2_covered_blocks, prev); | 235 | 26.0M | uint32_t u_coeff = PackSigned(coeff); | 236 | 26.0M | output->emplace_back(ctx, u_coeff); | 237 | 26.0M | prev = (coeff != 0) ? 1 : 0; | 238 | 26.0M | nzeros -= prev; | 239 | 26.0M | } | 240 | 725k | JXL_ENSURE(nzeros == 0); | 241 | 725k | offset[c] += size; | 242 | 725k | } | 243 | 241k | } | 244 | 14.4k | } | 245 | 595 | return true; | 246 | 595 | } |
Unexecuted instantiation: jxl::N_SSE2::TokenizeCoefficients(unsigned int const*, jxl::RectT<unsigned long> const&, int const* restrict*, jxl::AcStrategyImage const&, jxl::YCbCrChromaSubsampling const&, jxl::Image3<int>*, std::__1::vector<jxl::Token, std::__1::allocator<jxl::Token> >*, jxl::Plane<unsigned char> const&, jxl::Plane<int> const&, jxl::BlockCtxMap const&) |
247 | | |
248 | | // NOLINTNEXTLINE(google-readability-namespace-comments) |
249 | | } // namespace HWY_NAMESPACE |
250 | | } // namespace jxl |
251 | | HWY_AFTER_NAMESPACE(); |
252 | | |
253 | | #if HWY_ONCE |
254 | | namespace jxl { |
255 | | HWY_EXPORT(TokenizeCoefficients); |
256 | | Status TokenizeCoefficients(const coeff_order_t* JXL_RESTRICT orders, |
257 | | const Rect& rect, |
258 | | const int32_t* JXL_RESTRICT* JXL_RESTRICT ac_rows, |
259 | | const AcStrategyImage& ac_strategy, |
260 | | const YCbCrChromaSubsampling& cs, |
261 | | Image3I* JXL_RESTRICT tmp_num_nzeroes, |
262 | | std::vector<Token>* JXL_RESTRICT output, |
263 | | const ImageB& qdc, const ImageI& qf, |
264 | 595 | const BlockCtxMap& block_ctx_map) { |
265 | 595 | return HWY_DYNAMIC_DISPATCH(TokenizeCoefficients)( |
266 | 595 | orders, rect, ac_rows, ac_strategy, cs, tmp_num_nzeroes, output, qdc, qf, |
267 | 595 | block_ctx_map); |
268 | 595 | } |
269 | | |
270 | | } // namespace jxl |
271 | | #endif // HWY_ONCE |