/proc/self/cwd/libfaad/sbr_qmf.c
Line | Count | Source |
1 | | /* |
2 | | ** FAAD2 - Freeware Advanced Audio (AAC) Decoder including SBR decoding |
3 | | ** Copyright (C) 2003-2005 M. Bakker, Nero AG, http://www.nero.com |
4 | | ** |
5 | | ** This program is free software; you can redistribute it and/or modify |
6 | | ** it under the terms of the GNU General Public License as published by |
7 | | ** the Free Software Foundation; either version 2 of the License, or |
8 | | ** (at your option) any later version. |
9 | | ** |
10 | | ** This program is distributed in the hope that it will be useful, |
11 | | ** but WITHOUT ANY WARRANTY; without even the implied warranty of |
12 | | ** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
13 | | ** GNU General Public License for more details. |
14 | | ** |
15 | | ** You should have received a copy of the GNU General Public License |
16 | | ** along with this program; if not, write to the Free Software |
17 | | ** Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. |
18 | | ** |
19 | | ** Any non-GPL usage of this software or parts of this software is strictly |
20 | | ** forbidden. |
21 | | ** |
22 | | ** The "appropriate copyright message" mentioned in section 2c of the GPLv2 |
23 | | ** must read: "Code from FAAD2 is copyright (c) Nero AG, www.nero.com" |
24 | | ** |
25 | | ** Commercial non-GPL licensing of this software is possible. |
26 | | ** For more info contact Nero AG through Mpeg4AAClicense@nero.com. |
27 | | ** |
28 | | ** $Id: sbr_qmf.c,v 1.32 2007/11/01 12:33:36 menno Exp $ |
29 | | **/ |
30 | | |
31 | | #include "common.h" |
32 | | #include "structs.h" |
33 | | |
34 | | #ifdef SBR_DEC |
35 | | |
36 | | |
37 | | #include <stdlib.h> |
38 | | #include "sbr_dct.h" |
39 | | #include "sbr_qmf.h" |
40 | | #include "sbr_qmf_c.h" |
41 | | #include "sbr_syntax.h" |
42 | | |
43 | | qmfa_info *qmfa_init(uint8_t channels) |
44 | 425k | { |
45 | 425k | qmfa_info *qmfa = (qmfa_info*)faad_malloc(sizeof(qmfa_info)); |
46 | | |
47 | | /* x is implemented as double ringbuffer */ |
48 | 425k | qmfa->x = (real_t*)faad_malloc(2 * channels * 10 * sizeof(real_t)); |
49 | 425k | memset(qmfa->x, 0, 2 * channels * 10 * sizeof(real_t)); |
50 | | |
51 | | /* ringbuffer index */ |
52 | 425k | qmfa->x_index = 0; |
53 | | |
54 | 425k | qmfa->channels = channels; |
55 | | |
56 | 425k | return qmfa; |
57 | 425k | } |
58 | | |
59 | | void qmfa_end(qmfa_info *qmfa) |
60 | 436k | { |
61 | 436k | if (qmfa) |
62 | 425k | { |
63 | 425k | if (qmfa->x) faad_free(qmfa->x); |
64 | 425k | faad_free(qmfa); |
65 | 425k | } |
66 | 436k | } |
67 | | |
68 | | void sbr_qmf_analysis_32(sbr_info *sbr, qmfa_info *qmfa, const real_t *input, |
69 | | qmf_t X[MAX_NTSRHFG][64], uint8_t offset, uint8_t kx) |
70 | 988k | { |
71 | 988k | ALIGN real_t u[64]; |
72 | 988k | #ifndef SBR_LOW_POWER |
73 | 988k | ALIGN real_t in_real[32], in_imag[32], out_real[32], out_imag[32]; |
74 | | #else |
75 | | ALIGN real_t y[32]; |
76 | | #endif |
77 | 988k | uint32_t in = 0; |
78 | 988k | uint8_t l; |
79 | | |
80 | | /* qmf subsample l */ |
81 | 32.2M | for (l = 0; l < sbr->numTimeSlotsRate; l++) |
82 | 31.2M | { |
83 | 31.2M | int16_t n; |
84 | | |
85 | | /* shift input buffer x */ |
86 | | /* input buffer is not shifted anymore, x is implemented as double ringbuffer */ |
87 | | //memmove(qmfa->x + 32, qmfa->x, (320-32)*sizeof(real_t)); |
88 | | |
89 | | /* add new samples to input buffer x */ |
90 | 1.03G | for (n = 32 - 1; n >= 0; n--) |
91 | 1.00G | { |
92 | | #ifdef FIXED_POINT |
93 | | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = (input[in++]) >> 4; |
94 | | #else |
95 | | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = input[in++]; |
96 | | #endif |
97 | 1.00G | } |
98 | | |
99 | | /* window and summation to create array u */ |
100 | 2.03G | for (n = 0; n < 64; n++) |
101 | 2.00G | { |
102 | 2.00G | u[n] = MUL_F(qmfa->x[qmfa->x_index + n], qmf_c[2*n]) + |
103 | 2.00G | MUL_F(qmfa->x[qmfa->x_index + n + 64], qmf_c[2*(n + 64)]) + |
104 | 2.00G | MUL_F(qmfa->x[qmfa->x_index + n + 128], qmf_c[2*(n + 128)]) + |
105 | 2.00G | MUL_F(qmfa->x[qmfa->x_index + n + 192], qmf_c[2*(n + 192)]) + |
106 | 2.00G | MUL_F(qmfa->x[qmfa->x_index + n + 256], qmf_c[2*(n + 256)]); |
107 | 2.00G | } |
108 | | |
109 | | /* update ringbuffer index */ |
110 | 31.2M | qmfa->x_index -= 32; |
111 | 31.2M | if (qmfa->x_index < 0) |
112 | 3.69M | qmfa->x_index = (320-32); |
113 | | |
114 | | /* calculate 32 subband samples by introducing X */ |
115 | | #ifdef SBR_LOW_POWER |
116 | | y[0] = u[48]; |
117 | | for (n = 1; n < 16; n++) |
118 | | y[n] = u[n+48] + u[48-n]; |
119 | | for (n = 16; n < 32; n++) |
120 | | y[n] = -u[n-16] + u[48-n]; |
121 | | |
122 | | DCT3_32_unscaled(u, y); |
123 | | |
124 | | for (n = 0; n < 32; n++) |
125 | | { |
126 | | if (n < kx) |
127 | | { |
128 | | #ifdef FIXED_POINT |
129 | | QMF_RE(X[l + offset][n]) = u[n] /*<< 1*/; |
130 | | #else |
131 | | QMF_RE(X[l + offset][n]) = 2. * u[n]; |
132 | | #endif |
133 | | } else { |
134 | | QMF_RE(X[l + offset][n]) = 0; |
135 | | } |
136 | | } |
137 | | #else |
138 | | |
139 | | // Reordering of data moved from DCT_IV to here |
140 | 31.2M | in_imag[31] = u[1]; |
141 | 31.2M | in_real[0] = u[0]; |
142 | 970M | for (n = 1; n < 31; n++) |
143 | 938M | { |
144 | 938M | in_imag[31 - n] = u[n+1]; |
145 | 938M | in_real[n] = -u[64-n]; |
146 | 938M | } |
147 | 31.2M | in_imag[0] = u[32]; |
148 | 31.2M | in_real[31] = -u[33]; |
149 | | |
150 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT |
151 | 31.2M | dct4_kernel(in_real, in_imag, out_real, out_imag); |
152 | | |
153 | | // Reordering of data moved from DCT_IV to here |
154 | 531M | for (n = 0; n < 16; n++) { |
155 | 500M | if (2*n+1 < kx) { |
156 | | #ifdef FIXED_POINT |
157 | 244M | QMF_RE(X[l + offset][2*n]) = out_real[n]; |
158 | 244M | QMF_IM(X[l + offset][2*n]) = out_imag[n]; |
159 | 244M | QMF_RE(X[l + offset][2*n+1]) = -out_imag[31-n]; |
160 | 244M | QMF_IM(X[l + offset][2*n+1]) = -out_real[31-n]; |
161 | | #else |
162 | 244M | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; |
163 | 244M | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; |
164 | 244M | QMF_RE(X[l + offset][2*n+1]) = -2. * out_imag[31-n]; |
165 | 244M | QMF_IM(X[l + offset][2*n+1]) = -2. * out_real[31-n]; |
166 | | #endif |
167 | 489M | } else { |
168 | 11.0M | if (2*n < kx) { |
169 | | #ifdef FIXED_POINT |
170 | 437k | QMF_RE(X[l + offset][2*n]) = out_real[n]; |
171 | 437k | QMF_IM(X[l + offset][2*n]) = out_imag[n]; |
172 | | #else |
173 | 437k | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; |
174 | 437k | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; |
175 | | #endif |
176 | 874k | } |
177 | 10.1M | else { |
178 | 10.1M | QMF_RE(X[l + offset][2*n]) = 0; |
179 | 10.1M | QMF_IM(X[l + offset][2*n]) = 0; |
180 | 10.1M | } |
181 | 11.0M | QMF_RE(X[l + offset][2*n+1]) = 0; |
182 | 11.0M | QMF_IM(X[l + offset][2*n+1]) = 0; |
183 | 11.0M | } |
184 | 500M | } |
185 | 31.2M | #endif |
186 | 31.2M | } |
187 | 988k | } Line | Count | Source | 70 | 494k | { | 71 | 494k | ALIGN real_t u[64]; | 72 | 494k | #ifndef SBR_LOW_POWER | 73 | 494k | ALIGN real_t in_real[32], in_imag[32], out_real[32], out_imag[32]; | 74 | | #else | 75 | | ALIGN real_t y[32]; | 76 | | #endif | 77 | 494k | uint32_t in = 0; | 78 | 494k | uint8_t l; | 79 | | | 80 | | /* qmf subsample l */ | 81 | 16.1M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 82 | 15.6M | { | 83 | 15.6M | int16_t n; | 84 | | | 85 | | /* shift input buffer x */ | 86 | | /* input buffer is not shifted anymore, x is implemented as double ringbuffer */ | 87 | | //memmove(qmfa->x + 32, qmfa->x, (320-32)*sizeof(real_t)); | 88 | | | 89 | | /* add new samples to input buffer x */ | 90 | 516M | for (n = 32 - 1; n >= 0; n--) | 91 | 500M | { | 92 | 500M | #ifdef FIXED_POINT | 93 | 500M | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = (input[in++]) >> 4; | 94 | | #else | 95 | | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = input[in++]; | 96 | | #endif | 97 | 500M | } | 98 | | | 99 | | /* window and summation to create array u */ | 100 | 1.01G | for (n = 0; n < 64; n++) | 101 | 1.00G | { | 102 | 1.00G | u[n] = MUL_F(qmfa->x[qmfa->x_index + n], qmf_c[2*n]) + | 103 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 64], qmf_c[2*(n + 64)]) + | 104 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 128], qmf_c[2*(n + 128)]) + | 105 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 192], qmf_c[2*(n + 192)]) + | 106 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 256], qmf_c[2*(n + 256)]); | 107 | 1.00G | } | 108 | | | 109 | | /* update ringbuffer index */ | 110 | 15.6M | qmfa->x_index -= 32; | 111 | 15.6M | if (qmfa->x_index < 0) | 112 | 1.84M | qmfa->x_index = (320-32); | 113 | | | 114 | | /* calculate 32 subband samples by introducing X */ | 115 | | #ifdef SBR_LOW_POWER | 116 | | y[0] = u[48]; | 117 | | for (n = 1; n < 16; n++) | 118 | | y[n] = u[n+48] + u[48-n]; | 119 | | for (n = 16; n < 32; n++) | 120 | | y[n] = -u[n-16] + u[48-n]; | 121 | | | 122 | | DCT3_32_unscaled(u, y); | 123 | | | 124 | | for (n = 0; n < 32; n++) | 125 | | { | 126 | | if (n < kx) | 127 | | { | 128 | | #ifdef FIXED_POINT | 129 | | QMF_RE(X[l + offset][n]) = u[n] /*<< 1*/; | 130 | | #else | 131 | | QMF_RE(X[l + offset][n]) = 2. * u[n]; | 132 | | #endif | 133 | | } else { | 134 | | QMF_RE(X[l + offset][n]) = 0; | 135 | | } | 136 | | } | 137 | | #else | 138 | | | 139 | | // Reordering of data moved from DCT_IV to here | 140 | 15.6M | in_imag[31] = u[1]; | 141 | 15.6M | in_real[0] = u[0]; | 142 | 485M | for (n = 1; n < 31; n++) | 143 | 469M | { | 144 | 469M | in_imag[31 - n] = u[n+1]; | 145 | 469M | in_real[n] = -u[64-n]; | 146 | 469M | } | 147 | 15.6M | in_imag[0] = u[32]; | 148 | 15.6M | in_real[31] = -u[33]; | 149 | | | 150 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT | 151 | 15.6M | dct4_kernel(in_real, in_imag, out_real, out_imag); | 152 | | | 153 | | // Reordering of data moved from DCT_IV to here | 154 | 265M | for (n = 0; n < 16; n++) { | 155 | 250M | if (2*n+1 < kx) { | 156 | 244M | #ifdef FIXED_POINT | 157 | 244M | QMF_RE(X[l + offset][2*n]) = out_real[n]; | 158 | 244M | QMF_IM(X[l + offset][2*n]) = out_imag[n]; | 159 | 244M | QMF_RE(X[l + offset][2*n+1]) = -out_imag[31-n]; | 160 | 244M | QMF_IM(X[l + offset][2*n+1]) = -out_real[31-n]; | 161 | | #else | 162 | | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; | 163 | | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; | 164 | | QMF_RE(X[l + offset][2*n+1]) = -2. * out_imag[31-n]; | 165 | | QMF_IM(X[l + offset][2*n+1]) = -2. * out_real[31-n]; | 166 | | #endif | 167 | 244M | } else { | 168 | 5.53M | if (2*n < kx) { | 169 | 437k | #ifdef FIXED_POINT | 170 | 437k | QMF_RE(X[l + offset][2*n]) = out_real[n]; | 171 | 437k | QMF_IM(X[l + offset][2*n]) = out_imag[n]; | 172 | | #else | 173 | | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; | 174 | | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; | 175 | | #endif | 176 | 437k | } | 177 | 5.09M | else { | 178 | 5.09M | QMF_RE(X[l + offset][2*n]) = 0; | 179 | 5.09M | QMF_IM(X[l + offset][2*n]) = 0; | 180 | 5.09M | } | 181 | 5.53M | QMF_RE(X[l + offset][2*n+1]) = 0; | 182 | 5.53M | QMF_IM(X[l + offset][2*n+1]) = 0; | 183 | 5.53M | } | 184 | 250M | } | 185 | 15.6M | #endif | 186 | 15.6M | } | 187 | 494k | } |
Line | Count | Source | 70 | 494k | { | 71 | 494k | ALIGN real_t u[64]; | 72 | 494k | #ifndef SBR_LOW_POWER | 73 | 494k | ALIGN real_t in_real[32], in_imag[32], out_real[32], out_imag[32]; | 74 | | #else | 75 | | ALIGN real_t y[32]; | 76 | | #endif | 77 | 494k | uint32_t in = 0; | 78 | 494k | uint8_t l; | 79 | | | 80 | | /* qmf subsample l */ | 81 | 16.1M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 82 | 15.6M | { | 83 | 15.6M | int16_t n; | 84 | | | 85 | | /* shift input buffer x */ | 86 | | /* input buffer is not shifted anymore, x is implemented as double ringbuffer */ | 87 | | //memmove(qmfa->x + 32, qmfa->x, (320-32)*sizeof(real_t)); | 88 | | | 89 | | /* add new samples to input buffer x */ | 90 | 516M | for (n = 32 - 1; n >= 0; n--) | 91 | 500M | { | 92 | | #ifdef FIXED_POINT | 93 | | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = (input[in++]) >> 4; | 94 | | #else | 95 | 500M | qmfa->x[qmfa->x_index + n] = qmfa->x[qmfa->x_index + n + 320] = input[in++]; | 96 | 500M | #endif | 97 | 500M | } | 98 | | | 99 | | /* window and summation to create array u */ | 100 | 1.01G | for (n = 0; n < 64; n++) | 101 | 1.00G | { | 102 | 1.00G | u[n] = MUL_F(qmfa->x[qmfa->x_index + n], qmf_c[2*n]) + | 103 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 64], qmf_c[2*(n + 64)]) + | 104 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 128], qmf_c[2*(n + 128)]) + | 105 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 192], qmf_c[2*(n + 192)]) + | 106 | 1.00G | MUL_F(qmfa->x[qmfa->x_index + n + 256], qmf_c[2*(n + 256)]); | 107 | 1.00G | } | 108 | | | 109 | | /* update ringbuffer index */ | 110 | 15.6M | qmfa->x_index -= 32; | 111 | 15.6M | if (qmfa->x_index < 0) | 112 | 1.84M | qmfa->x_index = (320-32); | 113 | | | 114 | | /* calculate 32 subband samples by introducing X */ | 115 | | #ifdef SBR_LOW_POWER | 116 | | y[0] = u[48]; | 117 | | for (n = 1; n < 16; n++) | 118 | | y[n] = u[n+48] + u[48-n]; | 119 | | for (n = 16; n < 32; n++) | 120 | | y[n] = -u[n-16] + u[48-n]; | 121 | | | 122 | | DCT3_32_unscaled(u, y); | 123 | | | 124 | | for (n = 0; n < 32; n++) | 125 | | { | 126 | | if (n < kx) | 127 | | { | 128 | | #ifdef FIXED_POINT | 129 | | QMF_RE(X[l + offset][n]) = u[n] /*<< 1*/; | 130 | | #else | 131 | | QMF_RE(X[l + offset][n]) = 2. * u[n]; | 132 | | #endif | 133 | | } else { | 134 | | QMF_RE(X[l + offset][n]) = 0; | 135 | | } | 136 | | } | 137 | | #else | 138 | | | 139 | | // Reordering of data moved from DCT_IV to here | 140 | 15.6M | in_imag[31] = u[1]; | 141 | 15.6M | in_real[0] = u[0]; | 142 | 485M | for (n = 1; n < 31; n++) | 143 | 469M | { | 144 | 469M | in_imag[31 - n] = u[n+1]; | 145 | 469M | in_real[n] = -u[64-n]; | 146 | 469M | } | 147 | 15.6M | in_imag[0] = u[32]; | 148 | 15.6M | in_real[31] = -u[33]; | 149 | | | 150 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT | 151 | 15.6M | dct4_kernel(in_real, in_imag, out_real, out_imag); | 152 | | | 153 | | // Reordering of data moved from DCT_IV to here | 154 | 265M | for (n = 0; n < 16; n++) { | 155 | 250M | if (2*n+1 < kx) { | 156 | | #ifdef FIXED_POINT | 157 | | QMF_RE(X[l + offset][2*n]) = out_real[n]; | 158 | | QMF_IM(X[l + offset][2*n]) = out_imag[n]; | 159 | | QMF_RE(X[l + offset][2*n+1]) = -out_imag[31-n]; | 160 | | QMF_IM(X[l + offset][2*n+1]) = -out_real[31-n]; | 161 | | #else | 162 | 244M | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; | 163 | 244M | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; | 164 | 244M | QMF_RE(X[l + offset][2*n+1]) = -2. * out_imag[31-n]; | 165 | 244M | QMF_IM(X[l + offset][2*n+1]) = -2. * out_real[31-n]; | 166 | 244M | #endif | 167 | 244M | } else { | 168 | 5.53M | if (2*n < kx) { | 169 | | #ifdef FIXED_POINT | 170 | | QMF_RE(X[l + offset][2*n]) = out_real[n]; | 171 | | QMF_IM(X[l + offset][2*n]) = out_imag[n]; | 172 | | #else | 173 | 437k | QMF_RE(X[l + offset][2*n]) = 2. * out_real[n]; | 174 | 437k | QMF_IM(X[l + offset][2*n]) = 2. * out_imag[n]; | 175 | 437k | #endif | 176 | 437k | } | 177 | 5.09M | else { | 178 | 5.09M | QMF_RE(X[l + offset][2*n]) = 0; | 179 | 5.09M | QMF_IM(X[l + offset][2*n]) = 0; | 180 | 5.09M | } | 181 | 5.53M | QMF_RE(X[l + offset][2*n+1]) = 0; | 182 | 5.53M | QMF_IM(X[l + offset][2*n+1]) = 0; | 183 | 5.53M | } | 184 | 250M | } | 185 | 15.6M | #endif | 186 | 15.6M | } | 187 | 494k | } |
|
188 | | |
189 | | static const complex_t qmf32_pre_twiddle[] = |
190 | | { |
191 | | { FRAC_CONST(0.999924701839145), FRAC_CONST(-0.012271538285720) }, |
192 | | { FRAC_CONST(0.999322384588350), FRAC_CONST(-0.036807222941359) }, |
193 | | { FRAC_CONST(0.998118112900149), FRAC_CONST(-0.061320736302209) }, |
194 | | { FRAC_CONST(0.996312612182778), FRAC_CONST(-0.085797312344440) }, |
195 | | { FRAC_CONST(0.993906970002356), FRAC_CONST(-0.110222207293883) }, |
196 | | { FRAC_CONST(0.990902635427780), FRAC_CONST(-0.134580708507126) }, |
197 | | { FRAC_CONST(0.987301418157858), FRAC_CONST(-0.158858143333861) }, |
198 | | { FRAC_CONST(0.983105487431216), FRAC_CONST(-0.183039887955141) }, |
199 | | { FRAC_CONST(0.978317370719628), FRAC_CONST(-0.207111376192219) }, |
200 | | { FRAC_CONST(0.972939952205560), FRAC_CONST(-0.231058108280671) }, |
201 | | { FRAC_CONST(0.966976471044852), FRAC_CONST(-0.254865659604515) }, |
202 | | { FRAC_CONST(0.960430519415566), FRAC_CONST(-0.278519689385053) }, |
203 | | { FRAC_CONST(0.953306040354194), FRAC_CONST(-0.302005949319228) }, |
204 | | { FRAC_CONST(0.945607325380521), FRAC_CONST(-0.325310292162263) }, |
205 | | { FRAC_CONST(0.937339011912575), FRAC_CONST(-0.348418680249435) }, |
206 | | { FRAC_CONST(0.928506080473216), FRAC_CONST(-0.371317193951838) }, |
207 | | { FRAC_CONST(0.919113851690058), FRAC_CONST(-0.393992040061048) }, |
208 | | { FRAC_CONST(0.909167983090522), FRAC_CONST(-0.416429560097637) }, |
209 | | { FRAC_CONST(0.898674465693954), FRAC_CONST(-0.438616238538528) }, |
210 | | { FRAC_CONST(0.887639620402854), FRAC_CONST(-0.460538710958240) }, |
211 | | { FRAC_CONST(0.876070094195407), FRAC_CONST(-0.482183772079123) }, |
212 | | { FRAC_CONST(0.863972856121587), FRAC_CONST(-0.503538383725718) }, |
213 | | { FRAC_CONST(0.851355193105265), FRAC_CONST(-0.524589682678469) }, |
214 | | { FRAC_CONST(0.838224705554838), FRAC_CONST(-0.545324988422046) }, |
215 | | { FRAC_CONST(0.824589302785025), FRAC_CONST(-0.565731810783613) }, |
216 | | { FRAC_CONST(0.810457198252595), FRAC_CONST(-0.585797857456439) }, |
217 | | { FRAC_CONST(0.795836904608884), FRAC_CONST(-0.605511041404326) }, |
218 | | { FRAC_CONST(0.780737228572094), FRAC_CONST(-0.624859488142386) }, |
219 | | { FRAC_CONST(0.765167265622459), FRAC_CONST(-0.643831542889791) }, |
220 | | { FRAC_CONST(0.749136394523459), FRAC_CONST(-0.662415777590172) }, |
221 | | { FRAC_CONST(0.732654271672413), FRAC_CONST(-0.680600997795453) }, |
222 | | { FRAC_CONST(0.715730825283819), FRAC_CONST(-0.698376249408973) } |
223 | | }; |
224 | | |
225 | | qmfs_info *qmfs_init(uint8_t channels) |
226 | 436k | { |
227 | 436k | qmfs_info *qmfs = (qmfs_info*)faad_malloc(sizeof(qmfs_info)); |
228 | | |
229 | | /* v is a double ringbuffer */ |
230 | 436k | qmfs->v = (real_t*)faad_malloc(2 * channels * 20 * sizeof(real_t)); |
231 | 436k | memset(qmfs->v, 0, 2 * channels * 20 * sizeof(real_t)); |
232 | | |
233 | 436k | qmfs->v_index = 0; |
234 | | |
235 | 436k | qmfs->channels = channels; |
236 | | |
237 | 436k | return qmfs; |
238 | 436k | } |
239 | | |
240 | | void qmfs_end(qmfs_info *qmfs) |
241 | 436k | { |
242 | 436k | if (qmfs) |
243 | 436k | { |
244 | 436k | if (qmfs->v) faad_free(qmfs->v); |
245 | 436k | faad_free(qmfs); |
246 | 436k | } |
247 | 436k | } |
248 | | |
249 | | #ifdef SBR_LOW_POWER |
250 | | |
251 | | void sbr_qmf_synthesis_32(sbr_info *sbr, qmfs_info *qmfs, qmf_t X[MAX_NTSRHFG][64], |
252 | | real_t *output) |
253 | | { |
254 | | ALIGN real_t x[16]; |
255 | | ALIGN real_t y[16]; |
256 | | int32_t n, k, out = 0; |
257 | | uint8_t l; |
258 | | |
259 | | /* qmf subsample l */ |
260 | | for (l = 0; l < sbr->numTimeSlotsRate; l++) |
261 | | { |
262 | | /* shift buffers */ |
263 | | /* we are not shifting v, it is a double ringbuffer */ |
264 | | //memmove(qmfs->v + 64, qmfs->v, (640-64)*sizeof(real_t)); |
265 | | |
266 | | /* calculate 64 samples */ |
267 | | for (k = 0; k < 16; k++) |
268 | | { |
269 | | #ifdef FIXED_POINT |
270 | | y[k] = (QMF_RE(X[l][k]) - QMF_RE(X[l][31 - k])); |
271 | | x[k] = (QMF_RE(X[l][k]) + QMF_RE(X[l][31 - k])); |
272 | | #else |
273 | | y[k] = (QMF_RE(X[l][k]) - QMF_RE(X[l][31 - k])) / 32.0; |
274 | | x[k] = (QMF_RE(X[l][k]) + QMF_RE(X[l][31 - k])) / 32.0; |
275 | | #endif |
276 | | } |
277 | | |
278 | | /* even n samples */ |
279 | | DCT2_16_unscaled(x, x); |
280 | | /* odd n samples */ |
281 | | DCT4_16(y, y); |
282 | | |
283 | | for (n = 8; n < 24; n++) |
284 | | { |
285 | | qmfs->v[qmfs->v_index + n*2] = qmfs->v[qmfs->v_index + 640 + n*2] = x[n-8]; |
286 | | qmfs->v[qmfs->v_index + n*2+1] = qmfs->v[qmfs->v_index + 640 + n*2+1] = y[n-8]; |
287 | | } |
288 | | for (n = 0; n < 16; n++) |
289 | | { |
290 | | qmfs->v[qmfs->v_index + n] = qmfs->v[qmfs->v_index + 640 + n] = qmfs->v[qmfs->v_index + 32-n]; |
291 | | } |
292 | | qmfs->v[qmfs->v_index + 48] = qmfs->v[qmfs->v_index + 640 + 48] = 0; |
293 | | for (n = 1; n < 16; n++) |
294 | | { |
295 | | qmfs->v[qmfs->v_index + 48+n] = qmfs->v[qmfs->v_index + 640 + 48+n] = -qmfs->v[qmfs->v_index + 48-n]; |
296 | | } |
297 | | |
298 | | /* calculate 32 output samples and window */ |
299 | | for (k = 0; k < 32; k++) |
300 | | { |
301 | | output[out++] = MUL_F(qmfs->v[qmfs->v_index + k], qmf_c[2*k]) + |
302 | | MUL_F(qmfs->v[qmfs->v_index + 96 + k], qmf_c[64 + 2*k]) + |
303 | | MUL_F(qmfs->v[qmfs->v_index + 128 + k], qmf_c[128 + 2*k]) + |
304 | | MUL_F(qmfs->v[qmfs->v_index + 224 + k], qmf_c[192 + 2*k]) + |
305 | | MUL_F(qmfs->v[qmfs->v_index + 256 + k], qmf_c[256 + 2*k]) + |
306 | | MUL_F(qmfs->v[qmfs->v_index + 352 + k], qmf_c[320 + 2*k]) + |
307 | | MUL_F(qmfs->v[qmfs->v_index + 384 + k], qmf_c[384 + 2*k]) + |
308 | | MUL_F(qmfs->v[qmfs->v_index + 480 + k], qmf_c[448 + 2*k]) + |
309 | | MUL_F(qmfs->v[qmfs->v_index + 512 + k], qmf_c[512 + 2*k]) + |
310 | | MUL_F(qmfs->v[qmfs->v_index + 608 + k], qmf_c[576 + 2*k]); |
311 | | } |
312 | | |
313 | | /* update the ringbuffer index */ |
314 | | qmfs->v_index -= 64; |
315 | | if (qmfs->v_index < 0) |
316 | | qmfs->v_index = (640-64); |
317 | | } |
318 | | } |
319 | | |
320 | | void sbr_qmf_synthesis_64(sbr_info *sbr, qmfs_info *qmfs, qmf_t X[MAX_NTSRHFG][64], |
321 | | real_t *output) |
322 | | { |
323 | | ALIGN real_t x[64]; |
324 | | ALIGN real_t y[64]; |
325 | | int32_t n, k, out = 0; |
326 | | uint8_t l; |
327 | | |
328 | | |
329 | | /* qmf subsample l */ |
330 | | for (l = 0; l < sbr->numTimeSlotsRate; l++) |
331 | | { |
332 | | /* shift buffers */ |
333 | | /* we are not shifting v, it is a double ringbuffer */ |
334 | | //memmove(qmfs->v + 128, qmfs->v, (1280-128)*sizeof(real_t)); |
335 | | |
336 | | /* calculate 128 samples */ |
337 | | for (k = 0; k < 32; k++) |
338 | | { |
339 | | #ifdef FIXED_POINT |
340 | | y[k] = (QMF_RE(X[l][k]) - QMF_RE(X[l][63 - k])); |
341 | | x[k] = (QMF_RE(X[l][k]) + QMF_RE(X[l][63 - k])); |
342 | | #else |
343 | | y[k] = (QMF_RE(X[l][k]) - QMF_RE(X[l][63 - k])) / 32.0; |
344 | | x[k] = (QMF_RE(X[l][k]) + QMF_RE(X[l][63 - k])) / 32.0; |
345 | | #endif |
346 | | } |
347 | | |
348 | | /* even n samples */ |
349 | | DCT2_32_unscaled(x, x); |
350 | | /* odd n samples */ |
351 | | DCT4_32(y, y); |
352 | | |
353 | | for (n = 16; n < 48; n++) |
354 | | { |
355 | | qmfs->v[qmfs->v_index + n*2] = qmfs->v[qmfs->v_index + 1280 + n*2] = x[n-16]; |
356 | | qmfs->v[qmfs->v_index + n*2+1] = qmfs->v[qmfs->v_index + 1280 + n*2+1] = y[n-16]; |
357 | | } |
358 | | for (n = 0; n < 32; n++) |
359 | | { |
360 | | qmfs->v[qmfs->v_index + n] = qmfs->v[qmfs->v_index + 1280 + n] = qmfs->v[qmfs->v_index + 64-n]; |
361 | | } |
362 | | qmfs->v[qmfs->v_index + 96] = qmfs->v[qmfs->v_index + 1280 + 96] = 0; |
363 | | for (n = 1; n < 32; n++) |
364 | | { |
365 | | qmfs->v[qmfs->v_index + 96+n] = qmfs->v[qmfs->v_index + 1280 + 96+n] = -qmfs->v[qmfs->v_index + 96-n]; |
366 | | } |
367 | | |
368 | | /* calculate 64 output samples and window */ |
369 | | for (k = 0; k < 64; k++) |
370 | | { |
371 | | output[out++] = MUL_F(qmfs->v[qmfs->v_index + k], qmf_c[k]) + |
372 | | MUL_F(qmfs->v[qmfs->v_index + 192 + k], qmf_c[64 + k]) + |
373 | | MUL_F(qmfs->v[qmfs->v_index + 256 + k], qmf_c[128 + k]) + |
374 | | MUL_F(qmfs->v[qmfs->v_index + 256 + 192 + k], qmf_c[128 + 64 + k]) + |
375 | | MUL_F(qmfs->v[qmfs->v_index + 512 + k], qmf_c[256 + k]) + |
376 | | MUL_F(qmfs->v[qmfs->v_index + 512 + 192 + k], qmf_c[256 + 64 + k]) + |
377 | | MUL_F(qmfs->v[qmfs->v_index + 768 + k], qmf_c[384 + k]) + |
378 | | MUL_F(qmfs->v[qmfs->v_index + 768 + 192 + k], qmf_c[384 + 64 + k]) + |
379 | | MUL_F(qmfs->v[qmfs->v_index + 1024 + k], qmf_c[512 + k]) + |
380 | | MUL_F(qmfs->v[qmfs->v_index + 1024 + 192 + k], qmf_c[512 + 64 + k]); |
381 | | } |
382 | | |
383 | | /* update the ringbuffer index */ |
384 | | qmfs->v_index -= 128; |
385 | | if (qmfs->v_index < 0) |
386 | | qmfs->v_index = (1280-128); |
387 | | } |
388 | | } |
389 | | #else |
390 | | void sbr_qmf_synthesis_32(sbr_info *sbr, qmfs_info *qmfs, qmf_t X[MAX_NTSRHFG][64], |
391 | | real_t *output) |
392 | 358k | { |
393 | 358k | ALIGN real_t x1[32], x2[32]; |
394 | | #ifndef FIXED_POINT |
395 | | real_t scale = 1.f/64.f; |
396 | | #endif |
397 | 358k | int32_t n, k, out = 0; |
398 | 358k | uint8_t l; |
399 | | |
400 | | |
401 | | /* qmf subsample l */ |
402 | 11.7M | for (l = 0; l < sbr->numTimeSlotsRate; l++) |
403 | 11.3M | { |
404 | | /* shift buffer v */ |
405 | | /* buffer is not shifted, we are using a ringbuffer */ |
406 | | //memmove(qmfs->v + 64, qmfs->v, (640-64)*sizeof(real_t)); |
407 | | |
408 | | /* calculate 64 samples */ |
409 | | /* complex pre-twiddle */ |
410 | 375M | for (k = 0; k < 32; k++) |
411 | 363M | { |
412 | 363M | x1[k] = MUL_F(QMF_RE(X[l][k]), RE(qmf32_pre_twiddle[k])) - MUL_F(QMF_IM(X[l][k]), IM(qmf32_pre_twiddle[k])); |
413 | 363M | x2[k] = MUL_F(QMF_IM(X[l][k]), RE(qmf32_pre_twiddle[k])) + MUL_F(QMF_RE(X[l][k]), IM(qmf32_pre_twiddle[k])); |
414 | | |
415 | | #ifndef FIXED_POINT |
416 | | x1[k] *= scale; |
417 | | x2[k] *= scale; |
418 | | #else |
419 | | x1[k] >>= 1; |
420 | | x2[k] >>= 1; |
421 | | #endif |
422 | 363M | } |
423 | | |
424 | | /* transform */ |
425 | 11.3M | DCT4_32(x1, x1); |
426 | 11.3M | DST4_32(x2, x2); |
427 | | |
428 | 375M | for (n = 0; n < 32; n++) |
429 | 363M | { |
430 | 363M | qmfs->v[qmfs->v_index + n] = qmfs->v[qmfs->v_index + 640 + n] = -x1[n] + x2[n]; |
431 | 363M | qmfs->v[qmfs->v_index + 63 - n] = qmfs->v[qmfs->v_index + 640 + 63 - n] = x1[n] + x2[n]; |
432 | 363M | } |
433 | | |
434 | | /* calculate 32 output samples and window */ |
435 | 375M | for (k = 0; k < 32; k++) |
436 | 363M | { |
437 | 363M | output[out++] = MUL_F(qmfs->v[qmfs->v_index + k], qmf_c[2*k]) + |
438 | 363M | MUL_F(qmfs->v[qmfs->v_index + 96 + k], qmf_c[64 + 2*k]) + |
439 | 363M | MUL_F(qmfs->v[qmfs->v_index + 128 + k], qmf_c[128 + 2*k]) + |
440 | 363M | MUL_F(qmfs->v[qmfs->v_index + 224 + k], qmf_c[192 + 2*k]) + |
441 | 363M | MUL_F(qmfs->v[qmfs->v_index + 256 + k], qmf_c[256 + 2*k]) + |
442 | 363M | MUL_F(qmfs->v[qmfs->v_index + 352 + k], qmf_c[320 + 2*k]) + |
443 | 363M | MUL_F(qmfs->v[qmfs->v_index + 384 + k], qmf_c[384 + 2*k]) + |
444 | 363M | MUL_F(qmfs->v[qmfs->v_index + 480 + k], qmf_c[448 + 2*k]) + |
445 | 363M | MUL_F(qmfs->v[qmfs->v_index + 512 + k], qmf_c[512 + 2*k]) + |
446 | 363M | MUL_F(qmfs->v[qmfs->v_index + 608 + k], qmf_c[576 + 2*k]); |
447 | 363M | } |
448 | | |
449 | | /* update ringbuffer index */ |
450 | 11.3M | qmfs->v_index -= 64; |
451 | 11.3M | if (qmfs->v_index < 0) |
452 | 1.34M | qmfs->v_index = (640 - 64); |
453 | 11.3M | } |
454 | 358k | } Line | Count | Source | 392 | 179k | { | 393 | 179k | ALIGN real_t x1[32], x2[32]; | 394 | | #ifndef FIXED_POINT | 395 | | real_t scale = 1.f/64.f; | 396 | | #endif | 397 | 179k | int32_t n, k, out = 0; | 398 | 179k | uint8_t l; | 399 | | | 400 | | | 401 | | /* qmf subsample l */ | 402 | 5.86M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 403 | 5.68M | { | 404 | | /* shift buffer v */ | 405 | | /* buffer is not shifted, we are using a ringbuffer */ | 406 | | //memmove(qmfs->v + 64, qmfs->v, (640-64)*sizeof(real_t)); | 407 | | | 408 | | /* calculate 64 samples */ | 409 | | /* complex pre-twiddle */ | 410 | 187M | for (k = 0; k < 32; k++) | 411 | 181M | { | 412 | 181M | x1[k] = MUL_F(QMF_RE(X[l][k]), RE(qmf32_pre_twiddle[k])) - MUL_F(QMF_IM(X[l][k]), IM(qmf32_pre_twiddle[k])); | 413 | 181M | x2[k] = MUL_F(QMF_IM(X[l][k]), RE(qmf32_pre_twiddle[k])) + MUL_F(QMF_RE(X[l][k]), IM(qmf32_pre_twiddle[k])); | 414 | | | 415 | | #ifndef FIXED_POINT | 416 | | x1[k] *= scale; | 417 | | x2[k] *= scale; | 418 | | #else | 419 | 181M | x1[k] >>= 1; | 420 | 181M | x2[k] >>= 1; | 421 | 181M | #endif | 422 | 181M | } | 423 | | | 424 | | /* transform */ | 425 | 5.68M | DCT4_32(x1, x1); | 426 | 5.68M | DST4_32(x2, x2); | 427 | | | 428 | 187M | for (n = 0; n < 32; n++) | 429 | 181M | { | 430 | 181M | qmfs->v[qmfs->v_index + n] = qmfs->v[qmfs->v_index + 640 + n] = -x1[n] + x2[n]; | 431 | 181M | qmfs->v[qmfs->v_index + 63 - n] = qmfs->v[qmfs->v_index + 640 + 63 - n] = x1[n] + x2[n]; | 432 | 181M | } | 433 | | | 434 | | /* calculate 32 output samples and window */ | 435 | 187M | for (k = 0; k < 32; k++) | 436 | 181M | { | 437 | 181M | output[out++] = MUL_F(qmfs->v[qmfs->v_index + k], qmf_c[2*k]) + | 438 | 181M | MUL_F(qmfs->v[qmfs->v_index + 96 + k], qmf_c[64 + 2*k]) + | 439 | 181M | MUL_F(qmfs->v[qmfs->v_index + 128 + k], qmf_c[128 + 2*k]) + | 440 | 181M | MUL_F(qmfs->v[qmfs->v_index + 224 + k], qmf_c[192 + 2*k]) + | 441 | 181M | MUL_F(qmfs->v[qmfs->v_index + 256 + k], qmf_c[256 + 2*k]) + | 442 | 181M | MUL_F(qmfs->v[qmfs->v_index + 352 + k], qmf_c[320 + 2*k]) + | 443 | 181M | MUL_F(qmfs->v[qmfs->v_index + 384 + k], qmf_c[384 + 2*k]) + | 444 | 181M | MUL_F(qmfs->v[qmfs->v_index + 480 + k], qmf_c[448 + 2*k]) + | 445 | 181M | MUL_F(qmfs->v[qmfs->v_index + 512 + k], qmf_c[512 + 2*k]) + | 446 | 181M | MUL_F(qmfs->v[qmfs->v_index + 608 + k], qmf_c[576 + 2*k]); | 447 | 181M | } | 448 | | | 449 | | /* update ringbuffer index */ | 450 | 5.68M | qmfs->v_index -= 64; | 451 | 5.68M | if (qmfs->v_index < 0) | 452 | 671k | qmfs->v_index = (640 - 64); | 453 | 5.68M | } | 454 | 179k | } |
Line | Count | Source | 392 | 179k | { | 393 | 179k | ALIGN real_t x1[32], x2[32]; | 394 | 179k | #ifndef FIXED_POINT | 395 | 179k | real_t scale = 1.f/64.f; | 396 | 179k | #endif | 397 | 179k | int32_t n, k, out = 0; | 398 | 179k | uint8_t l; | 399 | | | 400 | | | 401 | | /* qmf subsample l */ | 402 | 5.86M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 403 | 5.68M | { | 404 | | /* shift buffer v */ | 405 | | /* buffer is not shifted, we are using a ringbuffer */ | 406 | | //memmove(qmfs->v + 64, qmfs->v, (640-64)*sizeof(real_t)); | 407 | | | 408 | | /* calculate 64 samples */ | 409 | | /* complex pre-twiddle */ | 410 | 187M | for (k = 0; k < 32; k++) | 411 | 181M | { | 412 | 181M | x1[k] = MUL_F(QMF_RE(X[l][k]), RE(qmf32_pre_twiddle[k])) - MUL_F(QMF_IM(X[l][k]), IM(qmf32_pre_twiddle[k])); | 413 | 181M | x2[k] = MUL_F(QMF_IM(X[l][k]), RE(qmf32_pre_twiddle[k])) + MUL_F(QMF_RE(X[l][k]), IM(qmf32_pre_twiddle[k])); | 414 | | | 415 | 181M | #ifndef FIXED_POINT | 416 | 181M | x1[k] *= scale; | 417 | 181M | x2[k] *= scale; | 418 | | #else | 419 | | x1[k] >>= 1; | 420 | | x2[k] >>= 1; | 421 | | #endif | 422 | 181M | } | 423 | | | 424 | | /* transform */ | 425 | 5.68M | DCT4_32(x1, x1); | 426 | 5.68M | DST4_32(x2, x2); | 427 | | | 428 | 187M | for (n = 0; n < 32; n++) | 429 | 181M | { | 430 | 181M | qmfs->v[qmfs->v_index + n] = qmfs->v[qmfs->v_index + 640 + n] = -x1[n] + x2[n]; | 431 | 181M | qmfs->v[qmfs->v_index + 63 - n] = qmfs->v[qmfs->v_index + 640 + 63 - n] = x1[n] + x2[n]; | 432 | 181M | } | 433 | | | 434 | | /* calculate 32 output samples and window */ | 435 | 187M | for (k = 0; k < 32; k++) | 436 | 181M | { | 437 | 181M | output[out++] = MUL_F(qmfs->v[qmfs->v_index + k], qmf_c[2*k]) + | 438 | 181M | MUL_F(qmfs->v[qmfs->v_index + 96 + k], qmf_c[64 + 2*k]) + | 439 | 181M | MUL_F(qmfs->v[qmfs->v_index + 128 + k], qmf_c[128 + 2*k]) + | 440 | 181M | MUL_F(qmfs->v[qmfs->v_index + 224 + k], qmf_c[192 + 2*k]) + | 441 | 181M | MUL_F(qmfs->v[qmfs->v_index + 256 + k], qmf_c[256 + 2*k]) + | 442 | 181M | MUL_F(qmfs->v[qmfs->v_index + 352 + k], qmf_c[320 + 2*k]) + | 443 | 181M | MUL_F(qmfs->v[qmfs->v_index + 384 + k], qmf_c[384 + 2*k]) + | 444 | 181M | MUL_F(qmfs->v[qmfs->v_index + 480 + k], qmf_c[448 + 2*k]) + | 445 | 181M | MUL_F(qmfs->v[qmfs->v_index + 512 + k], qmf_c[512 + 2*k]) + | 446 | 181M | MUL_F(qmfs->v[qmfs->v_index + 608 + k], qmf_c[576 + 2*k]); | 447 | 181M | } | 448 | | | 449 | | /* update ringbuffer index */ | 450 | 5.68M | qmfs->v_index -= 64; | 451 | 5.68M | if (qmfs->v_index < 0) | 452 | 671k | qmfs->v_index = (640 - 64); | 453 | 5.68M | } | 454 | 179k | } |
|
455 | | |
456 | | void sbr_qmf_synthesis_64(sbr_info *sbr, qmfs_info *qmfs, qmf_t X[MAX_NTSRHFG][64], |
457 | | real_t *output) |
458 | 664k | { |
459 | | // ALIGN real_t x1[64], x2[64]; |
460 | 664k | #ifndef SBR_LOW_POWER |
461 | 664k | ALIGN real_t in_real1[32], in_imag1[32], out_real1[32], out_imag1[32]; |
462 | 664k | ALIGN real_t in_real2[32], in_imag2[32], out_real2[32], out_imag2[32]; |
463 | 664k | #endif |
464 | 664k | qmf_t * pX; |
465 | 664k | real_t * pring_buffer_1, * pring_buffer_3; |
466 | | // real_t * ptemp_1, * ptemp_2; |
467 | | #ifdef PREFER_POINTERS |
468 | | // These pointers are used if target platform has autoinc address generators |
469 | | real_t * pring_buffer_2, * pring_buffer_4; |
470 | | real_t * pring_buffer_5, * pring_buffer_6; |
471 | | real_t * pring_buffer_7, * pring_buffer_8; |
472 | | real_t * pring_buffer_9, * pring_buffer_10; |
473 | | const real_t * pqmf_c_1, * pqmf_c_2, * pqmf_c_3, * pqmf_c_4; |
474 | | const real_t * pqmf_c_5, * pqmf_c_6, * pqmf_c_7, * pqmf_c_8; |
475 | | const real_t * pqmf_c_9, * pqmf_c_10; |
476 | | #endif // #ifdef PREFER_POINTERS |
477 | | #ifndef FIXED_POINT |
478 | | real_t scale = 1.f/64.f; |
479 | | #endif |
480 | 664k | int32_t n, k, out = 0; |
481 | 664k | uint8_t l; |
482 | | |
483 | | |
484 | | /* qmf subsample l */ |
485 | 21.6M | for (l = 0; l < sbr->numTimeSlotsRate; l++) |
486 | 20.9M | { |
487 | | /* shift buffer v */ |
488 | | /* buffer is not shifted, we use double ringbuffer */ |
489 | | //memmove(qmfs->v + 128, qmfs->v, (1280-128)*sizeof(real_t)); |
490 | | |
491 | | /* calculate 128 samples */ |
492 | | #ifndef FIXED_POINT |
493 | | |
494 | | pX = X[l]; |
495 | | |
496 | 10.4M | in_imag1[31] = scale*QMF_RE(pX[1]); |
497 | 10.4M | in_real1[0] = scale*QMF_RE(pX[0]); |
498 | 10.4M | in_imag2[31] = scale*QMF_IM(pX[63-1]); |
499 | 10.4M | in_real2[0] = scale*QMF_IM(pX[63-0]); |
500 | 325M | for (k = 1; k < 31; k++) |
501 | 314M | { |
502 | 314M | in_imag1[31 - k] = scale*QMF_RE(pX[2*k + 1]); |
503 | 314M | in_real1[ k] = scale*QMF_RE(pX[2*k ]); |
504 | 314M | in_imag2[31 - k] = scale*QMF_IM(pX[63 - (2*k + 1)]); |
505 | 314M | in_real2[ k] = scale*QMF_IM(pX[63 - (2*k )]); |
506 | 314M | } |
507 | 10.4M | in_imag1[0] = scale*QMF_RE(pX[63]); |
508 | 10.4M | in_real1[31] = scale*QMF_RE(pX[62]); |
509 | 10.4M | in_imag2[0] = scale*QMF_IM(pX[63-63]); |
510 | 10.4M | in_real2[31] = scale*QMF_IM(pX[63-62]); |
511 | | |
512 | | #else |
513 | | |
514 | | pX = X[l]; |
515 | | |
516 | 10.4M | in_imag1[31] = QMF_RE(pX[1]) >> 1; |
517 | 10.4M | in_real1[0] = QMF_RE(pX[0]) >> 1; |
518 | 10.4M | in_imag2[31] = QMF_IM(pX[62]) >> 1; |
519 | 10.4M | in_real2[0] = QMF_IM(pX[63]) >> 1; |
520 | 325M | for (k = 1; k < 31; k++) |
521 | 314M | { |
522 | 314M | in_imag1[31 - k] = QMF_RE(pX[2*k + 1]) >> 1; |
523 | 314M | in_real1[ k] = QMF_RE(pX[2*k ]) >> 1; |
524 | 314M | in_imag2[31 - k] = QMF_IM(pX[63 - (2*k + 1)]) >> 1; |
525 | 314M | in_real2[ k] = QMF_IM(pX[63 - (2*k )]) >> 1; |
526 | 314M | } |
527 | 10.4M | in_imag1[0] = QMF_RE(pX[63]) >> 1; |
528 | 10.4M | in_real1[31] = QMF_RE(pX[62]) >> 1; |
529 | 10.4M | in_imag2[0] = QMF_IM(pX[0]) >> 1; |
530 | 10.4M | in_real2[31] = QMF_IM(pX[1]) >> 1; |
531 | | |
532 | | #endif |
533 | | |
534 | | |
535 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT |
536 | 20.9M | dct4_kernel(in_real1, in_imag1, out_real1, out_imag1); |
537 | 20.9M | dct4_kernel(in_real2, in_imag2, out_real2, out_imag2); |
538 | | |
539 | | |
540 | 20.9M | pring_buffer_1 = qmfs->v + qmfs->v_index; |
541 | 20.9M | pring_buffer_3 = pring_buffer_1 + 1280; |
542 | | #ifdef PREFER_POINTERS |
543 | | pring_buffer_2 = pring_buffer_1 + 127; |
544 | | pring_buffer_4 = pring_buffer_1 + (1280 + 127); |
545 | | #endif // #ifdef PREFER_POINTERS |
546 | | // ptemp_1 = x1; |
547 | | // ptemp_2 = x2; |
548 | | #ifdef PREFER_POINTERS |
549 | | for (n = 0; n < 32; n ++) |
550 | | { |
551 | | //real_t x1 = *ptemp_1++; |
552 | | //real_t x2 = *ptemp_2++; |
553 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer |
554 | | *pring_buffer_1++ = *pring_buffer_3++ = out_real2[n] - out_real1[n]; |
555 | | *pring_buffer_2-- = *pring_buffer_4-- = out_real2[n] + out_real1[n]; |
556 | | //x1 = *ptemp_1++; |
557 | | //x2 = *ptemp_2++; |
558 | | *pring_buffer_1++ = *pring_buffer_3++ = out_imag2[31-n] + out_imag1[31-n]; |
559 | | *pring_buffer_2-- = *pring_buffer_4-- = out_imag2[31-n] - out_imag1[31-n]; |
560 | | } |
561 | | #else // #ifdef PREFER_POINTERS |
562 | | |
563 | 692M | for (n = 0; n < 32; n++) |
564 | 671M | { |
565 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer |
566 | 671M | pring_buffer_1[2*n] = pring_buffer_3[2*n] = out_real2[n] - out_real1[n]; |
567 | 671M | pring_buffer_1[127-2*n] = pring_buffer_3[127-2*n] = out_real2[n] + out_real1[n]; |
568 | 671M | pring_buffer_1[2*n+1] = pring_buffer_3[2*n+1] = out_imag2[31-n] + out_imag1[31-n]; |
569 | 671M | pring_buffer_1[127-(2*n+1)] = pring_buffer_3[127-(2*n+1)] = out_imag2[31-n] - out_imag1[31-n]; |
570 | 671M | } |
571 | | |
572 | 20.9M | #endif // #ifdef PREFER_POINTERS |
573 | | |
574 | 20.9M | pring_buffer_1 = qmfs->v + qmfs->v_index; |
575 | | #ifdef PREFER_POINTERS |
576 | | pring_buffer_2 = pring_buffer_1 + 192; |
577 | | pring_buffer_3 = pring_buffer_1 + 256; |
578 | | pring_buffer_4 = pring_buffer_1 + (256 + 192); |
579 | | pring_buffer_5 = pring_buffer_1 + 512; |
580 | | pring_buffer_6 = pring_buffer_1 + (512 + 192); |
581 | | pring_buffer_7 = pring_buffer_1 + 768; |
582 | | pring_buffer_8 = pring_buffer_1 + (768 + 192); |
583 | | pring_buffer_9 = pring_buffer_1 + 1024; |
584 | | pring_buffer_10 = pring_buffer_1 + (1024 + 192); |
585 | | pqmf_c_1 = qmf_c; |
586 | | pqmf_c_2 = qmf_c + 64; |
587 | | pqmf_c_3 = qmf_c + 128; |
588 | | pqmf_c_4 = qmf_c + 192; |
589 | | pqmf_c_5 = qmf_c + 256; |
590 | | pqmf_c_6 = qmf_c + 320; |
591 | | pqmf_c_7 = qmf_c + 384; |
592 | | pqmf_c_8 = qmf_c + 448; |
593 | | pqmf_c_9 = qmf_c + 512; |
594 | | pqmf_c_10 = qmf_c + 576; |
595 | | #endif // #ifdef PREFER_POINTERS |
596 | | |
597 | | /* calculate 64 output samples and window */ |
598 | 1.36G | for (k = 0; k < 64; k++) |
599 | 1.34G | { |
600 | | #ifdef PREFER_POINTERS |
601 | | output[out++] = |
602 | | MUL_F(*pring_buffer_1++, *pqmf_c_1++) + |
603 | | MUL_F(*pring_buffer_2++, *pqmf_c_2++) + |
604 | | MUL_F(*pring_buffer_3++, *pqmf_c_3++) + |
605 | | MUL_F(*pring_buffer_4++, *pqmf_c_4++) + |
606 | | MUL_F(*pring_buffer_5++, *pqmf_c_5++) + |
607 | | MUL_F(*pring_buffer_6++, *pqmf_c_6++) + |
608 | | MUL_F(*pring_buffer_7++, *pqmf_c_7++) + |
609 | | MUL_F(*pring_buffer_8++, *pqmf_c_8++) + |
610 | | MUL_F(*pring_buffer_9++, *pqmf_c_9++) + |
611 | | MUL_F(*pring_buffer_10++, *pqmf_c_10++); |
612 | | #else // #ifdef PREFER_POINTERS |
613 | 1.34G | output[out++] = |
614 | 1.34G | MUL_F(pring_buffer_1[k+0], qmf_c[k+0]) + |
615 | 1.34G | MUL_F(pring_buffer_1[k+192], qmf_c[k+64]) + |
616 | 1.34G | MUL_F(pring_buffer_1[k+256], qmf_c[k+128]) + |
617 | 1.34G | MUL_F(pring_buffer_1[k+(256+192)], qmf_c[k+192]) + |
618 | 1.34G | MUL_F(pring_buffer_1[k+512], qmf_c[k+256]) + |
619 | 1.34G | MUL_F(pring_buffer_1[k+(512+192)], qmf_c[k+320]) + |
620 | 1.34G | MUL_F(pring_buffer_1[k+768], qmf_c[k+384]) + |
621 | 1.34G | MUL_F(pring_buffer_1[k+(768+192)], qmf_c[k+448]) + |
622 | 1.34G | MUL_F(pring_buffer_1[k+1024], qmf_c[k+512]) + |
623 | 1.34G | MUL_F(pring_buffer_1[k+(1024+192)], qmf_c[k+576]); |
624 | 1.34G | #endif // #ifdef PREFER_POINTERS |
625 | 1.34G | } |
626 | | |
627 | | /* update ringbuffer index */ |
628 | 20.9M | qmfs->v_index -= 128; |
629 | 20.9M | if (qmfs->v_index < 0) |
630 | 2.46M | qmfs->v_index = (1280 - 128); |
631 | 20.9M | } |
632 | 664k | } Line | Count | Source | 458 | 332k | { | 459 | | // ALIGN real_t x1[64], x2[64]; | 460 | 332k | #ifndef SBR_LOW_POWER | 461 | 332k | ALIGN real_t in_real1[32], in_imag1[32], out_real1[32], out_imag1[32]; | 462 | 332k | ALIGN real_t in_real2[32], in_imag2[32], out_real2[32], out_imag2[32]; | 463 | 332k | #endif | 464 | 332k | qmf_t * pX; | 465 | 332k | real_t * pring_buffer_1, * pring_buffer_3; | 466 | | // real_t * ptemp_1, * ptemp_2; | 467 | | #ifdef PREFER_POINTERS | 468 | | // These pointers are used if target platform has autoinc address generators | 469 | | real_t * pring_buffer_2, * pring_buffer_4; | 470 | | real_t * pring_buffer_5, * pring_buffer_6; | 471 | | real_t * pring_buffer_7, * pring_buffer_8; | 472 | | real_t * pring_buffer_9, * pring_buffer_10; | 473 | | const real_t * pqmf_c_1, * pqmf_c_2, * pqmf_c_3, * pqmf_c_4; | 474 | | const real_t * pqmf_c_5, * pqmf_c_6, * pqmf_c_7, * pqmf_c_8; | 475 | | const real_t * pqmf_c_9, * pqmf_c_10; | 476 | | #endif // #ifdef PREFER_POINTERS | 477 | | #ifndef FIXED_POINT | 478 | | real_t scale = 1.f/64.f; | 479 | | #endif | 480 | 332k | int32_t n, k, out = 0; | 481 | 332k | uint8_t l; | 482 | | | 483 | | | 484 | | /* qmf subsample l */ | 485 | 10.8M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 486 | 10.4M | { | 487 | | /* shift buffer v */ | 488 | | /* buffer is not shifted, we use double ringbuffer */ | 489 | | //memmove(qmfs->v + 128, qmfs->v, (1280-128)*sizeof(real_t)); | 490 | | | 491 | | /* calculate 128 samples */ | 492 | | #ifndef FIXED_POINT | 493 | | | 494 | | pX = X[l]; | 495 | | | 496 | | in_imag1[31] = scale*QMF_RE(pX[1]); | 497 | | in_real1[0] = scale*QMF_RE(pX[0]); | 498 | | in_imag2[31] = scale*QMF_IM(pX[63-1]); | 499 | | in_real2[0] = scale*QMF_IM(pX[63-0]); | 500 | | for (k = 1; k < 31; k++) | 501 | | { | 502 | | in_imag1[31 - k] = scale*QMF_RE(pX[2*k + 1]); | 503 | | in_real1[ k] = scale*QMF_RE(pX[2*k ]); | 504 | | in_imag2[31 - k] = scale*QMF_IM(pX[63 - (2*k + 1)]); | 505 | | in_real2[ k] = scale*QMF_IM(pX[63 - (2*k )]); | 506 | | } | 507 | | in_imag1[0] = scale*QMF_RE(pX[63]); | 508 | | in_real1[31] = scale*QMF_RE(pX[62]); | 509 | | in_imag2[0] = scale*QMF_IM(pX[63-63]); | 510 | | in_real2[31] = scale*QMF_IM(pX[63-62]); | 511 | | | 512 | | #else | 513 | | | 514 | 10.4M | pX = X[l]; | 515 | | | 516 | 10.4M | in_imag1[31] = QMF_RE(pX[1]) >> 1; | 517 | 10.4M | in_real1[0] = QMF_RE(pX[0]) >> 1; | 518 | 10.4M | in_imag2[31] = QMF_IM(pX[62]) >> 1; | 519 | 10.4M | in_real2[0] = QMF_IM(pX[63]) >> 1; | 520 | 325M | for (k = 1; k < 31; k++) | 521 | 314M | { | 522 | 314M | in_imag1[31 - k] = QMF_RE(pX[2*k + 1]) >> 1; | 523 | 314M | in_real1[ k] = QMF_RE(pX[2*k ]) >> 1; | 524 | 314M | in_imag2[31 - k] = QMF_IM(pX[63 - (2*k + 1)]) >> 1; | 525 | 314M | in_real2[ k] = QMF_IM(pX[63 - (2*k )]) >> 1; | 526 | 314M | } | 527 | 10.4M | in_imag1[0] = QMF_RE(pX[63]) >> 1; | 528 | 10.4M | in_real1[31] = QMF_RE(pX[62]) >> 1; | 529 | 10.4M | in_imag2[0] = QMF_IM(pX[0]) >> 1; | 530 | 10.4M | in_real2[31] = QMF_IM(pX[1]) >> 1; | 531 | | | 532 | 10.4M | #endif | 533 | | | 534 | | | 535 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT | 536 | 10.4M | dct4_kernel(in_real1, in_imag1, out_real1, out_imag1); | 537 | 10.4M | dct4_kernel(in_real2, in_imag2, out_real2, out_imag2); | 538 | | | 539 | | | 540 | 10.4M | pring_buffer_1 = qmfs->v + qmfs->v_index; | 541 | 10.4M | pring_buffer_3 = pring_buffer_1 + 1280; | 542 | | #ifdef PREFER_POINTERS | 543 | | pring_buffer_2 = pring_buffer_1 + 127; | 544 | | pring_buffer_4 = pring_buffer_1 + (1280 + 127); | 545 | | #endif // #ifdef PREFER_POINTERS | 546 | | // ptemp_1 = x1; | 547 | | // ptemp_2 = x2; | 548 | | #ifdef PREFER_POINTERS | 549 | | for (n = 0; n < 32; n ++) | 550 | | { | 551 | | //real_t x1 = *ptemp_1++; | 552 | | //real_t x2 = *ptemp_2++; | 553 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer | 554 | | *pring_buffer_1++ = *pring_buffer_3++ = out_real2[n] - out_real1[n]; | 555 | | *pring_buffer_2-- = *pring_buffer_4-- = out_real2[n] + out_real1[n]; | 556 | | //x1 = *ptemp_1++; | 557 | | //x2 = *ptemp_2++; | 558 | | *pring_buffer_1++ = *pring_buffer_3++ = out_imag2[31-n] + out_imag1[31-n]; | 559 | | *pring_buffer_2-- = *pring_buffer_4-- = out_imag2[31-n] - out_imag1[31-n]; | 560 | | } | 561 | | #else // #ifdef PREFER_POINTERS | 562 | | | 563 | 346M | for (n = 0; n < 32; n++) | 564 | 335M | { | 565 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer | 566 | 335M | pring_buffer_1[2*n] = pring_buffer_3[2*n] = out_real2[n] - out_real1[n]; | 567 | 335M | pring_buffer_1[127-2*n] = pring_buffer_3[127-2*n] = out_real2[n] + out_real1[n]; | 568 | 335M | pring_buffer_1[2*n+1] = pring_buffer_3[2*n+1] = out_imag2[31-n] + out_imag1[31-n]; | 569 | 335M | pring_buffer_1[127-(2*n+1)] = pring_buffer_3[127-(2*n+1)] = out_imag2[31-n] - out_imag1[31-n]; | 570 | 335M | } | 571 | | | 572 | 10.4M | #endif // #ifdef PREFER_POINTERS | 573 | | | 574 | 10.4M | pring_buffer_1 = qmfs->v + qmfs->v_index; | 575 | | #ifdef PREFER_POINTERS | 576 | | pring_buffer_2 = pring_buffer_1 + 192; | 577 | | pring_buffer_3 = pring_buffer_1 + 256; | 578 | | pring_buffer_4 = pring_buffer_1 + (256 + 192); | 579 | | pring_buffer_5 = pring_buffer_1 + 512; | 580 | | pring_buffer_6 = pring_buffer_1 + (512 + 192); | 581 | | pring_buffer_7 = pring_buffer_1 + 768; | 582 | | pring_buffer_8 = pring_buffer_1 + (768 + 192); | 583 | | pring_buffer_9 = pring_buffer_1 + 1024; | 584 | | pring_buffer_10 = pring_buffer_1 + (1024 + 192); | 585 | | pqmf_c_1 = qmf_c; | 586 | | pqmf_c_2 = qmf_c + 64; | 587 | | pqmf_c_3 = qmf_c + 128; | 588 | | pqmf_c_4 = qmf_c + 192; | 589 | | pqmf_c_5 = qmf_c + 256; | 590 | | pqmf_c_6 = qmf_c + 320; | 591 | | pqmf_c_7 = qmf_c + 384; | 592 | | pqmf_c_8 = qmf_c + 448; | 593 | | pqmf_c_9 = qmf_c + 512; | 594 | | pqmf_c_10 = qmf_c + 576; | 595 | | #endif // #ifdef PREFER_POINTERS | 596 | | | 597 | | /* calculate 64 output samples and window */ | 598 | 682M | for (k = 0; k < 64; k++) | 599 | 671M | { | 600 | | #ifdef PREFER_POINTERS | 601 | | output[out++] = | 602 | | MUL_F(*pring_buffer_1++, *pqmf_c_1++) + | 603 | | MUL_F(*pring_buffer_2++, *pqmf_c_2++) + | 604 | | MUL_F(*pring_buffer_3++, *pqmf_c_3++) + | 605 | | MUL_F(*pring_buffer_4++, *pqmf_c_4++) + | 606 | | MUL_F(*pring_buffer_5++, *pqmf_c_5++) + | 607 | | MUL_F(*pring_buffer_6++, *pqmf_c_6++) + | 608 | | MUL_F(*pring_buffer_7++, *pqmf_c_7++) + | 609 | | MUL_F(*pring_buffer_8++, *pqmf_c_8++) + | 610 | | MUL_F(*pring_buffer_9++, *pqmf_c_9++) + | 611 | | MUL_F(*pring_buffer_10++, *pqmf_c_10++); | 612 | | #else // #ifdef PREFER_POINTERS | 613 | 671M | output[out++] = | 614 | 671M | MUL_F(pring_buffer_1[k+0], qmf_c[k+0]) + | 615 | 671M | MUL_F(pring_buffer_1[k+192], qmf_c[k+64]) + | 616 | 671M | MUL_F(pring_buffer_1[k+256], qmf_c[k+128]) + | 617 | 671M | MUL_F(pring_buffer_1[k+(256+192)], qmf_c[k+192]) + | 618 | 671M | MUL_F(pring_buffer_1[k+512], qmf_c[k+256]) + | 619 | 671M | MUL_F(pring_buffer_1[k+(512+192)], qmf_c[k+320]) + | 620 | 671M | MUL_F(pring_buffer_1[k+768], qmf_c[k+384]) + | 621 | 671M | MUL_F(pring_buffer_1[k+(768+192)], qmf_c[k+448]) + | 622 | 671M | MUL_F(pring_buffer_1[k+1024], qmf_c[k+512]) + | 623 | 671M | MUL_F(pring_buffer_1[k+(1024+192)], qmf_c[k+576]); | 624 | 671M | #endif // #ifdef PREFER_POINTERS | 625 | 671M | } | 626 | | | 627 | | /* update ringbuffer index */ | 628 | 10.4M | qmfs->v_index -= 128; | 629 | 10.4M | if (qmfs->v_index < 0) | 630 | 1.23M | qmfs->v_index = (1280 - 128); | 631 | 10.4M | } | 632 | 332k | } |
Line | Count | Source | 458 | 332k | { | 459 | | // ALIGN real_t x1[64], x2[64]; | 460 | 332k | #ifndef SBR_LOW_POWER | 461 | 332k | ALIGN real_t in_real1[32], in_imag1[32], out_real1[32], out_imag1[32]; | 462 | 332k | ALIGN real_t in_real2[32], in_imag2[32], out_real2[32], out_imag2[32]; | 463 | 332k | #endif | 464 | 332k | qmf_t * pX; | 465 | 332k | real_t * pring_buffer_1, * pring_buffer_3; | 466 | | // real_t * ptemp_1, * ptemp_2; | 467 | | #ifdef PREFER_POINTERS | 468 | | // These pointers are used if target platform has autoinc address generators | 469 | | real_t * pring_buffer_2, * pring_buffer_4; | 470 | | real_t * pring_buffer_5, * pring_buffer_6; | 471 | | real_t * pring_buffer_7, * pring_buffer_8; | 472 | | real_t * pring_buffer_9, * pring_buffer_10; | 473 | | const real_t * pqmf_c_1, * pqmf_c_2, * pqmf_c_3, * pqmf_c_4; | 474 | | const real_t * pqmf_c_5, * pqmf_c_6, * pqmf_c_7, * pqmf_c_8; | 475 | | const real_t * pqmf_c_9, * pqmf_c_10; | 476 | | #endif // #ifdef PREFER_POINTERS | 477 | 332k | #ifndef FIXED_POINT | 478 | 332k | real_t scale = 1.f/64.f; | 479 | 332k | #endif | 480 | 332k | int32_t n, k, out = 0; | 481 | 332k | uint8_t l; | 482 | | | 483 | | | 484 | | /* qmf subsample l */ | 485 | 10.8M | for (l = 0; l < sbr->numTimeSlotsRate; l++) | 486 | 10.4M | { | 487 | | /* shift buffer v */ | 488 | | /* buffer is not shifted, we use double ringbuffer */ | 489 | | //memmove(qmfs->v + 128, qmfs->v, (1280-128)*sizeof(real_t)); | 490 | | | 491 | | /* calculate 128 samples */ | 492 | 10.4M | #ifndef FIXED_POINT | 493 | | | 494 | 10.4M | pX = X[l]; | 495 | | | 496 | 10.4M | in_imag1[31] = scale*QMF_RE(pX[1]); | 497 | 10.4M | in_real1[0] = scale*QMF_RE(pX[0]); | 498 | 10.4M | in_imag2[31] = scale*QMF_IM(pX[63-1]); | 499 | 10.4M | in_real2[0] = scale*QMF_IM(pX[63-0]); | 500 | 325M | for (k = 1; k < 31; k++) | 501 | 314M | { | 502 | 314M | in_imag1[31 - k] = scale*QMF_RE(pX[2*k + 1]); | 503 | 314M | in_real1[ k] = scale*QMF_RE(pX[2*k ]); | 504 | 314M | in_imag2[31 - k] = scale*QMF_IM(pX[63 - (2*k + 1)]); | 505 | 314M | in_real2[ k] = scale*QMF_IM(pX[63 - (2*k )]); | 506 | 314M | } | 507 | 10.4M | in_imag1[0] = scale*QMF_RE(pX[63]); | 508 | 10.4M | in_real1[31] = scale*QMF_RE(pX[62]); | 509 | 10.4M | in_imag2[0] = scale*QMF_IM(pX[63-63]); | 510 | 10.4M | in_real2[31] = scale*QMF_IM(pX[63-62]); | 511 | | | 512 | | #else | 513 | | | 514 | | pX = X[l]; | 515 | | | 516 | | in_imag1[31] = QMF_RE(pX[1]) >> 1; | 517 | | in_real1[0] = QMF_RE(pX[0]) >> 1; | 518 | | in_imag2[31] = QMF_IM(pX[62]) >> 1; | 519 | | in_real2[0] = QMF_IM(pX[63]) >> 1; | 520 | | for (k = 1; k < 31; k++) | 521 | | { | 522 | | in_imag1[31 - k] = QMF_RE(pX[2*k + 1]) >> 1; | 523 | | in_real1[ k] = QMF_RE(pX[2*k ]) >> 1; | 524 | | in_imag2[31 - k] = QMF_IM(pX[63 - (2*k + 1)]) >> 1; | 525 | | in_real2[ k] = QMF_IM(pX[63 - (2*k )]) >> 1; | 526 | | } | 527 | | in_imag1[0] = QMF_RE(pX[63]) >> 1; | 528 | | in_real1[31] = QMF_RE(pX[62]) >> 1; | 529 | | in_imag2[0] = QMF_IM(pX[0]) >> 1; | 530 | | in_real2[31] = QMF_IM(pX[1]) >> 1; | 531 | | | 532 | | #endif | 533 | | | 534 | | | 535 | | // dct4_kernel is DCT_IV without reordering which is done before and after FFT | 536 | 10.4M | dct4_kernel(in_real1, in_imag1, out_real1, out_imag1); | 537 | 10.4M | dct4_kernel(in_real2, in_imag2, out_real2, out_imag2); | 538 | | | 539 | | | 540 | 10.4M | pring_buffer_1 = qmfs->v + qmfs->v_index; | 541 | 10.4M | pring_buffer_3 = pring_buffer_1 + 1280; | 542 | | #ifdef PREFER_POINTERS | 543 | | pring_buffer_2 = pring_buffer_1 + 127; | 544 | | pring_buffer_4 = pring_buffer_1 + (1280 + 127); | 545 | | #endif // #ifdef PREFER_POINTERS | 546 | | // ptemp_1 = x1; | 547 | | // ptemp_2 = x2; | 548 | | #ifdef PREFER_POINTERS | 549 | | for (n = 0; n < 32; n ++) | 550 | | { | 551 | | //real_t x1 = *ptemp_1++; | 552 | | //real_t x2 = *ptemp_2++; | 553 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer | 554 | | *pring_buffer_1++ = *pring_buffer_3++ = out_real2[n] - out_real1[n]; | 555 | | *pring_buffer_2-- = *pring_buffer_4-- = out_real2[n] + out_real1[n]; | 556 | | //x1 = *ptemp_1++; | 557 | | //x2 = *ptemp_2++; | 558 | | *pring_buffer_1++ = *pring_buffer_3++ = out_imag2[31-n] + out_imag1[31-n]; | 559 | | *pring_buffer_2-- = *pring_buffer_4-- = out_imag2[31-n] - out_imag1[31-n]; | 560 | | } | 561 | | #else // #ifdef PREFER_POINTERS | 562 | | | 563 | 346M | for (n = 0; n < 32; n++) | 564 | 335M | { | 565 | | // pring_buffer_3 and pring_buffer_4 are needed only for double ring buffer | 566 | 335M | pring_buffer_1[2*n] = pring_buffer_3[2*n] = out_real2[n] - out_real1[n]; | 567 | 335M | pring_buffer_1[127-2*n] = pring_buffer_3[127-2*n] = out_real2[n] + out_real1[n]; | 568 | 335M | pring_buffer_1[2*n+1] = pring_buffer_3[2*n+1] = out_imag2[31-n] + out_imag1[31-n]; | 569 | 335M | pring_buffer_1[127-(2*n+1)] = pring_buffer_3[127-(2*n+1)] = out_imag2[31-n] - out_imag1[31-n]; | 570 | 335M | } | 571 | | | 572 | 10.4M | #endif // #ifdef PREFER_POINTERS | 573 | | | 574 | 10.4M | pring_buffer_1 = qmfs->v + qmfs->v_index; | 575 | | #ifdef PREFER_POINTERS | 576 | | pring_buffer_2 = pring_buffer_1 + 192; | 577 | | pring_buffer_3 = pring_buffer_1 + 256; | 578 | | pring_buffer_4 = pring_buffer_1 + (256 + 192); | 579 | | pring_buffer_5 = pring_buffer_1 + 512; | 580 | | pring_buffer_6 = pring_buffer_1 + (512 + 192); | 581 | | pring_buffer_7 = pring_buffer_1 + 768; | 582 | | pring_buffer_8 = pring_buffer_1 + (768 + 192); | 583 | | pring_buffer_9 = pring_buffer_1 + 1024; | 584 | | pring_buffer_10 = pring_buffer_1 + (1024 + 192); | 585 | | pqmf_c_1 = qmf_c; | 586 | | pqmf_c_2 = qmf_c + 64; | 587 | | pqmf_c_3 = qmf_c + 128; | 588 | | pqmf_c_4 = qmf_c + 192; | 589 | | pqmf_c_5 = qmf_c + 256; | 590 | | pqmf_c_6 = qmf_c + 320; | 591 | | pqmf_c_7 = qmf_c + 384; | 592 | | pqmf_c_8 = qmf_c + 448; | 593 | | pqmf_c_9 = qmf_c + 512; | 594 | | pqmf_c_10 = qmf_c + 576; | 595 | | #endif // #ifdef PREFER_POINTERS | 596 | | | 597 | | /* calculate 64 output samples and window */ | 598 | 682M | for (k = 0; k < 64; k++) | 599 | 671M | { | 600 | | #ifdef PREFER_POINTERS | 601 | | output[out++] = | 602 | | MUL_F(*pring_buffer_1++, *pqmf_c_1++) + | 603 | | MUL_F(*pring_buffer_2++, *pqmf_c_2++) + | 604 | | MUL_F(*pring_buffer_3++, *pqmf_c_3++) + | 605 | | MUL_F(*pring_buffer_4++, *pqmf_c_4++) + | 606 | | MUL_F(*pring_buffer_5++, *pqmf_c_5++) + | 607 | | MUL_F(*pring_buffer_6++, *pqmf_c_6++) + | 608 | | MUL_F(*pring_buffer_7++, *pqmf_c_7++) + | 609 | | MUL_F(*pring_buffer_8++, *pqmf_c_8++) + | 610 | | MUL_F(*pring_buffer_9++, *pqmf_c_9++) + | 611 | | MUL_F(*pring_buffer_10++, *pqmf_c_10++); | 612 | | #else // #ifdef PREFER_POINTERS | 613 | 671M | output[out++] = | 614 | 671M | MUL_F(pring_buffer_1[k+0], qmf_c[k+0]) + | 615 | 671M | MUL_F(pring_buffer_1[k+192], qmf_c[k+64]) + | 616 | 671M | MUL_F(pring_buffer_1[k+256], qmf_c[k+128]) + | 617 | 671M | MUL_F(pring_buffer_1[k+(256+192)], qmf_c[k+192]) + | 618 | 671M | MUL_F(pring_buffer_1[k+512], qmf_c[k+256]) + | 619 | 671M | MUL_F(pring_buffer_1[k+(512+192)], qmf_c[k+320]) + | 620 | 671M | MUL_F(pring_buffer_1[k+768], qmf_c[k+384]) + | 621 | 671M | MUL_F(pring_buffer_1[k+(768+192)], qmf_c[k+448]) + | 622 | 671M | MUL_F(pring_buffer_1[k+1024], qmf_c[k+512]) + | 623 | 671M | MUL_F(pring_buffer_1[k+(1024+192)], qmf_c[k+576]); | 624 | 671M | #endif // #ifdef PREFER_POINTERS | 625 | 671M | } | 626 | | | 627 | | /* update ringbuffer index */ | 628 | 10.4M | qmfs->v_index -= 128; | 629 | 10.4M | if (qmfs->v_index < 0) | 630 | 1.23M | qmfs->v_index = (1280 - 128); | 631 | 10.4M | } | 632 | 332k | } |
|
633 | | #endif |
634 | | |
635 | | #endif |