1*a58d3d2aSXin Li /* Copyright (c) 2014-2020, Cisco Systems, INC
2*a58d3d2aSXin Li Written by XiangMingZhu WeiZhou MinPeng YanWang FrancisQuiers
3*a58d3d2aSXin Li
4*a58d3d2aSXin Li Redistribution and use in source and binary forms, with or without
5*a58d3d2aSXin Li modification, are permitted provided that the following conditions
6*a58d3d2aSXin Li are met:
7*a58d3d2aSXin Li
8*a58d3d2aSXin Li - Redistributions of source code must retain the above copyright
9*a58d3d2aSXin Li notice, this list of conditions and the following disclaimer.
10*a58d3d2aSXin Li
11*a58d3d2aSXin Li - Redistributions in binary form must reproduce the above copyright
12*a58d3d2aSXin Li notice, this list of conditions and the following disclaimer in the
13*a58d3d2aSXin Li documentation and/or other materials provided with the distribution.
14*a58d3d2aSXin Li
15*a58d3d2aSXin Li THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
16*a58d3d2aSXin Li ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
17*a58d3d2aSXin Li LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
18*a58d3d2aSXin Li A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER
19*a58d3d2aSXin Li OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
20*a58d3d2aSXin Li EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
21*a58d3d2aSXin Li PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
22*a58d3d2aSXin Li PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
23*a58d3d2aSXin Li LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
24*a58d3d2aSXin Li NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
25*a58d3d2aSXin Li SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26*a58d3d2aSXin Li */
27*a58d3d2aSXin Li
28*a58d3d2aSXin Li #ifdef HAVE_CONFIG_H
29*a58d3d2aSXin Li #include "config.h"
30*a58d3d2aSXin Li #endif
31*a58d3d2aSXin Li
32*a58d3d2aSXin Li #include <xmmintrin.h>
33*a58d3d2aSXin Li #include <emmintrin.h>
34*a58d3d2aSXin Li #include <smmintrin.h>
35*a58d3d2aSXin Li #include "main.h"
36*a58d3d2aSXin Li #include "celt/x86/x86cpu.h"
37*a58d3d2aSXin Li #include "stack_alloc.h"
38*a58d3d2aSXin Li
39*a58d3d2aSXin Li static OPUS_INLINE void silk_nsq_scale_states_sse4_1(
40*a58d3d2aSXin Li const silk_encoder_state *psEncC, /* I Encoder State */
41*a58d3d2aSXin Li silk_nsq_state *NSQ, /* I/O NSQ state */
42*a58d3d2aSXin Li const opus_int16 x16[], /* I input */
43*a58d3d2aSXin Li opus_int32 x_sc_Q10[], /* O input scaled with 1/Gain */
44*a58d3d2aSXin Li const opus_int16 sLTP[], /* I re-whitened LTP state in Q0 */
45*a58d3d2aSXin Li opus_int32 sLTP_Q15[], /* O LTP state matching scaled input */
46*a58d3d2aSXin Li opus_int subfr, /* I subframe number */
47*a58d3d2aSXin Li const opus_int LTP_scale_Q14, /* I */
48*a58d3d2aSXin Li const opus_int32 Gains_Q16[ MAX_NB_SUBFR ], /* I */
49*a58d3d2aSXin Li const opus_int pitchL[ MAX_NB_SUBFR ], /* I Pitch lag */
50*a58d3d2aSXin Li const opus_int signal_type /* I Signal type */
51*a58d3d2aSXin Li );
52*a58d3d2aSXin Li
53*a58d3d2aSXin Li static OPUS_INLINE void silk_noise_shape_quantizer_10_16_sse4_1(
54*a58d3d2aSXin Li silk_nsq_state *NSQ, /* I/O NSQ state */
55*a58d3d2aSXin Li opus_int signalType, /* I Signal type */
56*a58d3d2aSXin Li const opus_int32 x_sc_Q10[], /* I */
57*a58d3d2aSXin Li opus_int8 pulses[], /* O */
58*a58d3d2aSXin Li opus_int16 xq[], /* O */
59*a58d3d2aSXin Li opus_int32 sLTP_Q15[], /* I/O LTP state */
60*a58d3d2aSXin Li const opus_int16 a_Q12[], /* I Short term prediction coefs */
61*a58d3d2aSXin Li const opus_int16 b_Q14[], /* I Long term prediction coefs */
62*a58d3d2aSXin Li const opus_int16 AR_shp_Q13[], /* I Noise shaping AR coefs */
63*a58d3d2aSXin Li opus_int lag, /* I Pitch lag */
64*a58d3d2aSXin Li opus_int32 HarmShapeFIRPacked_Q14, /* I */
65*a58d3d2aSXin Li opus_int Tilt_Q14, /* I Spectral tilt */
66*a58d3d2aSXin Li opus_int32 LF_shp_Q14, /* I */
67*a58d3d2aSXin Li opus_int32 Gain_Q16, /* I */
68*a58d3d2aSXin Li opus_int Lambda_Q10, /* I */
69*a58d3d2aSXin Li opus_int offset_Q10, /* I */
70*a58d3d2aSXin Li opus_int length, /* I Input length */
71*a58d3d2aSXin Li opus_int32 table[][4] /* I */
72*a58d3d2aSXin Li );
73*a58d3d2aSXin Li
silk_NSQ_sse4_1(const silk_encoder_state * psEncC,silk_nsq_state * NSQ,SideInfoIndices * psIndices,const opus_int16 x16[],opus_int8 pulses[],const opus_int16 * PredCoef_Q12,const opus_int16 LTPCoef_Q14[LTP_ORDER * MAX_NB_SUBFR],const opus_int16 AR_Q13[MAX_NB_SUBFR * MAX_SHAPE_LPC_ORDER],const opus_int HarmShapeGain_Q14[MAX_NB_SUBFR],const opus_int Tilt_Q14[MAX_NB_SUBFR],const opus_int32 LF_shp_Q14[MAX_NB_SUBFR],const opus_int32 Gains_Q16[MAX_NB_SUBFR],const opus_int pitchL[MAX_NB_SUBFR],const opus_int Lambda_Q10,const opus_int LTP_scale_Q14)74*a58d3d2aSXin Li void silk_NSQ_sse4_1(
75*a58d3d2aSXin Li const silk_encoder_state *psEncC, /* I Encoder State */
76*a58d3d2aSXin Li silk_nsq_state *NSQ, /* I/O NSQ state */
77*a58d3d2aSXin Li SideInfoIndices *psIndices, /* I/O Quantization Indices */
78*a58d3d2aSXin Li const opus_int16 x16[], /* I Input */
79*a58d3d2aSXin Li opus_int8 pulses[], /* O Quantized pulse signal */
80*a58d3d2aSXin Li const opus_int16 *PredCoef_Q12, /* I Short term prediction coefs */
81*a58d3d2aSXin Li const opus_int16 LTPCoef_Q14[ LTP_ORDER * MAX_NB_SUBFR ], /* I Long term prediction coefs */
82*a58d3d2aSXin Li const opus_int16 AR_Q13[ MAX_NB_SUBFR * MAX_SHAPE_LPC_ORDER ], /* I Noise shaping coefs */
83*a58d3d2aSXin Li const opus_int HarmShapeGain_Q14[ MAX_NB_SUBFR ], /* I Long term shaping coefs */
84*a58d3d2aSXin Li const opus_int Tilt_Q14[ MAX_NB_SUBFR ], /* I Spectral tilt */
85*a58d3d2aSXin Li const opus_int32 LF_shp_Q14[ MAX_NB_SUBFR ], /* I Low frequency shaping coefs */
86*a58d3d2aSXin Li const opus_int32 Gains_Q16[ MAX_NB_SUBFR ], /* I Quantization step sizes */
87*a58d3d2aSXin Li const opus_int pitchL[ MAX_NB_SUBFR ], /* I Pitch lags */
88*a58d3d2aSXin Li const opus_int Lambda_Q10, /* I Rate/distortion tradeoff */
89*a58d3d2aSXin Li const opus_int LTP_scale_Q14 /* I LTP state scaling */
90*a58d3d2aSXin Li )
91*a58d3d2aSXin Li {
92*a58d3d2aSXin Li opus_int k, lag, start_idx, LSF_interpolation_flag;
93*a58d3d2aSXin Li const opus_int16 *A_Q12, *B_Q14, *AR_shp_Q13;
94*a58d3d2aSXin Li opus_int16 *pxq;
95*a58d3d2aSXin Li VARDECL( opus_int32, sLTP_Q15 );
96*a58d3d2aSXin Li VARDECL( opus_int16, sLTP );
97*a58d3d2aSXin Li opus_int32 HarmShapeFIRPacked_Q14;
98*a58d3d2aSXin Li opus_int offset_Q10;
99*a58d3d2aSXin Li VARDECL( opus_int32, x_sc_Q10 );
100*a58d3d2aSXin Li
101*a58d3d2aSXin Li opus_int32 table[ 64 ][ 4 ];
102*a58d3d2aSXin Li opus_int32 tmp1;
103*a58d3d2aSXin Li opus_int32 q1_Q10, q2_Q10, rd1_Q20, rd2_Q20;
104*a58d3d2aSXin Li
105*a58d3d2aSXin Li #ifdef OPUS_CHECK_ASM
106*a58d3d2aSXin Li silk_nsq_state NSQ_c;
107*a58d3d2aSXin Li SideInfoIndices psIndices_c;
108*a58d3d2aSXin Li opus_int8 pulses_c[ MAX_FRAME_LENGTH ];
109*a58d3d2aSXin Li const opus_int8 *const pulses_a = pulses;
110*a58d3d2aSXin Li #endif
111*a58d3d2aSXin Li
112*a58d3d2aSXin Li SAVE_STACK;
113*a58d3d2aSXin Li
114*a58d3d2aSXin Li #ifdef OPUS_CHECK_ASM
115*a58d3d2aSXin Li ( void )pulses_a;
116*a58d3d2aSXin Li silk_memcpy( &NSQ_c, NSQ, sizeof( NSQ_c ) );
117*a58d3d2aSXin Li silk_memcpy( &psIndices_c, psIndices, sizeof( psIndices_c ) );
118*a58d3d2aSXin Li silk_assert( psEncC->nb_subfr * psEncC->subfr_length <= MAX_FRAME_LENGTH );
119*a58d3d2aSXin Li silk_memcpy( pulses_c, pulses, psEncC->nb_subfr * psEncC->subfr_length * sizeof( pulses[0] ) );
120*a58d3d2aSXin Li
121*a58d3d2aSXin Li silk_NSQ_c(
122*a58d3d2aSXin Li psEncC,
123*a58d3d2aSXin Li &NSQ_c,
124*a58d3d2aSXin Li &psIndices_c,
125*a58d3d2aSXin Li x16,
126*a58d3d2aSXin Li pulses_c,
127*a58d3d2aSXin Li PredCoef_Q12,
128*a58d3d2aSXin Li LTPCoef_Q14,
129*a58d3d2aSXin Li AR_Q13,
130*a58d3d2aSXin Li HarmShapeGain_Q14,
131*a58d3d2aSXin Li Tilt_Q14,
132*a58d3d2aSXin Li LF_shp_Q14,
133*a58d3d2aSXin Li Gains_Q16,
134*a58d3d2aSXin Li pitchL,
135*a58d3d2aSXin Li Lambda_Q10,
136*a58d3d2aSXin Li LTP_scale_Q14
137*a58d3d2aSXin Li );
138*a58d3d2aSXin Li #endif
139*a58d3d2aSXin Li
140*a58d3d2aSXin Li NSQ->rand_seed = psIndices->Seed;
141*a58d3d2aSXin Li
142*a58d3d2aSXin Li /* Set unvoiced lag to the previous one, overwrite later for voiced */
143*a58d3d2aSXin Li lag = NSQ->lagPrev;
144*a58d3d2aSXin Li
145*a58d3d2aSXin Li silk_assert( NSQ->prev_gain_Q16 != 0 );
146*a58d3d2aSXin Li
147*a58d3d2aSXin Li offset_Q10 = silk_Quantization_Offsets_Q10[ psIndices->signalType >> 1 ][ psIndices->quantOffsetType ];
148*a58d3d2aSXin Li
149*a58d3d2aSXin Li /* 0 */
150*a58d3d2aSXin Li q1_Q10 = offset_Q10;
151*a58d3d2aSXin Li q2_Q10 = offset_Q10 + ( 1024 - QUANT_LEVEL_ADJUST_Q10 );
152*a58d3d2aSXin Li rd1_Q20 = q1_Q10 * Lambda_Q10;
153*a58d3d2aSXin Li rd2_Q20 = q2_Q10 * Lambda_Q10;
154*a58d3d2aSXin Li
155*a58d3d2aSXin Li table[ 32 ][ 0 ] = q1_Q10;
156*a58d3d2aSXin Li table[ 32 ][ 1 ] = q2_Q10;
157*a58d3d2aSXin Li table[ 32 ][ 2 ] = 2 * (q1_Q10 - q2_Q10);
158*a58d3d2aSXin Li table[ 32 ][ 3 ] = (rd1_Q20 - rd2_Q20) + (q1_Q10 * q1_Q10 - q2_Q10 * q2_Q10);
159*a58d3d2aSXin Li
160*a58d3d2aSXin Li /* -1 */
161*a58d3d2aSXin Li q1_Q10 = offset_Q10 - ( 1024 - QUANT_LEVEL_ADJUST_Q10 );
162*a58d3d2aSXin Li q2_Q10 = offset_Q10;
163*a58d3d2aSXin Li rd1_Q20 = - q1_Q10 * Lambda_Q10;
164*a58d3d2aSXin Li rd2_Q20 = q2_Q10 * Lambda_Q10;
165*a58d3d2aSXin Li
166*a58d3d2aSXin Li table[ 31 ][ 0 ] = q1_Q10;
167*a58d3d2aSXin Li table[ 31 ][ 1 ] = q2_Q10;
168*a58d3d2aSXin Li table[ 31 ][ 2 ] = 2 * (q1_Q10 - q2_Q10);
169*a58d3d2aSXin Li table[ 31 ][ 3 ] = (rd1_Q20 - rd2_Q20) + (q1_Q10 * q1_Q10 - q2_Q10 * q2_Q10);
170*a58d3d2aSXin Li
171*a58d3d2aSXin Li /* > 0 */
172*a58d3d2aSXin Li for (k = 1; k <= 31; k++)
173*a58d3d2aSXin Li {
174*a58d3d2aSXin Li tmp1 = offset_Q10 + silk_LSHIFT( k, 10 );
175*a58d3d2aSXin Li
176*a58d3d2aSXin Li q1_Q10 = tmp1 - QUANT_LEVEL_ADJUST_Q10;
177*a58d3d2aSXin Li q2_Q10 = tmp1 - QUANT_LEVEL_ADJUST_Q10 + 1024;
178*a58d3d2aSXin Li rd1_Q20 = q1_Q10 * Lambda_Q10;
179*a58d3d2aSXin Li rd2_Q20 = q2_Q10 * Lambda_Q10;
180*a58d3d2aSXin Li
181*a58d3d2aSXin Li table[ 32 + k ][ 0 ] = q1_Q10;
182*a58d3d2aSXin Li table[ 32 + k ][ 1 ] = q2_Q10;
183*a58d3d2aSXin Li table[ 32 + k ][ 2 ] = 2 * (q1_Q10 - q2_Q10);
184*a58d3d2aSXin Li table[ 32 + k ][ 3 ] = (rd1_Q20 - rd2_Q20) + (q1_Q10 * q1_Q10 - q2_Q10 * q2_Q10);
185*a58d3d2aSXin Li }
186*a58d3d2aSXin Li
187*a58d3d2aSXin Li /* < -1 */
188*a58d3d2aSXin Li for (k = -32; k <= -2; k++)
189*a58d3d2aSXin Li {
190*a58d3d2aSXin Li tmp1 = offset_Q10 + silk_LSHIFT( k, 10 );
191*a58d3d2aSXin Li
192*a58d3d2aSXin Li q1_Q10 = tmp1 + QUANT_LEVEL_ADJUST_Q10;
193*a58d3d2aSXin Li q2_Q10 = tmp1 + QUANT_LEVEL_ADJUST_Q10 + 1024;
194*a58d3d2aSXin Li rd1_Q20 = - q1_Q10 * Lambda_Q10;
195*a58d3d2aSXin Li rd2_Q20 = - q2_Q10 * Lambda_Q10;
196*a58d3d2aSXin Li
197*a58d3d2aSXin Li table[ 32 + k ][ 0 ] = q1_Q10;
198*a58d3d2aSXin Li table[ 32 + k ][ 1 ] = q2_Q10;
199*a58d3d2aSXin Li table[ 32 + k ][ 2 ] = 2 * (q1_Q10 - q2_Q10);
200*a58d3d2aSXin Li table[ 32 + k ][ 3 ] = (rd1_Q20 - rd2_Q20) + (q1_Q10 * q1_Q10 - q2_Q10 * q2_Q10);
201*a58d3d2aSXin Li }
202*a58d3d2aSXin Li
203*a58d3d2aSXin Li if( psIndices->NLSFInterpCoef_Q2 == 4 ) {
204*a58d3d2aSXin Li LSF_interpolation_flag = 0;
205*a58d3d2aSXin Li } else {
206*a58d3d2aSXin Li LSF_interpolation_flag = 1;
207*a58d3d2aSXin Li }
208*a58d3d2aSXin Li
209*a58d3d2aSXin Li ALLOC( sLTP_Q15, psEncC->ltp_mem_length + psEncC->frame_length, opus_int32 );
210*a58d3d2aSXin Li ALLOC( sLTP, psEncC->ltp_mem_length + psEncC->frame_length, opus_int16 );
211*a58d3d2aSXin Li ALLOC( x_sc_Q10, psEncC->subfr_length, opus_int32 );
212*a58d3d2aSXin Li /* Set up pointers to start of sub frame */
213*a58d3d2aSXin Li NSQ->sLTP_shp_buf_idx = psEncC->ltp_mem_length;
214*a58d3d2aSXin Li NSQ->sLTP_buf_idx = psEncC->ltp_mem_length;
215*a58d3d2aSXin Li pxq = &NSQ->xq[ psEncC->ltp_mem_length ];
216*a58d3d2aSXin Li for( k = 0; k < psEncC->nb_subfr; k++ ) {
217*a58d3d2aSXin Li A_Q12 = &PredCoef_Q12[ (( k >> 1 ) | ( 1 - LSF_interpolation_flag )) * MAX_LPC_ORDER ];
218*a58d3d2aSXin Li B_Q14 = <PCoef_Q14[ k * LTP_ORDER ];
219*a58d3d2aSXin Li AR_shp_Q13 = &AR_Q13[ k * MAX_SHAPE_LPC_ORDER ];
220*a58d3d2aSXin Li
221*a58d3d2aSXin Li /* Noise shape parameters */
222*a58d3d2aSXin Li silk_assert( HarmShapeGain_Q14[ k ] >= 0 );
223*a58d3d2aSXin Li HarmShapeFIRPacked_Q14 = silk_RSHIFT( HarmShapeGain_Q14[ k ], 2 );
224*a58d3d2aSXin Li HarmShapeFIRPacked_Q14 |= silk_LSHIFT( (opus_int32)silk_RSHIFT( HarmShapeGain_Q14[ k ], 1 ), 16 );
225*a58d3d2aSXin Li
226*a58d3d2aSXin Li NSQ->rewhite_flag = 0;
227*a58d3d2aSXin Li if( psIndices->signalType == TYPE_VOICED ) {
228*a58d3d2aSXin Li /* Voiced */
229*a58d3d2aSXin Li lag = pitchL[ k ];
230*a58d3d2aSXin Li
231*a58d3d2aSXin Li /* Re-whitening */
232*a58d3d2aSXin Li if( ( k & ( 3 - silk_LSHIFT( LSF_interpolation_flag, 1 ) ) ) == 0 ) {
233*a58d3d2aSXin Li /* Rewhiten with new A coefs */
234*a58d3d2aSXin Li start_idx = psEncC->ltp_mem_length - lag - psEncC->predictLPCOrder - LTP_ORDER / 2;
235*a58d3d2aSXin Li celt_assert( start_idx > 0 );
236*a58d3d2aSXin Li
237*a58d3d2aSXin Li silk_LPC_analysis_filter( &sLTP[ start_idx ], &NSQ->xq[ start_idx + k * psEncC->subfr_length ],
238*a58d3d2aSXin Li A_Q12, psEncC->ltp_mem_length - start_idx, psEncC->predictLPCOrder, psEncC->arch );
239*a58d3d2aSXin Li
240*a58d3d2aSXin Li NSQ->rewhite_flag = 1;
241*a58d3d2aSXin Li NSQ->sLTP_buf_idx = psEncC->ltp_mem_length;
242*a58d3d2aSXin Li }
243*a58d3d2aSXin Li }
244*a58d3d2aSXin Li
245*a58d3d2aSXin Li silk_nsq_scale_states_sse4_1( psEncC, NSQ, x16, x_sc_Q10, sLTP, sLTP_Q15, k, LTP_scale_Q14, Gains_Q16, pitchL, psIndices->signalType );
246*a58d3d2aSXin Li
247*a58d3d2aSXin Li if ( opus_likely( ( 10 == psEncC->shapingLPCOrder ) && ( 16 == psEncC->predictLPCOrder) ) )
248*a58d3d2aSXin Li {
249*a58d3d2aSXin Li silk_noise_shape_quantizer_10_16_sse4_1( NSQ, psIndices->signalType, x_sc_Q10, pulses, pxq, sLTP_Q15, A_Q12, B_Q14,
250*a58d3d2aSXin Li AR_shp_Q13, lag, HarmShapeFIRPacked_Q14, Tilt_Q14[ k ], LF_shp_Q14[ k ], Gains_Q16[ k ], Lambda_Q10,
251*a58d3d2aSXin Li offset_Q10, psEncC->subfr_length, &(table[32]) );
252*a58d3d2aSXin Li }
253*a58d3d2aSXin Li else
254*a58d3d2aSXin Li {
255*a58d3d2aSXin Li silk_noise_shape_quantizer( NSQ, psIndices->signalType, x_sc_Q10, pulses, pxq, sLTP_Q15, A_Q12, B_Q14,
256*a58d3d2aSXin Li AR_shp_Q13, lag, HarmShapeFIRPacked_Q14, Tilt_Q14[ k ], LF_shp_Q14[ k ], Gains_Q16[ k ], Lambda_Q10,
257*a58d3d2aSXin Li offset_Q10, psEncC->subfr_length, psEncC->shapingLPCOrder, psEncC->predictLPCOrder, psEncC->arch );
258*a58d3d2aSXin Li }
259*a58d3d2aSXin Li
260*a58d3d2aSXin Li x16 += psEncC->subfr_length;
261*a58d3d2aSXin Li pulses += psEncC->subfr_length;
262*a58d3d2aSXin Li pxq += psEncC->subfr_length;
263*a58d3d2aSXin Li }
264*a58d3d2aSXin Li
265*a58d3d2aSXin Li /* Update lagPrev for next frame */
266*a58d3d2aSXin Li NSQ->lagPrev = pitchL[ psEncC->nb_subfr - 1 ];
267*a58d3d2aSXin Li
268*a58d3d2aSXin Li /* Save quantized speech and noise shaping signals */
269*a58d3d2aSXin Li silk_memmove( NSQ->xq, &NSQ->xq[ psEncC->frame_length ], psEncC->ltp_mem_length * sizeof( opus_int16 ) );
270*a58d3d2aSXin Li silk_memmove( NSQ->sLTP_shp_Q14, &NSQ->sLTP_shp_Q14[ psEncC->frame_length ], psEncC->ltp_mem_length * sizeof( opus_int32 ) );
271*a58d3d2aSXin Li
272*a58d3d2aSXin Li #ifdef OPUS_CHECK_ASM
273*a58d3d2aSXin Li silk_assert( !memcmp( &NSQ_c, NSQ, sizeof( NSQ_c ) ) );
274*a58d3d2aSXin Li silk_assert( !memcmp( &psIndices_c, psIndices, sizeof( psIndices_c ) ) );
275*a58d3d2aSXin Li silk_assert( !memcmp( pulses_c, pulses_a, psEncC->nb_subfr * psEncC->subfr_length * sizeof( pulses[0] ) ) );
276*a58d3d2aSXin Li #endif
277*a58d3d2aSXin Li
278*a58d3d2aSXin Li RESTORE_STACK;
279*a58d3d2aSXin Li }
280*a58d3d2aSXin Li
281*a58d3d2aSXin Li /************************************/
282*a58d3d2aSXin Li /* silk_noise_shape_quantizer_10_16 */
283*a58d3d2aSXin Li /************************************/
silk_noise_shape_quantizer_10_16_sse4_1(silk_nsq_state * NSQ,opus_int signalType,const opus_int32 x_sc_Q10[],opus_int8 pulses[],opus_int16 xq[],opus_int32 sLTP_Q15[],const opus_int16 a_Q12[],const opus_int16 b_Q14[],const opus_int16 AR_shp_Q13[],opus_int lag,opus_int32 HarmShapeFIRPacked_Q14,opus_int Tilt_Q14,opus_int32 LF_shp_Q14,opus_int32 Gain_Q16,opus_int Lambda_Q10,opus_int offset_Q10,opus_int length,opus_int32 table[][4])284*a58d3d2aSXin Li static OPUS_INLINE void silk_noise_shape_quantizer_10_16_sse4_1(
285*a58d3d2aSXin Li silk_nsq_state *NSQ, /* I/O NSQ state */
286*a58d3d2aSXin Li opus_int signalType, /* I Signal type */
287*a58d3d2aSXin Li const opus_int32 x_sc_Q10[], /* I */
288*a58d3d2aSXin Li opus_int8 pulses[], /* O */
289*a58d3d2aSXin Li opus_int16 xq[], /* O */
290*a58d3d2aSXin Li opus_int32 sLTP_Q15[], /* I/O LTP state */
291*a58d3d2aSXin Li const opus_int16 a_Q12[], /* I Short term prediction coefs */
292*a58d3d2aSXin Li const opus_int16 b_Q14[], /* I Long term prediction coefs */
293*a58d3d2aSXin Li const opus_int16 AR_shp_Q13[], /* I Noise shaping AR coefs */
294*a58d3d2aSXin Li opus_int lag, /* I Pitch lag */
295*a58d3d2aSXin Li opus_int32 HarmShapeFIRPacked_Q14, /* I */
296*a58d3d2aSXin Li opus_int Tilt_Q14, /* I Spectral tilt */
297*a58d3d2aSXin Li opus_int32 LF_shp_Q14, /* I */
298*a58d3d2aSXin Li opus_int32 Gain_Q16, /* I */
299*a58d3d2aSXin Li opus_int Lambda_Q10, /* I */
300*a58d3d2aSXin Li opus_int offset_Q10, /* I */
301*a58d3d2aSXin Li opus_int length, /* I Input length */
302*a58d3d2aSXin Li opus_int32 table[][4] /* I */
303*a58d3d2aSXin Li )
304*a58d3d2aSXin Li {
305*a58d3d2aSXin Li opus_int i;
306*a58d3d2aSXin Li opus_int32 LTP_pred_Q13, LPC_pred_Q10, n_AR_Q12, n_LTP_Q13;
307*a58d3d2aSXin Li opus_int32 n_LF_Q12, r_Q10, q1_Q0, q1_Q10, q2_Q10;
308*a58d3d2aSXin Li opus_int32 exc_Q14, LPC_exc_Q14, xq_Q14, Gain_Q10, sDiff_shp_Q14;
309*a58d3d2aSXin Li opus_int32 tmp1, tmp2, sLF_AR_shp_Q14;
310*a58d3d2aSXin Li opus_int32 *psLPC_Q14, *shp_lag_ptr, *pred_lag_ptr;
311*a58d3d2aSXin Li
312*a58d3d2aSXin Li __m128i xmm_tempa, xmm_tempb;
313*a58d3d2aSXin Li
314*a58d3d2aSXin Li __m128i xmm_one;
315*a58d3d2aSXin Li
316*a58d3d2aSXin Li __m128i psLPC_Q14_hi_01234567, psLPC_Q14_hi_89ABCDEF;
317*a58d3d2aSXin Li __m128i psLPC_Q14_lo_01234567, psLPC_Q14_lo_89ABCDEF;
318*a58d3d2aSXin Li __m128i a_Q12_01234567, a_Q12_89ABCDEF;
319*a58d3d2aSXin Li
320*a58d3d2aSXin Li __m128i sAR2_Q14_hi_76543210, sAR2_Q14_lo_76543210;
321*a58d3d2aSXin Li __m128i AR_shp_Q13_76543210;
322*a58d3d2aSXin Li
323*a58d3d2aSXin Li int rdo_offset = (Lambda_Q10 >> 1) - 512;
324*a58d3d2aSXin Li
325*a58d3d2aSXin Li shp_lag_ptr = &NSQ->sLTP_shp_Q14[ NSQ->sLTP_shp_buf_idx - lag + HARM_SHAPE_FIR_TAPS / 2 ];
326*a58d3d2aSXin Li pred_lag_ptr = &sLTP_Q15[ NSQ->sLTP_buf_idx - lag + LTP_ORDER / 2 ];
327*a58d3d2aSXin Li Gain_Q10 = silk_RSHIFT( Gain_Q16, 6 );
328*a58d3d2aSXin Li
329*a58d3d2aSXin Li /* Set up short term AR state */
330*a58d3d2aSXin Li psLPC_Q14 = &NSQ->sLPC_Q14[ NSQ_LPC_BUF_LENGTH - 1 ];
331*a58d3d2aSXin Li
332*a58d3d2aSXin Li sLF_AR_shp_Q14 = NSQ->sLF_AR_shp_Q14;
333*a58d3d2aSXin Li xq_Q14 = psLPC_Q14[ 0 ];
334*a58d3d2aSXin Li sDiff_shp_Q14 = NSQ->sDiff_shp_Q14;
335*a58d3d2aSXin Li LTP_pred_Q13 = 0;
336*a58d3d2aSXin Li
337*a58d3d2aSXin Li /* load a_Q12 */
338*a58d3d2aSXin Li xmm_one = _mm_set_epi8( 1, 0, 3, 2, 5, 4, 7, 6, 9, 8, 11, 10, 13, 12, 15, 14 );
339*a58d3d2aSXin Li
340*a58d3d2aSXin Li /* load a_Q12[0] - a_Q12[7] */
341*a58d3d2aSXin Li a_Q12_01234567 = _mm_loadu_si128( (__m128i *)(void*)(&a_Q12[ 0 ] ) );
342*a58d3d2aSXin Li /* load a_Q12[ 8 ] - a_Q12[ 15 ] */
343*a58d3d2aSXin Li a_Q12_89ABCDEF = _mm_loadu_si128( (__m128i *)(void*)(&a_Q12[ 8 ] ) );
344*a58d3d2aSXin Li
345*a58d3d2aSXin Li a_Q12_01234567 = _mm_shuffle_epi8( a_Q12_01234567, xmm_one );
346*a58d3d2aSXin Li a_Q12_89ABCDEF = _mm_shuffle_epi8( a_Q12_89ABCDEF, xmm_one );
347*a58d3d2aSXin Li
348*a58d3d2aSXin Li /* load AR_shp_Q13 */
349*a58d3d2aSXin Li AR_shp_Q13_76543210 = _mm_loadu_si128( (__m128i *)(void*)(&AR_shp_Q13[0] ) );
350*a58d3d2aSXin Li
351*a58d3d2aSXin Li /* load psLPC_Q14 */
352*a58d3d2aSXin Li xmm_one = _mm_set_epi8(15, 14, 11, 10, 7, 6, 3, 2, 13, 12, 9, 8, 5, 4, 1, 0 );
353*a58d3d2aSXin Li
354*a58d3d2aSXin Li xmm_tempa = _mm_loadu_si128( (__m128i *)(void*)(&psLPC_Q14[-16]) );
355*a58d3d2aSXin Li xmm_tempb = _mm_loadu_si128( (__m128i *)(void*)(&psLPC_Q14[-12]) );
356*a58d3d2aSXin Li
357*a58d3d2aSXin Li xmm_tempa = _mm_shuffle_epi8( xmm_tempa, xmm_one );
358*a58d3d2aSXin Li xmm_tempb = _mm_shuffle_epi8( xmm_tempb, xmm_one );
359*a58d3d2aSXin Li
360*a58d3d2aSXin Li psLPC_Q14_hi_89ABCDEF = _mm_unpackhi_epi64( xmm_tempa, xmm_tempb );
361*a58d3d2aSXin Li psLPC_Q14_lo_89ABCDEF = _mm_unpacklo_epi64( xmm_tempa, xmm_tempb );
362*a58d3d2aSXin Li
363*a58d3d2aSXin Li xmm_tempa = _mm_loadu_si128( (__m128i *)(void*)(&psLPC_Q14[ -8 ]) );
364*a58d3d2aSXin Li xmm_tempb = _mm_loadu_si128( (__m128i *)(void*)(&psLPC_Q14[ -4 ]) );
365*a58d3d2aSXin Li
366*a58d3d2aSXin Li xmm_tempa = _mm_shuffle_epi8( xmm_tempa, xmm_one );
367*a58d3d2aSXin Li xmm_tempb = _mm_shuffle_epi8( xmm_tempb, xmm_one );
368*a58d3d2aSXin Li
369*a58d3d2aSXin Li psLPC_Q14_hi_01234567 = _mm_unpackhi_epi64( xmm_tempa, xmm_tempb );
370*a58d3d2aSXin Li psLPC_Q14_lo_01234567 = _mm_unpacklo_epi64( xmm_tempa, xmm_tempb );
371*a58d3d2aSXin Li
372*a58d3d2aSXin Li /* load sAR2_Q14 */
373*a58d3d2aSXin Li xmm_tempa = _mm_loadu_si128( (__m128i *)(void*)(&(NSQ->sAR2_Q14[ 0 ]) ) );
374*a58d3d2aSXin Li xmm_tempb = _mm_loadu_si128( (__m128i *)(void*)(&(NSQ->sAR2_Q14[ 4 ]) ) );
375*a58d3d2aSXin Li
376*a58d3d2aSXin Li xmm_tempa = _mm_shuffle_epi8( xmm_tempa, xmm_one );
377*a58d3d2aSXin Li xmm_tempb = _mm_shuffle_epi8( xmm_tempb, xmm_one );
378*a58d3d2aSXin Li
379*a58d3d2aSXin Li sAR2_Q14_hi_76543210 = _mm_unpackhi_epi64( xmm_tempa, xmm_tempb );
380*a58d3d2aSXin Li sAR2_Q14_lo_76543210 = _mm_unpacklo_epi64( xmm_tempa, xmm_tempb );
381*a58d3d2aSXin Li
382*a58d3d2aSXin Li /* prepare 1 in 8 * 16bit */
383*a58d3d2aSXin Li xmm_one = _mm_set1_epi16(1);
384*a58d3d2aSXin Li
385*a58d3d2aSXin Li for( i = 0; i < length; i++ )
386*a58d3d2aSXin Li {
387*a58d3d2aSXin Li /* Short-term prediction */
388*a58d3d2aSXin Li __m128i xmm_hi_07, xmm_hi_8F, xmm_lo_07, xmm_lo_8F;
389*a58d3d2aSXin Li
390*a58d3d2aSXin Li /* Avoids introducing a bias because silk_SMLAWB() always rounds to -inf */
391*a58d3d2aSXin Li LPC_pred_Q10 = 8; /* silk_RSHIFT( predictLPCOrder, 1 ); */
392*a58d3d2aSXin Li
393*a58d3d2aSXin Li /* shift psLPC_Q14 */
394*a58d3d2aSXin Li psLPC_Q14_hi_89ABCDEF = _mm_alignr_epi8( psLPC_Q14_hi_01234567, psLPC_Q14_hi_89ABCDEF, 2 );
395*a58d3d2aSXin Li psLPC_Q14_lo_89ABCDEF = _mm_alignr_epi8( psLPC_Q14_lo_01234567, psLPC_Q14_lo_89ABCDEF, 2 );
396*a58d3d2aSXin Li
397*a58d3d2aSXin Li psLPC_Q14_hi_01234567 = _mm_srli_si128( psLPC_Q14_hi_01234567, 2 );
398*a58d3d2aSXin Li psLPC_Q14_lo_01234567 = _mm_srli_si128( psLPC_Q14_lo_01234567, 2 );
399*a58d3d2aSXin Li
400*a58d3d2aSXin Li psLPC_Q14_hi_01234567 = _mm_insert_epi16( psLPC_Q14_hi_01234567, (xq_Q14 >> 16), 7 );
401*a58d3d2aSXin Li psLPC_Q14_lo_01234567 = _mm_insert_epi16( psLPC_Q14_lo_01234567, (xq_Q14), 7 );
402*a58d3d2aSXin Li
403*a58d3d2aSXin Li /* high part, use pmaddwd, results in 4 32-bit */
404*a58d3d2aSXin Li xmm_hi_07 = _mm_madd_epi16( psLPC_Q14_hi_01234567, a_Q12_01234567 );
405*a58d3d2aSXin Li xmm_hi_8F = _mm_madd_epi16( psLPC_Q14_hi_89ABCDEF, a_Q12_89ABCDEF );
406*a58d3d2aSXin Li
407*a58d3d2aSXin Li /* low part, use pmulhw, results in 8 16-bit, note we need simulate unsigned * signed, _mm_srai_epi16(psLPC_Q14_lo_01234567, 15) */
408*a58d3d2aSXin Li xmm_tempa = _mm_cmpgt_epi16( _mm_setzero_si128(), psLPC_Q14_lo_01234567 );
409*a58d3d2aSXin Li xmm_tempb = _mm_cmpgt_epi16( _mm_setzero_si128(), psLPC_Q14_lo_89ABCDEF );
410*a58d3d2aSXin Li
411*a58d3d2aSXin Li xmm_tempa = _mm_and_si128( xmm_tempa, a_Q12_01234567 );
412*a58d3d2aSXin Li xmm_tempb = _mm_and_si128( xmm_tempb, a_Q12_89ABCDEF );
413*a58d3d2aSXin Li
414*a58d3d2aSXin Li xmm_lo_07 = _mm_mulhi_epi16( psLPC_Q14_lo_01234567, a_Q12_01234567 );
415*a58d3d2aSXin Li xmm_lo_8F = _mm_mulhi_epi16( psLPC_Q14_lo_89ABCDEF, a_Q12_89ABCDEF );
416*a58d3d2aSXin Li
417*a58d3d2aSXin Li xmm_lo_07 = _mm_add_epi16( xmm_lo_07, xmm_tempa );
418*a58d3d2aSXin Li xmm_lo_8F = _mm_add_epi16( xmm_lo_8F, xmm_tempb );
419*a58d3d2aSXin Li
420*a58d3d2aSXin Li xmm_lo_07 = _mm_madd_epi16( xmm_lo_07, xmm_one );
421*a58d3d2aSXin Li xmm_lo_8F = _mm_madd_epi16( xmm_lo_8F, xmm_one );
422*a58d3d2aSXin Li
423*a58d3d2aSXin Li /* accumulate */
424*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, xmm_hi_8F );
425*a58d3d2aSXin Li xmm_lo_07 = _mm_add_epi32( xmm_lo_07, xmm_lo_8F );
426*a58d3d2aSXin Li
427*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, xmm_lo_07 );
428*a58d3d2aSXin Li
429*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, _mm_unpackhi_epi64(xmm_hi_07, xmm_hi_07 ) );
430*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, _mm_shufflelo_epi16(xmm_hi_07, 0x0E ) );
431*a58d3d2aSXin Li
432*a58d3d2aSXin Li LPC_pred_Q10 += _mm_cvtsi128_si32( xmm_hi_07 );
433*a58d3d2aSXin Li
434*a58d3d2aSXin Li /* Long-term prediction */
435*a58d3d2aSXin Li if ( opus_likely( signalType == TYPE_VOICED ) ) {
436*a58d3d2aSXin Li /* Unrolled loop */
437*a58d3d2aSXin Li /* Avoids introducing a bias because silk_SMLAWB() always rounds to -inf */
438*a58d3d2aSXin Li LTP_pred_Q13 = 2;
439*a58d3d2aSXin Li {
440*a58d3d2aSXin Li __m128i b_Q14_3210, b_Q14_0123, pred_lag_ptr_0123;
441*a58d3d2aSXin Li
442*a58d3d2aSXin Li b_Q14_3210 = OP_CVTEPI16_EPI32_M64( b_Q14 );
443*a58d3d2aSXin Li b_Q14_0123 = _mm_shuffle_epi32( b_Q14_3210, 0x1B );
444*a58d3d2aSXin Li
445*a58d3d2aSXin Li /* loaded: [0] [-1] [-2] [-3] */
446*a58d3d2aSXin Li pred_lag_ptr_0123 = _mm_loadu_si128( (__m128i *)(void*)(&pred_lag_ptr[ -3 ] ) );
447*a58d3d2aSXin Li /* shuffle to [-3] [-2] [-1] [0] and to new xmm */
448*a58d3d2aSXin Li xmm_tempa = _mm_shuffle_epi32( pred_lag_ptr_0123, 0x1B );
449*a58d3d2aSXin Li /*64-bit multiply, a[2] * b[-2], a[0] * b[0] */
450*a58d3d2aSXin Li xmm_tempa = _mm_mul_epi32( xmm_tempa, b_Q14_3210 );
451*a58d3d2aSXin Li /* right shift 2 bytes (16 bits), zero extended */
452*a58d3d2aSXin Li xmm_tempa = _mm_srli_si128( xmm_tempa, 2 );
453*a58d3d2aSXin Li
454*a58d3d2aSXin Li /* a[1] * b[-1], a[3] * b[-3] */
455*a58d3d2aSXin Li pred_lag_ptr_0123 = _mm_mul_epi32( pred_lag_ptr_0123, b_Q14_0123 );
456*a58d3d2aSXin Li pred_lag_ptr_0123 = _mm_srli_si128( pred_lag_ptr_0123, 2 );
457*a58d3d2aSXin Li
458*a58d3d2aSXin Li pred_lag_ptr_0123 = _mm_add_epi32( pred_lag_ptr_0123, xmm_tempa );
459*a58d3d2aSXin Li /* equal shift right 8 bytes*/
460*a58d3d2aSXin Li xmm_tempa = _mm_shuffle_epi32( pred_lag_ptr_0123, _MM_SHUFFLE( 0, 0, 3, 2 ) );
461*a58d3d2aSXin Li xmm_tempa = _mm_add_epi32( xmm_tempa, pred_lag_ptr_0123 );
462*a58d3d2aSXin Li
463*a58d3d2aSXin Li LTP_pred_Q13 += _mm_cvtsi128_si32( xmm_tempa );
464*a58d3d2aSXin Li
465*a58d3d2aSXin Li LTP_pred_Q13 = silk_SMLAWB( LTP_pred_Q13, pred_lag_ptr[ -4 ], b_Q14[ 4 ] );
466*a58d3d2aSXin Li pred_lag_ptr++;
467*a58d3d2aSXin Li }
468*a58d3d2aSXin Li }
469*a58d3d2aSXin Li
470*a58d3d2aSXin Li /* Noise shape feedback */
471*a58d3d2aSXin Li NSQ->sAR2_Q14[ 9 ] = NSQ->sAR2_Q14[ 8 ];
472*a58d3d2aSXin Li NSQ->sAR2_Q14[ 8 ] = _mm_cvtsi128_si32( _mm_srli_si128(_mm_unpackhi_epi16( sAR2_Q14_lo_76543210, sAR2_Q14_hi_76543210 ), 12 ) );
473*a58d3d2aSXin Li
474*a58d3d2aSXin Li sAR2_Q14_hi_76543210 = _mm_slli_si128( sAR2_Q14_hi_76543210, 2 );
475*a58d3d2aSXin Li sAR2_Q14_lo_76543210 = _mm_slli_si128( sAR2_Q14_lo_76543210, 2 );
476*a58d3d2aSXin Li
477*a58d3d2aSXin Li sAR2_Q14_hi_76543210 = _mm_insert_epi16( sAR2_Q14_hi_76543210, (sDiff_shp_Q14 >> 16), 0 );
478*a58d3d2aSXin Li sAR2_Q14_lo_76543210 = _mm_insert_epi16( sAR2_Q14_lo_76543210, (sDiff_shp_Q14), 0 );
479*a58d3d2aSXin Li
480*a58d3d2aSXin Li /* high part, use pmaddwd, results in 4 32-bit */
481*a58d3d2aSXin Li xmm_hi_07 = _mm_madd_epi16( sAR2_Q14_hi_76543210, AR_shp_Q13_76543210 );
482*a58d3d2aSXin Li
483*a58d3d2aSXin Li /* low part, use pmulhw, results in 8 16-bit, note we need simulate unsigned * signed,_mm_srai_epi16(sAR2_Q14_lo_76543210, 15) */
484*a58d3d2aSXin Li xmm_tempa = _mm_cmpgt_epi16( _mm_setzero_si128(), sAR2_Q14_lo_76543210 );
485*a58d3d2aSXin Li xmm_tempa = _mm_and_si128( xmm_tempa, AR_shp_Q13_76543210 );
486*a58d3d2aSXin Li
487*a58d3d2aSXin Li xmm_lo_07 = _mm_mulhi_epi16( sAR2_Q14_lo_76543210, AR_shp_Q13_76543210 );
488*a58d3d2aSXin Li xmm_lo_07 = _mm_add_epi16( xmm_lo_07, xmm_tempa );
489*a58d3d2aSXin Li
490*a58d3d2aSXin Li xmm_lo_07 = _mm_madd_epi16( xmm_lo_07, xmm_one );
491*a58d3d2aSXin Li
492*a58d3d2aSXin Li /* accumulate */
493*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, xmm_lo_07 );
494*a58d3d2aSXin Li
495*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, _mm_unpackhi_epi64(xmm_hi_07, xmm_hi_07 ) );
496*a58d3d2aSXin Li xmm_hi_07 = _mm_add_epi32( xmm_hi_07, _mm_shufflelo_epi16(xmm_hi_07, 0x0E ) );
497*a58d3d2aSXin Li
498*a58d3d2aSXin Li n_AR_Q12 = 5 + _mm_cvtsi128_si32( xmm_hi_07 );
499*a58d3d2aSXin Li
500*a58d3d2aSXin Li n_AR_Q12 = silk_SMLAWB( n_AR_Q12, NSQ->sAR2_Q14[ 8 ], AR_shp_Q13[ 8 ] );
501*a58d3d2aSXin Li n_AR_Q12 = silk_SMLAWB( n_AR_Q12, NSQ->sAR2_Q14[ 9 ], AR_shp_Q13[ 9 ] );
502*a58d3d2aSXin Li
503*a58d3d2aSXin Li n_AR_Q12 = silk_LSHIFT32( n_AR_Q12, 1 ); /* Q11 -> Q12 */
504*a58d3d2aSXin Li n_AR_Q12 = silk_SMLAWB( n_AR_Q12, sLF_AR_shp_Q14, Tilt_Q14 );
505*a58d3d2aSXin Li
506*a58d3d2aSXin Li n_LF_Q12 = silk_SMULWB( NSQ->sLTP_shp_Q14[ NSQ->sLTP_shp_buf_idx - 1 ], LF_shp_Q14 );
507*a58d3d2aSXin Li n_LF_Q12 = silk_SMLAWT( n_LF_Q12, sLF_AR_shp_Q14, LF_shp_Q14 );
508*a58d3d2aSXin Li
509*a58d3d2aSXin Li celt_assert( lag > 0 || signalType != TYPE_VOICED );
510*a58d3d2aSXin Li
511*a58d3d2aSXin Li /* Combine prediction and noise shaping signals */
512*a58d3d2aSXin Li tmp1 = silk_SUB32( silk_LSHIFT32( LPC_pred_Q10, 2 ), n_AR_Q12 ); /* Q12 */
513*a58d3d2aSXin Li tmp1 = silk_SUB32( tmp1, n_LF_Q12 ); /* Q12 */
514*a58d3d2aSXin Li if( lag > 0 ) {
515*a58d3d2aSXin Li /* Symmetric, packed FIR coefficients */
516*a58d3d2aSXin Li n_LTP_Q13 = silk_SMULWB( silk_ADD_SAT32( shp_lag_ptr[ 0 ], shp_lag_ptr[ -2 ] ), HarmShapeFIRPacked_Q14 );
517*a58d3d2aSXin Li n_LTP_Q13 = silk_SMLAWT( n_LTP_Q13, shp_lag_ptr[ -1 ], HarmShapeFIRPacked_Q14 );
518*a58d3d2aSXin Li n_LTP_Q13 = silk_LSHIFT( n_LTP_Q13, 1 );
519*a58d3d2aSXin Li shp_lag_ptr++;
520*a58d3d2aSXin Li
521*a58d3d2aSXin Li tmp2 = silk_SUB32( LTP_pred_Q13, n_LTP_Q13 ); /* Q13 */
522*a58d3d2aSXin Li tmp1 = silk_ADD_LSHIFT32( tmp2, tmp1, 1 ); /* Q13 */
523*a58d3d2aSXin Li tmp1 = silk_RSHIFT_ROUND( tmp1, 3 ); /* Q10 */
524*a58d3d2aSXin Li } else {
525*a58d3d2aSXin Li tmp1 = silk_RSHIFT_ROUND( tmp1, 2 ); /* Q10 */
526*a58d3d2aSXin Li }
527*a58d3d2aSXin Li
528*a58d3d2aSXin Li r_Q10 = silk_SUB32( x_sc_Q10[ i ], tmp1 ); /* residual error Q10 */
529*a58d3d2aSXin Li
530*a58d3d2aSXin Li /* Generate dither */
531*a58d3d2aSXin Li NSQ->rand_seed = silk_RAND( NSQ->rand_seed );
532*a58d3d2aSXin Li
533*a58d3d2aSXin Li /* Flip sign depending on dither */
534*a58d3d2aSXin Li tmp2 = -r_Q10;
535*a58d3d2aSXin Li if ( NSQ->rand_seed < 0 ) r_Q10 = tmp2;
536*a58d3d2aSXin Li
537*a58d3d2aSXin Li r_Q10 = silk_LIMIT_32( r_Q10, -(31 << 10), 30 << 10 );
538*a58d3d2aSXin Li
539*a58d3d2aSXin Li /* Find two quantization level candidates and measure their rate-distortion */
540*a58d3d2aSXin Li q1_Q10 = silk_SUB32( r_Q10, offset_Q10 );
541*a58d3d2aSXin Li q1_Q0 = silk_RSHIFT( q1_Q10, 10 );
542*a58d3d2aSXin Li if (Lambda_Q10 > 2048) {
543*a58d3d2aSXin Li /* For aggressive RDO, the bias becomes more than one pulse. */
544*a58d3d2aSXin Li if (q1_Q10 > rdo_offset) {
545*a58d3d2aSXin Li q1_Q0 = silk_RSHIFT( q1_Q10 - rdo_offset, 10 );
546*a58d3d2aSXin Li } else if (q1_Q10 < -rdo_offset) {
547*a58d3d2aSXin Li q1_Q0 = silk_RSHIFT( q1_Q10 + rdo_offset, 10 );
548*a58d3d2aSXin Li } else if (q1_Q10 < 0) {
549*a58d3d2aSXin Li q1_Q0 = -1;
550*a58d3d2aSXin Li } else {
551*a58d3d2aSXin Li q1_Q0 = 0;
552*a58d3d2aSXin Li }
553*a58d3d2aSXin Li }
554*a58d3d2aSXin Li
555*a58d3d2aSXin Li q1_Q10 = table[q1_Q0][0];
556*a58d3d2aSXin Li q2_Q10 = table[q1_Q0][1];
557*a58d3d2aSXin Li
558*a58d3d2aSXin Li if (r_Q10 * table[q1_Q0][2] - table[q1_Q0][3] < 0)
559*a58d3d2aSXin Li {
560*a58d3d2aSXin Li q1_Q10 = q2_Q10;
561*a58d3d2aSXin Li }
562*a58d3d2aSXin Li
563*a58d3d2aSXin Li pulses[ i ] = (opus_int8)silk_RSHIFT_ROUND( q1_Q10, 10 );
564*a58d3d2aSXin Li
565*a58d3d2aSXin Li /* Excitation */
566*a58d3d2aSXin Li exc_Q14 = silk_LSHIFT( q1_Q10, 4 );
567*a58d3d2aSXin Li
568*a58d3d2aSXin Li tmp2 = -exc_Q14;
569*a58d3d2aSXin Li if ( NSQ->rand_seed < 0 ) exc_Q14 = tmp2;
570*a58d3d2aSXin Li
571*a58d3d2aSXin Li /* Add predictions */
572*a58d3d2aSXin Li LPC_exc_Q14 = silk_ADD_LSHIFT32( exc_Q14, LTP_pred_Q13, 1 );
573*a58d3d2aSXin Li xq_Q14 = silk_ADD_LSHIFT32( LPC_exc_Q14, LPC_pred_Q10, 4 );
574*a58d3d2aSXin Li
575*a58d3d2aSXin Li /* Update states */
576*a58d3d2aSXin Li psLPC_Q14++;
577*a58d3d2aSXin Li *psLPC_Q14 = xq_Q14;
578*a58d3d2aSXin Li NSQ->sDiff_shp_Q14 = silk_SUB_LSHIFT32( xq_Q14, x_sc_Q10[ i ], 4 );
579*a58d3d2aSXin Li sLF_AR_shp_Q14 = silk_SUB_LSHIFT32( NSQ->sDiff_shp_Q14, n_AR_Q12, 2 );
580*a58d3d2aSXin Li
581*a58d3d2aSXin Li NSQ->sLTP_shp_Q14[ NSQ->sLTP_shp_buf_idx ] = silk_SUB_LSHIFT32( sLF_AR_shp_Q14, n_LF_Q12, 2 );
582*a58d3d2aSXin Li sLTP_Q15[ NSQ->sLTP_buf_idx ] = silk_LSHIFT( LPC_exc_Q14, 1 );
583*a58d3d2aSXin Li NSQ->sLTP_shp_buf_idx++;
584*a58d3d2aSXin Li NSQ->sLTP_buf_idx++;
585*a58d3d2aSXin Li
586*a58d3d2aSXin Li /* Make dither dependent on quantized signal */
587*a58d3d2aSXin Li NSQ->rand_seed = silk_ADD32_ovflw( NSQ->rand_seed, pulses[ i ] );
588*a58d3d2aSXin Li }
589*a58d3d2aSXin Li
590*a58d3d2aSXin Li NSQ->sLF_AR_shp_Q14 = sLF_AR_shp_Q14;
591*a58d3d2aSXin Li
592*a58d3d2aSXin Li /* Scale XQ back to normal level before saving */
593*a58d3d2aSXin Li psLPC_Q14 = &NSQ->sLPC_Q14[ NSQ_LPC_BUF_LENGTH ];
594*a58d3d2aSXin Li
595*a58d3d2aSXin Li /* write back sAR2_Q14 */
596*a58d3d2aSXin Li xmm_tempa = _mm_unpackhi_epi16( sAR2_Q14_lo_76543210, sAR2_Q14_hi_76543210 );
597*a58d3d2aSXin Li xmm_tempb = _mm_unpacklo_epi16( sAR2_Q14_lo_76543210, sAR2_Q14_hi_76543210 );
598*a58d3d2aSXin Li _mm_storeu_si128( (__m128i *)(void*)(&NSQ->sAR2_Q14[ 4 ]), xmm_tempa );
599*a58d3d2aSXin Li _mm_storeu_si128( (__m128i *)(void*)(&NSQ->sAR2_Q14[ 0 ]), xmm_tempb );
600*a58d3d2aSXin Li
601*a58d3d2aSXin Li /* xq[ i ] = (opus_int16)silk_SAT16( silk_RSHIFT_ROUND( silk_SMULWW( psLPC_Q14[ i ], Gain_Q10 ), 8 ) ); */
602*a58d3d2aSXin Li {
603*a58d3d2aSXin Li __m128i xmm_Gain_Q10;
604*a58d3d2aSXin Li __m128i xmm_xq_Q14_3210, xmm_xq_Q14_x3x1, xmm_xq_Q14_7654, xmm_xq_Q14_x7x5;
605*a58d3d2aSXin Li
606*a58d3d2aSXin Li /* prepare (1 << 7) in packed 4 32-bits */
607*a58d3d2aSXin Li xmm_tempa = _mm_set1_epi32( (1 << 7) );
608*a58d3d2aSXin Li
609*a58d3d2aSXin Li /* prepare Gain_Q10 in packed 4 32-bits */
610*a58d3d2aSXin Li xmm_Gain_Q10 = _mm_set1_epi32( Gain_Q10 );
611*a58d3d2aSXin Li
612*a58d3d2aSXin Li /* process xq */
613*a58d3d2aSXin Li for (i = 0; i < length - 7; i += 8)
614*a58d3d2aSXin Li {
615*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_loadu_si128( (__m128i *)(void*)(&(psLPC_Q14[ i + 0 ] ) ) );
616*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_loadu_si128( (__m128i *)(void*)(&(psLPC_Q14[ i + 4 ] ) ) );
617*a58d3d2aSXin Li
618*a58d3d2aSXin Li /* equal shift right 4 bytes*/
619*a58d3d2aSXin Li xmm_xq_Q14_x3x1 = _mm_shuffle_epi32( xmm_xq_Q14_3210, _MM_SHUFFLE( 0, 3, 2, 1 ) );
620*a58d3d2aSXin Li /* equal shift right 4 bytes*/
621*a58d3d2aSXin Li xmm_xq_Q14_x7x5 = _mm_shuffle_epi32( xmm_xq_Q14_7654, _MM_SHUFFLE( 0, 3, 2, 1 ) );
622*a58d3d2aSXin Li
623*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_mul_epi32( xmm_xq_Q14_3210, xmm_Gain_Q10 );
624*a58d3d2aSXin Li xmm_xq_Q14_x3x1 = _mm_mul_epi32( xmm_xq_Q14_x3x1, xmm_Gain_Q10 );
625*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_mul_epi32( xmm_xq_Q14_7654, xmm_Gain_Q10 );
626*a58d3d2aSXin Li xmm_xq_Q14_x7x5 = _mm_mul_epi32( xmm_xq_Q14_x7x5, xmm_Gain_Q10 );
627*a58d3d2aSXin Li
628*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_srli_epi64( xmm_xq_Q14_3210, 16 );
629*a58d3d2aSXin Li xmm_xq_Q14_x3x1 = _mm_slli_epi64( xmm_xq_Q14_x3x1, 16 );
630*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_srli_epi64( xmm_xq_Q14_7654, 16 );
631*a58d3d2aSXin Li xmm_xq_Q14_x7x5 = _mm_slli_epi64( xmm_xq_Q14_x7x5, 16 );
632*a58d3d2aSXin Li
633*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_blend_epi16( xmm_xq_Q14_3210, xmm_xq_Q14_x3x1, 0xCC );
634*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_blend_epi16( xmm_xq_Q14_7654, xmm_xq_Q14_x7x5, 0xCC );
635*a58d3d2aSXin Li
636*a58d3d2aSXin Li /* silk_RSHIFT_ROUND(xq, 8) */
637*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_add_epi32( xmm_xq_Q14_3210, xmm_tempa );
638*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_add_epi32( xmm_xq_Q14_7654, xmm_tempa );
639*a58d3d2aSXin Li
640*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_srai_epi32( xmm_xq_Q14_3210, 8 );
641*a58d3d2aSXin Li xmm_xq_Q14_7654 = _mm_srai_epi32( xmm_xq_Q14_7654, 8 );
642*a58d3d2aSXin Li
643*a58d3d2aSXin Li /* silk_SAT16 */
644*a58d3d2aSXin Li xmm_xq_Q14_3210 = _mm_packs_epi32( xmm_xq_Q14_3210, xmm_xq_Q14_7654 );
645*a58d3d2aSXin Li
646*a58d3d2aSXin Li /* save to xq */
647*a58d3d2aSXin Li _mm_storeu_si128( (__m128i *)(void*)(&xq[ i ] ), xmm_xq_Q14_3210 );
648*a58d3d2aSXin Li }
649*a58d3d2aSXin Li }
650*a58d3d2aSXin Li for ( ; i < length; i++)
651*a58d3d2aSXin Li {
652*a58d3d2aSXin Li xq[i] = (opus_int16)silk_SAT16( silk_RSHIFT_ROUND( silk_SMULWW( psLPC_Q14[ i ], Gain_Q10 ), 8 ) );
653*a58d3d2aSXin Li }
654*a58d3d2aSXin Li
655*a58d3d2aSXin Li /* Update LPC synth buffer */
656*a58d3d2aSXin Li silk_memcpy( NSQ->sLPC_Q14, &NSQ->sLPC_Q14[ length ], NSQ_LPC_BUF_LENGTH * sizeof( opus_int32 ) );
657*a58d3d2aSXin Li }
658*a58d3d2aSXin Li
silk_nsq_scale_states_sse4_1(const silk_encoder_state * psEncC,silk_nsq_state * NSQ,const opus_int16 x16[],opus_int32 x_sc_Q10[],const opus_int16 sLTP[],opus_int32 sLTP_Q15[],opus_int subfr,const opus_int LTP_scale_Q14,const opus_int32 Gains_Q16[MAX_NB_SUBFR],const opus_int pitchL[MAX_NB_SUBFR],const opus_int signal_type)659*a58d3d2aSXin Li static OPUS_INLINE void silk_nsq_scale_states_sse4_1(
660*a58d3d2aSXin Li const silk_encoder_state *psEncC, /* I Encoder State */
661*a58d3d2aSXin Li silk_nsq_state *NSQ, /* I/O NSQ state */
662*a58d3d2aSXin Li const opus_int16 x16[], /* I input */
663*a58d3d2aSXin Li opus_int32 x_sc_Q10[], /* O input scaled with 1/Gain */
664*a58d3d2aSXin Li const opus_int16 sLTP[], /* I re-whitened LTP state in Q0 */
665*a58d3d2aSXin Li opus_int32 sLTP_Q15[], /* O LTP state matching scaled input */
666*a58d3d2aSXin Li opus_int subfr, /* I subframe number */
667*a58d3d2aSXin Li const opus_int LTP_scale_Q14, /* I */
668*a58d3d2aSXin Li const opus_int32 Gains_Q16[ MAX_NB_SUBFR ], /* I */
669*a58d3d2aSXin Li const opus_int pitchL[ MAX_NB_SUBFR ], /* I Pitch lag */
670*a58d3d2aSXin Li const opus_int signal_type /* I Signal type */
671*a58d3d2aSXin Li )
672*a58d3d2aSXin Li {
673*a58d3d2aSXin Li opus_int i, lag;
674*a58d3d2aSXin Li opus_int32 gain_adj_Q16, inv_gain_Q31, inv_gain_Q26;
675*a58d3d2aSXin Li __m128i xmm_inv_gain_Q26, xmm_x16_x2x0, xmm_x16_x3x1;
676*a58d3d2aSXin Li
677*a58d3d2aSXin Li lag = pitchL[ subfr ];
678*a58d3d2aSXin Li inv_gain_Q31 = silk_INVERSE32_varQ( silk_max( Gains_Q16[ subfr ], 1 ), 47 );
679*a58d3d2aSXin Li silk_assert( inv_gain_Q31 != 0 );
680*a58d3d2aSXin Li
681*a58d3d2aSXin Li /* Scale input */
682*a58d3d2aSXin Li inv_gain_Q26 = silk_RSHIFT_ROUND( inv_gain_Q31, 5 );
683*a58d3d2aSXin Li
684*a58d3d2aSXin Li /* prepare inv_gain_Q26 in packed 4 32-bits */
685*a58d3d2aSXin Li xmm_inv_gain_Q26 = _mm_set1_epi32(inv_gain_Q26);
686*a58d3d2aSXin Li
687*a58d3d2aSXin Li for( i = 0; i < psEncC->subfr_length - 3; i += 4 ) {
688*a58d3d2aSXin Li xmm_x16_x2x0 = OP_CVTEPI16_EPI32_M64( &(x16[ i ] ) );
689*a58d3d2aSXin Li
690*a58d3d2aSXin Li /* equal shift right 4 bytes*/
691*a58d3d2aSXin Li xmm_x16_x3x1 = _mm_shuffle_epi32( xmm_x16_x2x0, _MM_SHUFFLE( 0, 3, 2, 1 ) );
692*a58d3d2aSXin Li
693*a58d3d2aSXin Li xmm_x16_x2x0 = _mm_mul_epi32( xmm_x16_x2x0, xmm_inv_gain_Q26 );
694*a58d3d2aSXin Li xmm_x16_x3x1 = _mm_mul_epi32( xmm_x16_x3x1, xmm_inv_gain_Q26 );
695*a58d3d2aSXin Li
696*a58d3d2aSXin Li xmm_x16_x2x0 = _mm_srli_epi64( xmm_x16_x2x0, 16 );
697*a58d3d2aSXin Li xmm_x16_x3x1 = _mm_slli_epi64( xmm_x16_x3x1, 16 );
698*a58d3d2aSXin Li
699*a58d3d2aSXin Li xmm_x16_x2x0 = _mm_blend_epi16( xmm_x16_x2x0, xmm_x16_x3x1, 0xCC );
700*a58d3d2aSXin Li
701*a58d3d2aSXin Li _mm_storeu_si128( (__m128i *)(void*)(&(x_sc_Q10[ i ] ) ), xmm_x16_x2x0 );
702*a58d3d2aSXin Li }
703*a58d3d2aSXin Li
704*a58d3d2aSXin Li for( ; i < psEncC->subfr_length; i++ ) {
705*a58d3d2aSXin Li x_sc_Q10[ i ] = silk_SMULWW( x16[ i ], inv_gain_Q26 );
706*a58d3d2aSXin Li }
707*a58d3d2aSXin Li
708*a58d3d2aSXin Li /* After rewhitening the LTP state is un-scaled, so scale with inv_gain_Q16 */
709*a58d3d2aSXin Li if( NSQ->rewhite_flag ) {
710*a58d3d2aSXin Li if( subfr == 0 ) {
711*a58d3d2aSXin Li /* Do LTP downscaling */
712*a58d3d2aSXin Li inv_gain_Q31 = silk_LSHIFT( silk_SMULWB( inv_gain_Q31, LTP_scale_Q14 ), 2 );
713*a58d3d2aSXin Li }
714*a58d3d2aSXin Li for( i = NSQ->sLTP_buf_idx - lag - LTP_ORDER / 2; i < NSQ->sLTP_buf_idx; i++ ) {
715*a58d3d2aSXin Li silk_assert( i < MAX_FRAME_LENGTH );
716*a58d3d2aSXin Li sLTP_Q15[ i ] = silk_SMULWB( inv_gain_Q31, sLTP[ i ] );
717*a58d3d2aSXin Li }
718*a58d3d2aSXin Li }
719*a58d3d2aSXin Li
720*a58d3d2aSXin Li /* Adjust for changing gain */
721*a58d3d2aSXin Li if( Gains_Q16[ subfr ] != NSQ->prev_gain_Q16 ) {
722*a58d3d2aSXin Li __m128i xmm_gain_adj_Q16, xmm_sLTP_shp_Q14_x2x0, xmm_sLTP_shp_Q14_x3x1;
723*a58d3d2aSXin Li gain_adj_Q16 = silk_DIV32_varQ( NSQ->prev_gain_Q16, Gains_Q16[ subfr ], 16 );
724*a58d3d2aSXin Li
725*a58d3d2aSXin Li /* Scale long-term shaping state */
726*a58d3d2aSXin Li
727*a58d3d2aSXin Li /* prepare gain_adj_Q16 in packed 4 32-bits */
728*a58d3d2aSXin Li xmm_gain_adj_Q16 = _mm_set1_epi32(gain_adj_Q16);
729*a58d3d2aSXin Li
730*a58d3d2aSXin Li for( i = NSQ->sLTP_shp_buf_idx - psEncC->ltp_mem_length; i < NSQ->sLTP_shp_buf_idx - 3; i += 4 )
731*a58d3d2aSXin Li {
732*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x2x0 = _mm_loadu_si128( (__m128i *)(void*)(&(NSQ->sLTP_shp_Q14[ i ] ) ) );
733*a58d3d2aSXin Li /* equal shift right 4 bytes*/
734*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x3x1 = _mm_shuffle_epi32( xmm_sLTP_shp_Q14_x2x0, _MM_SHUFFLE( 0, 3, 2, 1 ) );
735*a58d3d2aSXin Li
736*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x2x0 = _mm_mul_epi32( xmm_sLTP_shp_Q14_x2x0, xmm_gain_adj_Q16 );
737*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x3x1 = _mm_mul_epi32( xmm_sLTP_shp_Q14_x3x1, xmm_gain_adj_Q16 );
738*a58d3d2aSXin Li
739*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x2x0 = _mm_srli_epi64( xmm_sLTP_shp_Q14_x2x0, 16 );
740*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x3x1 = _mm_slli_epi64( xmm_sLTP_shp_Q14_x3x1, 16 );
741*a58d3d2aSXin Li
742*a58d3d2aSXin Li xmm_sLTP_shp_Q14_x2x0 = _mm_blend_epi16( xmm_sLTP_shp_Q14_x2x0, xmm_sLTP_shp_Q14_x3x1, 0xCC );
743*a58d3d2aSXin Li
744*a58d3d2aSXin Li _mm_storeu_si128( (__m128i *)(void*)(&(NSQ->sLTP_shp_Q14[ i ] ) ), xmm_sLTP_shp_Q14_x2x0 );
745*a58d3d2aSXin Li }
746*a58d3d2aSXin Li
747*a58d3d2aSXin Li for( ; i < NSQ->sLTP_shp_buf_idx; i++ ) {
748*a58d3d2aSXin Li NSQ->sLTP_shp_Q14[ i ] = silk_SMULWW( gain_adj_Q16, NSQ->sLTP_shp_Q14[ i ] );
749*a58d3d2aSXin Li }
750*a58d3d2aSXin Li
751*a58d3d2aSXin Li /* Scale long-term prediction state */
752*a58d3d2aSXin Li if( signal_type == TYPE_VOICED && NSQ->rewhite_flag == 0 ) {
753*a58d3d2aSXin Li for( i = NSQ->sLTP_buf_idx - lag - LTP_ORDER / 2; i < NSQ->sLTP_buf_idx; i++ ) {
754*a58d3d2aSXin Li sLTP_Q15[ i ] = silk_SMULWW( gain_adj_Q16, sLTP_Q15[ i ] );
755*a58d3d2aSXin Li }
756*a58d3d2aSXin Li }
757*a58d3d2aSXin Li
758*a58d3d2aSXin Li NSQ->sLF_AR_shp_Q14 = silk_SMULWW( gain_adj_Q16, NSQ->sLF_AR_shp_Q14 );
759*a58d3d2aSXin Li NSQ->sDiff_shp_Q14 = silk_SMULWW( gain_adj_Q16, NSQ->sDiff_shp_Q14 );
760*a58d3d2aSXin Li
761*a58d3d2aSXin Li /* Scale short-term prediction and shaping states */
762*a58d3d2aSXin Li for( i = 0; i < NSQ_LPC_BUF_LENGTH; i++ ) {
763*a58d3d2aSXin Li NSQ->sLPC_Q14[ i ] = silk_SMULWW( gain_adj_Q16, NSQ->sLPC_Q14[ i ] );
764*a58d3d2aSXin Li }
765*a58d3d2aSXin Li for( i = 0; i < MAX_SHAPE_LPC_ORDER; i++ ) {
766*a58d3d2aSXin Li NSQ->sAR2_Q14[ i ] = silk_SMULWW( gain_adj_Q16, NSQ->sAR2_Q14[ i ] );
767*a58d3d2aSXin Li }
768*a58d3d2aSXin Li
769*a58d3d2aSXin Li /* Save inverse gain */
770*a58d3d2aSXin Li NSQ->prev_gain_Q16 = Gains_Q16[ subfr ];
771*a58d3d2aSXin Li }
772*a58d3d2aSXin Li }
773