xref: /aosp_15_r20/external/webp/src/dsp/lossless_neon.c (revision b2055c353e87c8814eb2b6b1b11112a1562253bd)
1*b2055c35SXin Li // Copyright 2014 Google Inc. All Rights Reserved.
2*b2055c35SXin Li //
3*b2055c35SXin Li // Use of this source code is governed by a BSD-style license
4*b2055c35SXin Li // that can be found in the COPYING file in the root of the source
5*b2055c35SXin Li // tree. An additional intellectual property rights grant can be found
6*b2055c35SXin Li // in the file PATENTS. All contributing project authors may
7*b2055c35SXin Li // be found in the AUTHORS file in the root of the source tree.
8*b2055c35SXin Li // -----------------------------------------------------------------------------
9*b2055c35SXin Li //
10*b2055c35SXin Li // NEON variant of methods for lossless decoder
11*b2055c35SXin Li //
12*b2055c35SXin Li // Author: Skal ([email protected])
13*b2055c35SXin Li 
14*b2055c35SXin Li #include "src/dsp/dsp.h"
15*b2055c35SXin Li 
16*b2055c35SXin Li #if defined(WEBP_USE_NEON)
17*b2055c35SXin Li 
18*b2055c35SXin Li #include <arm_neon.h>
19*b2055c35SXin Li 
20*b2055c35SXin Li #include "src/dsp/lossless.h"
21*b2055c35SXin Li #include "src/dsp/neon.h"
22*b2055c35SXin Li 
23*b2055c35SXin Li //------------------------------------------------------------------------------
24*b2055c35SXin Li // Colorspace conversion functions
25*b2055c35SXin Li 
26*b2055c35SXin Li #if !defined(WORK_AROUND_GCC)
27*b2055c35SXin Li // gcc 4.6.0 had some trouble (NDK-r9) with this code. We only use it for
28*b2055c35SXin Li // gcc-4.8.x at least.
ConvertBGRAToRGBA_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)29*b2055c35SXin Li static void ConvertBGRAToRGBA_NEON(const uint32_t* src,
30*b2055c35SXin Li                                    int num_pixels, uint8_t* dst) {
31*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~15);
32*b2055c35SXin Li   for (; src < end; src += 16) {
33*b2055c35SXin Li     uint8x16x4_t pixel = vld4q_u8((uint8_t*)src);
34*b2055c35SXin Li     // swap B and R. (VSWP d0,d2 has no intrinsics equivalent!)
35*b2055c35SXin Li     const uint8x16_t tmp = pixel.val[0];
36*b2055c35SXin Li     pixel.val[0] = pixel.val[2];
37*b2055c35SXin Li     pixel.val[2] = tmp;
38*b2055c35SXin Li     vst4q_u8(dst, pixel);
39*b2055c35SXin Li     dst += 64;
40*b2055c35SXin Li   }
41*b2055c35SXin Li   VP8LConvertBGRAToRGBA_C(src, num_pixels & 15, dst);  // left-overs
42*b2055c35SXin Li }
43*b2055c35SXin Li 
ConvertBGRAToBGR_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)44*b2055c35SXin Li static void ConvertBGRAToBGR_NEON(const uint32_t* src,
45*b2055c35SXin Li                                   int num_pixels, uint8_t* dst) {
46*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~15);
47*b2055c35SXin Li   for (; src < end; src += 16) {
48*b2055c35SXin Li     const uint8x16x4_t pixel = vld4q_u8((uint8_t*)src);
49*b2055c35SXin Li     const uint8x16x3_t tmp = { { pixel.val[0], pixel.val[1], pixel.val[2] } };
50*b2055c35SXin Li     vst3q_u8(dst, tmp);
51*b2055c35SXin Li     dst += 48;
52*b2055c35SXin Li   }
53*b2055c35SXin Li   VP8LConvertBGRAToBGR_C(src, num_pixels & 15, dst);  // left-overs
54*b2055c35SXin Li }
55*b2055c35SXin Li 
ConvertBGRAToRGB_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)56*b2055c35SXin Li static void ConvertBGRAToRGB_NEON(const uint32_t* src,
57*b2055c35SXin Li                                   int num_pixels, uint8_t* dst) {
58*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~15);
59*b2055c35SXin Li   for (; src < end; src += 16) {
60*b2055c35SXin Li     const uint8x16x4_t pixel = vld4q_u8((uint8_t*)src);
61*b2055c35SXin Li     const uint8x16x3_t tmp = { { pixel.val[2], pixel.val[1], pixel.val[0] } };
62*b2055c35SXin Li     vst3q_u8(dst, tmp);
63*b2055c35SXin Li     dst += 48;
64*b2055c35SXin Li   }
65*b2055c35SXin Li   VP8LConvertBGRAToRGB_C(src, num_pixels & 15, dst);  // left-overs
66*b2055c35SXin Li }
67*b2055c35SXin Li 
68*b2055c35SXin Li #else  // WORK_AROUND_GCC
69*b2055c35SXin Li 
70*b2055c35SXin Li // gcc-4.6.0 fallback
71*b2055c35SXin Li 
72*b2055c35SXin Li static const uint8_t kRGBAShuffle[8] = { 2, 1, 0, 3, 6, 5, 4, 7 };
73*b2055c35SXin Li 
ConvertBGRAToRGBA_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)74*b2055c35SXin Li static void ConvertBGRAToRGBA_NEON(const uint32_t* src,
75*b2055c35SXin Li                                    int num_pixels, uint8_t* dst) {
76*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~1);
77*b2055c35SXin Li   const uint8x8_t shuffle = vld1_u8(kRGBAShuffle);
78*b2055c35SXin Li   for (; src < end; src += 2) {
79*b2055c35SXin Li     const uint8x8_t pixels = vld1_u8((uint8_t*)src);
80*b2055c35SXin Li     vst1_u8(dst, vtbl1_u8(pixels, shuffle));
81*b2055c35SXin Li     dst += 8;
82*b2055c35SXin Li   }
83*b2055c35SXin Li   VP8LConvertBGRAToRGBA_C(src, num_pixels & 1, dst);  // left-overs
84*b2055c35SXin Li }
85*b2055c35SXin Li 
86*b2055c35SXin Li static const uint8_t kBGRShuffle[3][8] = {
87*b2055c35SXin Li   {  0,  1,  2,  4,  5,  6,  8,  9 },
88*b2055c35SXin Li   { 10, 12, 13, 14, 16, 17, 18, 20 },
89*b2055c35SXin Li   { 21, 22, 24, 25, 26, 28, 29, 30 }
90*b2055c35SXin Li };
91*b2055c35SXin Li 
ConvertBGRAToBGR_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)92*b2055c35SXin Li static void ConvertBGRAToBGR_NEON(const uint32_t* src,
93*b2055c35SXin Li                                   int num_pixels, uint8_t* dst) {
94*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~7);
95*b2055c35SXin Li   const uint8x8_t shuffle0 = vld1_u8(kBGRShuffle[0]);
96*b2055c35SXin Li   const uint8x8_t shuffle1 = vld1_u8(kBGRShuffle[1]);
97*b2055c35SXin Li   const uint8x8_t shuffle2 = vld1_u8(kBGRShuffle[2]);
98*b2055c35SXin Li   for (; src < end; src += 8) {
99*b2055c35SXin Li     uint8x8x4_t pixels;
100*b2055c35SXin Li     INIT_VECTOR4(pixels,
101*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 0)),
102*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 2)),
103*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 4)),
104*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 6)));
105*b2055c35SXin Li     vst1_u8(dst +  0, vtbl4_u8(pixels, shuffle0));
106*b2055c35SXin Li     vst1_u8(dst +  8, vtbl4_u8(pixels, shuffle1));
107*b2055c35SXin Li     vst1_u8(dst + 16, vtbl4_u8(pixels, shuffle2));
108*b2055c35SXin Li     dst += 8 * 3;
109*b2055c35SXin Li   }
110*b2055c35SXin Li   VP8LConvertBGRAToBGR_C(src, num_pixels & 7, dst);  // left-overs
111*b2055c35SXin Li }
112*b2055c35SXin Li 
113*b2055c35SXin Li static const uint8_t kRGBShuffle[3][8] = {
114*b2055c35SXin Li   {  2,  1,  0,  6,  5,  4, 10,  9 },
115*b2055c35SXin Li   {  8, 14, 13, 12, 18, 17, 16, 22 },
116*b2055c35SXin Li   { 21, 20, 26, 25, 24, 30, 29, 28 }
117*b2055c35SXin Li };
118*b2055c35SXin Li 
ConvertBGRAToRGB_NEON(const uint32_t * src,int num_pixels,uint8_t * dst)119*b2055c35SXin Li static void ConvertBGRAToRGB_NEON(const uint32_t* src,
120*b2055c35SXin Li                                   int num_pixels, uint8_t* dst) {
121*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~7);
122*b2055c35SXin Li   const uint8x8_t shuffle0 = vld1_u8(kRGBShuffle[0]);
123*b2055c35SXin Li   const uint8x8_t shuffle1 = vld1_u8(kRGBShuffle[1]);
124*b2055c35SXin Li   const uint8x8_t shuffle2 = vld1_u8(kRGBShuffle[2]);
125*b2055c35SXin Li   for (; src < end; src += 8) {
126*b2055c35SXin Li     uint8x8x4_t pixels;
127*b2055c35SXin Li     INIT_VECTOR4(pixels,
128*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 0)),
129*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 2)),
130*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 4)),
131*b2055c35SXin Li                  vld1_u8((const uint8_t*)(src + 6)));
132*b2055c35SXin Li     vst1_u8(dst +  0, vtbl4_u8(pixels, shuffle0));
133*b2055c35SXin Li     vst1_u8(dst +  8, vtbl4_u8(pixels, shuffle1));
134*b2055c35SXin Li     vst1_u8(dst + 16, vtbl4_u8(pixels, shuffle2));
135*b2055c35SXin Li     dst += 8 * 3;
136*b2055c35SXin Li   }
137*b2055c35SXin Li   VP8LConvertBGRAToRGB_C(src, num_pixels & 7, dst);  // left-overs
138*b2055c35SXin Li }
139*b2055c35SXin Li 
140*b2055c35SXin Li #endif   // !WORK_AROUND_GCC
141*b2055c35SXin Li 
142*b2055c35SXin Li //------------------------------------------------------------------------------
143*b2055c35SXin Li // Predictor Transform
144*b2055c35SXin Li 
145*b2055c35SXin Li #define LOAD_U32_AS_U8(IN) vreinterpret_u8_u32(vdup_n_u32((IN)))
146*b2055c35SXin Li #define LOAD_U32P_AS_U8(IN) vreinterpret_u8_u32(vld1_u32((IN)))
147*b2055c35SXin Li #define LOADQ_U32_AS_U8(IN) vreinterpretq_u8_u32(vdupq_n_u32((IN)))
148*b2055c35SXin Li #define LOADQ_U32P_AS_U8(IN) vreinterpretq_u8_u32(vld1q_u32((IN)))
149*b2055c35SXin Li #define GET_U8_AS_U32(IN) vget_lane_u32(vreinterpret_u32_u8((IN)), 0)
150*b2055c35SXin Li #define GETQ_U8_AS_U32(IN) vgetq_lane_u32(vreinterpretq_u32_u8((IN)), 0)
151*b2055c35SXin Li #define STOREQ_U8_AS_U32P(OUT, IN) vst1q_u32((OUT), vreinterpretq_u32_u8((IN)))
152*b2055c35SXin Li #define ROTATE32_LEFT(L) vextq_u8((L), (L), 12)    // D|C|B|A -> C|B|A|D
153*b2055c35SXin Li 
Average2_u8_NEON(uint32_t a0,uint32_t a1)154*b2055c35SXin Li static WEBP_INLINE uint8x8_t Average2_u8_NEON(uint32_t a0, uint32_t a1) {
155*b2055c35SXin Li   const uint8x8_t A0 = LOAD_U32_AS_U8(a0);
156*b2055c35SXin Li   const uint8x8_t A1 = LOAD_U32_AS_U8(a1);
157*b2055c35SXin Li   return vhadd_u8(A0, A1);
158*b2055c35SXin Li }
159*b2055c35SXin Li 
ClampedAddSubtractHalf_NEON(uint32_t c0,uint32_t c1,uint32_t c2)160*b2055c35SXin Li static WEBP_INLINE uint32_t ClampedAddSubtractHalf_NEON(uint32_t c0,
161*b2055c35SXin Li                                                         uint32_t c1,
162*b2055c35SXin Li                                                         uint32_t c2) {
163*b2055c35SXin Li   const uint8x8_t avg = Average2_u8_NEON(c0, c1);
164*b2055c35SXin Li   // Remove one to c2 when bigger than avg.
165*b2055c35SXin Li   const uint8x8_t C2 = LOAD_U32_AS_U8(c2);
166*b2055c35SXin Li   const uint8x8_t cmp = vcgt_u8(C2, avg);
167*b2055c35SXin Li   const uint8x8_t C2_1 = vadd_u8(C2, cmp);
168*b2055c35SXin Li   // Compute half of the difference between avg and c2.
169*b2055c35SXin Li   const int8x8_t diff_avg = vreinterpret_s8_u8(vhsub_u8(avg, C2_1));
170*b2055c35SXin Li   // Compute the sum with avg and saturate.
171*b2055c35SXin Li   const int16x8_t avg_16 = vreinterpretq_s16_u16(vmovl_u8(avg));
172*b2055c35SXin Li   const uint8x8_t res = vqmovun_s16(vaddw_s8(avg_16, diff_avg));
173*b2055c35SXin Li   const uint32_t output = GET_U8_AS_U32(res);
174*b2055c35SXin Li   return output;
175*b2055c35SXin Li }
176*b2055c35SXin Li 
Average2_NEON(uint32_t a0,uint32_t a1)177*b2055c35SXin Li static WEBP_INLINE uint32_t Average2_NEON(uint32_t a0, uint32_t a1) {
178*b2055c35SXin Li   const uint8x8_t avg_u8x8 = Average2_u8_NEON(a0, a1);
179*b2055c35SXin Li   const uint32_t avg = GET_U8_AS_U32(avg_u8x8);
180*b2055c35SXin Li   return avg;
181*b2055c35SXin Li }
182*b2055c35SXin Li 
Average3_NEON(uint32_t a0,uint32_t a1,uint32_t a2)183*b2055c35SXin Li static WEBP_INLINE uint32_t Average3_NEON(uint32_t a0, uint32_t a1,
184*b2055c35SXin Li                                           uint32_t a2) {
185*b2055c35SXin Li   const uint8x8_t avg0 = Average2_u8_NEON(a0, a2);
186*b2055c35SXin Li   const uint8x8_t A1 = LOAD_U32_AS_U8(a1);
187*b2055c35SXin Li   const uint32_t avg = GET_U8_AS_U32(vhadd_u8(avg0, A1));
188*b2055c35SXin Li   return avg;
189*b2055c35SXin Li }
190*b2055c35SXin Li 
Predictor5_NEON(const uint32_t * const left,const uint32_t * const top)191*b2055c35SXin Li static uint32_t Predictor5_NEON(const uint32_t* const left,
192*b2055c35SXin Li                                 const uint32_t* const top) {
193*b2055c35SXin Li   return Average3_NEON(*left, top[0], top[1]);
194*b2055c35SXin Li }
Predictor6_NEON(const uint32_t * const left,const uint32_t * const top)195*b2055c35SXin Li static uint32_t Predictor6_NEON(const uint32_t* const left,
196*b2055c35SXin Li                                 const uint32_t* const top) {
197*b2055c35SXin Li   return Average2_NEON(*left, top[-1]);
198*b2055c35SXin Li }
Predictor7_NEON(const uint32_t * const left,const uint32_t * const top)199*b2055c35SXin Li static uint32_t Predictor7_NEON(const uint32_t* const left,
200*b2055c35SXin Li                                 const uint32_t* const top) {
201*b2055c35SXin Li   return Average2_NEON(*left, top[0]);
202*b2055c35SXin Li }
Predictor13_NEON(const uint32_t * const left,const uint32_t * const top)203*b2055c35SXin Li static uint32_t Predictor13_NEON(const uint32_t* const left,
204*b2055c35SXin Li                                  const uint32_t* const top) {
205*b2055c35SXin Li   return ClampedAddSubtractHalf_NEON(*left, top[0], top[-1]);
206*b2055c35SXin Li }
207*b2055c35SXin Li 
208*b2055c35SXin Li // Batch versions of those functions.
209*b2055c35SXin Li 
210*b2055c35SXin Li // Predictor0: ARGB_BLACK.
PredictorAdd0_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)211*b2055c35SXin Li static void PredictorAdd0_NEON(const uint32_t* in, const uint32_t* upper,
212*b2055c35SXin Li                                int num_pixels, uint32_t* out) {
213*b2055c35SXin Li   int i;
214*b2055c35SXin Li   const uint8x16_t black = vreinterpretq_u8_u32(vdupq_n_u32(ARGB_BLACK));
215*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
216*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
217*b2055c35SXin Li     const uint8x16_t res = vaddq_u8(src, black);
218*b2055c35SXin Li     STOREQ_U8_AS_U32P(&out[i], res);
219*b2055c35SXin Li   }
220*b2055c35SXin Li   VP8LPredictorsAdd_C[0](in + i, upper + i, num_pixels - i, out + i);
221*b2055c35SXin Li }
222*b2055c35SXin Li 
223*b2055c35SXin Li // Predictor1: left.
PredictorAdd1_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)224*b2055c35SXin Li static void PredictorAdd1_NEON(const uint32_t* in, const uint32_t* upper,
225*b2055c35SXin Li                                int num_pixels, uint32_t* out) {
226*b2055c35SXin Li   int i;
227*b2055c35SXin Li   const uint8x16_t zero = LOADQ_U32_AS_U8(0);
228*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
229*b2055c35SXin Li     // a | b | c | d
230*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
231*b2055c35SXin Li     // 0 | a | b | c
232*b2055c35SXin Li     const uint8x16_t shift0 = vextq_u8(zero, src, 12);
233*b2055c35SXin Li     // a | a + b | b + c | c + d
234*b2055c35SXin Li     const uint8x16_t sum0 = vaddq_u8(src, shift0);
235*b2055c35SXin Li     // 0 | 0 | a | a + b
236*b2055c35SXin Li     const uint8x16_t shift1 = vextq_u8(zero, sum0, 8);
237*b2055c35SXin Li     // a | a + b | a + b + c | a + b + c + d
238*b2055c35SXin Li     const uint8x16_t sum1 = vaddq_u8(sum0, shift1);
239*b2055c35SXin Li     const uint8x16_t prev = LOADQ_U32_AS_U8(out[i - 1]);
240*b2055c35SXin Li     const uint8x16_t res = vaddq_u8(sum1, prev);
241*b2055c35SXin Li     STOREQ_U8_AS_U32P(&out[i], res);
242*b2055c35SXin Li   }
243*b2055c35SXin Li   VP8LPredictorsAdd_C[1](in + i, upper + i, num_pixels - i, out + i);
244*b2055c35SXin Li }
245*b2055c35SXin Li 
246*b2055c35SXin Li // Macro that adds 32-bit integers from IN using mod 256 arithmetic
247*b2055c35SXin Li // per 8 bit channel.
248*b2055c35SXin Li #define GENERATE_PREDICTOR_1(X, IN)                                       \
249*b2055c35SXin Li static void PredictorAdd##X##_NEON(const uint32_t* in,                    \
250*b2055c35SXin Li                                    const uint32_t* upper, int num_pixels, \
251*b2055c35SXin Li                                    uint32_t* out) {                       \
252*b2055c35SXin Li   int i;                                                                  \
253*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {                              \
254*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);                      \
255*b2055c35SXin Li     const uint8x16_t other = LOADQ_U32P_AS_U8(&(IN));                     \
256*b2055c35SXin Li     const uint8x16_t res = vaddq_u8(src, other);                          \
257*b2055c35SXin Li     STOREQ_U8_AS_U32P(&out[i], res);                                      \
258*b2055c35SXin Li   }                                                                       \
259*b2055c35SXin Li   VP8LPredictorsAdd_C[(X)](in + i, upper + i, num_pixels - i, out + i);   \
260*b2055c35SXin Li }
261*b2055c35SXin Li // Predictor2: Top.
262*b2055c35SXin Li GENERATE_PREDICTOR_1(2, upper[i])
263*b2055c35SXin Li // Predictor3: Top-right.
264*b2055c35SXin Li GENERATE_PREDICTOR_1(3, upper[i + 1])
265*b2055c35SXin Li // Predictor4: Top-left.
266*b2055c35SXin Li GENERATE_PREDICTOR_1(4, upper[i - 1])
267*b2055c35SXin Li #undef GENERATE_PREDICTOR_1
268*b2055c35SXin Li 
269*b2055c35SXin Li // Predictor5: average(average(left, TR), T)
270*b2055c35SXin Li #define DO_PRED5(LANE) do {                                              \
271*b2055c35SXin Li   const uint8x16_t avgLTR = vhaddq_u8(L, TR);                            \
272*b2055c35SXin Li   const uint8x16_t avg = vhaddq_u8(avgLTR, T);                           \
273*b2055c35SXin Li   const uint8x16_t res = vaddq_u8(avg, src);                             \
274*b2055c35SXin Li   vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE));   \
275*b2055c35SXin Li   L = ROTATE32_LEFT(res);                                                \
276*b2055c35SXin Li } while (0)
277*b2055c35SXin Li 
PredictorAdd5_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)278*b2055c35SXin Li static void PredictorAdd5_NEON(const uint32_t* in, const uint32_t* upper,
279*b2055c35SXin Li                                int num_pixels, uint32_t* out) {
280*b2055c35SXin Li   int i;
281*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
282*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
283*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
284*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i + 0]);
285*b2055c35SXin Li     const uint8x16_t TR = LOADQ_U32P_AS_U8(&upper[i + 1]);
286*b2055c35SXin Li     DO_PRED5(0);
287*b2055c35SXin Li     DO_PRED5(1);
288*b2055c35SXin Li     DO_PRED5(2);
289*b2055c35SXin Li     DO_PRED5(3);
290*b2055c35SXin Li   }
291*b2055c35SXin Li   VP8LPredictorsAdd_C[5](in + i, upper + i, num_pixels - i, out + i);
292*b2055c35SXin Li }
293*b2055c35SXin Li #undef DO_PRED5
294*b2055c35SXin Li 
295*b2055c35SXin Li #define DO_PRED67(LANE) do {                                             \
296*b2055c35SXin Li   const uint8x16_t avg = vhaddq_u8(L, top);                              \
297*b2055c35SXin Li   const uint8x16_t res = vaddq_u8(avg, src);                             \
298*b2055c35SXin Li   vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE));   \
299*b2055c35SXin Li   L = ROTATE32_LEFT(res);                                                \
300*b2055c35SXin Li } while (0)
301*b2055c35SXin Li 
302*b2055c35SXin Li // Predictor6: average(left, TL)
PredictorAdd6_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)303*b2055c35SXin Li static void PredictorAdd6_NEON(const uint32_t* in, const uint32_t* upper,
304*b2055c35SXin Li                                int num_pixels, uint32_t* out) {
305*b2055c35SXin Li   int i;
306*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
307*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
308*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
309*b2055c35SXin Li     const uint8x16_t top = LOADQ_U32P_AS_U8(&upper[i - 1]);
310*b2055c35SXin Li     DO_PRED67(0);
311*b2055c35SXin Li     DO_PRED67(1);
312*b2055c35SXin Li     DO_PRED67(2);
313*b2055c35SXin Li     DO_PRED67(3);
314*b2055c35SXin Li   }
315*b2055c35SXin Li   VP8LPredictorsAdd_C[6](in + i, upper + i, num_pixels - i, out + i);
316*b2055c35SXin Li }
317*b2055c35SXin Li 
318*b2055c35SXin Li // Predictor7: average(left, T)
PredictorAdd7_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)319*b2055c35SXin Li static void PredictorAdd7_NEON(const uint32_t* in, const uint32_t* upper,
320*b2055c35SXin Li                                int num_pixels, uint32_t* out) {
321*b2055c35SXin Li   int i;
322*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
323*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
324*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
325*b2055c35SXin Li     const uint8x16_t top = LOADQ_U32P_AS_U8(&upper[i]);
326*b2055c35SXin Li     DO_PRED67(0);
327*b2055c35SXin Li     DO_PRED67(1);
328*b2055c35SXin Li     DO_PRED67(2);
329*b2055c35SXin Li     DO_PRED67(3);
330*b2055c35SXin Li   }
331*b2055c35SXin Li   VP8LPredictorsAdd_C[7](in + i, upper + i, num_pixels - i, out + i);
332*b2055c35SXin Li }
333*b2055c35SXin Li #undef DO_PRED67
334*b2055c35SXin Li 
335*b2055c35SXin Li #define GENERATE_PREDICTOR_2(X, IN)                                       \
336*b2055c35SXin Li static void PredictorAdd##X##_NEON(const uint32_t* in,                    \
337*b2055c35SXin Li                                    const uint32_t* upper, int num_pixels, \
338*b2055c35SXin Li                                    uint32_t* out) {                       \
339*b2055c35SXin Li   int i;                                                                  \
340*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {                              \
341*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);                      \
342*b2055c35SXin Li     const uint8x16_t Tother = LOADQ_U32P_AS_U8(&(IN));                    \
343*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);                     \
344*b2055c35SXin Li     const uint8x16_t avg = vhaddq_u8(T, Tother);                          \
345*b2055c35SXin Li     const uint8x16_t res = vaddq_u8(avg, src);                            \
346*b2055c35SXin Li     STOREQ_U8_AS_U32P(&out[i], res);                                      \
347*b2055c35SXin Li   }                                                                       \
348*b2055c35SXin Li   VP8LPredictorsAdd_C[(X)](in + i, upper + i, num_pixels - i, out + i);   \
349*b2055c35SXin Li }
350*b2055c35SXin Li // Predictor8: average TL T.
351*b2055c35SXin Li GENERATE_PREDICTOR_2(8, upper[i - 1])
352*b2055c35SXin Li // Predictor9: average T TR.
353*b2055c35SXin Li GENERATE_PREDICTOR_2(9, upper[i + 1])
354*b2055c35SXin Li #undef GENERATE_PREDICTOR_2
355*b2055c35SXin Li 
356*b2055c35SXin Li // Predictor10: average of (average of (L,TL), average of (T, TR)).
357*b2055c35SXin Li #define DO_PRED10(LANE) do {                                             \
358*b2055c35SXin Li   const uint8x16_t avgLTL = vhaddq_u8(L, TL);                            \
359*b2055c35SXin Li   const uint8x16_t avg = vhaddq_u8(avgTTR, avgLTL);                      \
360*b2055c35SXin Li   const uint8x16_t res = vaddq_u8(avg, src);                             \
361*b2055c35SXin Li   vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE));   \
362*b2055c35SXin Li   L = ROTATE32_LEFT(res);                                                \
363*b2055c35SXin Li } while (0)
364*b2055c35SXin Li 
PredictorAdd10_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)365*b2055c35SXin Li static void PredictorAdd10_NEON(const uint32_t* in, const uint32_t* upper,
366*b2055c35SXin Li                                 int num_pixels, uint32_t* out) {
367*b2055c35SXin Li   int i;
368*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
369*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
370*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
371*b2055c35SXin Li     const uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]);
372*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);
373*b2055c35SXin Li     const uint8x16_t TR = LOADQ_U32P_AS_U8(&upper[i + 1]);
374*b2055c35SXin Li     const uint8x16_t avgTTR = vhaddq_u8(T, TR);
375*b2055c35SXin Li     DO_PRED10(0);
376*b2055c35SXin Li     DO_PRED10(1);
377*b2055c35SXin Li     DO_PRED10(2);
378*b2055c35SXin Li     DO_PRED10(3);
379*b2055c35SXin Li   }
380*b2055c35SXin Li   VP8LPredictorsAdd_C[10](in + i, upper + i, num_pixels - i, out + i);
381*b2055c35SXin Li }
382*b2055c35SXin Li #undef DO_PRED10
383*b2055c35SXin Li 
384*b2055c35SXin Li // Predictor11: select.
385*b2055c35SXin Li #define DO_PRED11(LANE) do {                                                   \
386*b2055c35SXin Li   const uint8x16_t sumLin = vaddq_u8(L, src);  /* in + L */                    \
387*b2055c35SXin Li   const uint8x16_t pLTL = vabdq_u8(L, TL);  /* |L - TL| */                     \
388*b2055c35SXin Li   const uint16x8_t sum_LTL = vpaddlq_u8(pLTL);                                 \
389*b2055c35SXin Li   const uint32x4_t pa = vpaddlq_u16(sum_LTL);                                  \
390*b2055c35SXin Li   const uint32x4_t mask = vcleq_u32(pa, pb);                                   \
391*b2055c35SXin Li   const uint8x16_t res = vbslq_u8(vreinterpretq_u8_u32(mask), sumTin, sumLin); \
392*b2055c35SXin Li   vst1q_lane_u32(&out[i + (LANE)], vreinterpretq_u32_u8(res), (LANE));         \
393*b2055c35SXin Li   L = ROTATE32_LEFT(res);                                                      \
394*b2055c35SXin Li } while (0)
395*b2055c35SXin Li 
PredictorAdd11_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)396*b2055c35SXin Li static void PredictorAdd11_NEON(const uint32_t* in, const uint32_t* upper,
397*b2055c35SXin Li                                 int num_pixels, uint32_t* out) {
398*b2055c35SXin Li   int i;
399*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
400*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
401*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);
402*b2055c35SXin Li     const uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]);
403*b2055c35SXin Li     const uint8x16_t pTTL = vabdq_u8(T, TL);   // |T - TL|
404*b2055c35SXin Li     const uint16x8_t sum_TTL = vpaddlq_u8(pTTL);
405*b2055c35SXin Li     const uint32x4_t pb = vpaddlq_u16(sum_TTL);
406*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
407*b2055c35SXin Li     const uint8x16_t sumTin = vaddq_u8(T, src);   // in + T
408*b2055c35SXin Li     DO_PRED11(0);
409*b2055c35SXin Li     DO_PRED11(1);
410*b2055c35SXin Li     DO_PRED11(2);
411*b2055c35SXin Li     DO_PRED11(3);
412*b2055c35SXin Li   }
413*b2055c35SXin Li   VP8LPredictorsAdd_C[11](in + i, upper + i, num_pixels - i, out + i);
414*b2055c35SXin Li }
415*b2055c35SXin Li #undef DO_PRED11
416*b2055c35SXin Li 
417*b2055c35SXin Li // Predictor12: ClampedAddSubtractFull.
418*b2055c35SXin Li #define DO_PRED12(DIFF, LANE) do {                                       \
419*b2055c35SXin Li   const uint8x8_t pred =                                                 \
420*b2055c35SXin Li       vqmovun_s16(vaddq_s16(vreinterpretq_s16_u16(L), (DIFF)));          \
421*b2055c35SXin Li   const uint8x8_t res =                                                  \
422*b2055c35SXin Li       vadd_u8(pred, (LANE <= 1) ? vget_low_u8(src) : vget_high_u8(src)); \
423*b2055c35SXin Li   const uint16x8_t res16 = vmovl_u8(res);                                \
424*b2055c35SXin Li   vst1_lane_u32(&out[i + (LANE)], vreinterpret_u32_u8(res), (LANE) & 1); \
425*b2055c35SXin Li   /* rotate in the left predictor for next iteration */                  \
426*b2055c35SXin Li   L = vextq_u16(res16, res16, 4);                                        \
427*b2055c35SXin Li } while (0)
428*b2055c35SXin Li 
PredictorAdd12_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)429*b2055c35SXin Li static void PredictorAdd12_NEON(const uint32_t* in, const uint32_t* upper,
430*b2055c35SXin Li                                 int num_pixels, uint32_t* out) {
431*b2055c35SXin Li   int i;
432*b2055c35SXin Li   uint16x8_t L = vmovl_u8(LOAD_U32_AS_U8(out[-1]));
433*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
434*b2055c35SXin Li     // load four pixels of source
435*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
436*b2055c35SXin Li     // precompute the difference T - TL once for all, stored as s16
437*b2055c35SXin Li     const uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]);
438*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);
439*b2055c35SXin Li     const int16x8_t diff_lo =
440*b2055c35SXin Li         vreinterpretq_s16_u16(vsubl_u8(vget_low_u8(T), vget_low_u8(TL)));
441*b2055c35SXin Li     const int16x8_t diff_hi =
442*b2055c35SXin Li         vreinterpretq_s16_u16(vsubl_u8(vget_high_u8(T), vget_high_u8(TL)));
443*b2055c35SXin Li     // loop over the four reconstructed pixels
444*b2055c35SXin Li     DO_PRED12(diff_lo, 0);
445*b2055c35SXin Li     DO_PRED12(diff_lo, 1);
446*b2055c35SXin Li     DO_PRED12(diff_hi, 2);
447*b2055c35SXin Li     DO_PRED12(diff_hi, 3);
448*b2055c35SXin Li   }
449*b2055c35SXin Li   VP8LPredictorsAdd_C[12](in + i, upper + i, num_pixels - i, out + i);
450*b2055c35SXin Li }
451*b2055c35SXin Li #undef DO_PRED12
452*b2055c35SXin Li 
453*b2055c35SXin Li // Predictor13: ClampedAddSubtractHalf
454*b2055c35SXin Li #define DO_PRED13(LANE, LOW_OR_HI) do {                                        \
455*b2055c35SXin Li   const uint8x16_t avg = vhaddq_u8(L, T);                                      \
456*b2055c35SXin Li   const uint8x16_t cmp = vcgtq_u8(TL, avg);                                    \
457*b2055c35SXin Li   const uint8x16_t TL_1 = vaddq_u8(TL, cmp);                                   \
458*b2055c35SXin Li   /* Compute half of the difference between avg and TL'. */                    \
459*b2055c35SXin Li   const int8x8_t diff_avg =                                                    \
460*b2055c35SXin Li       vreinterpret_s8_u8(LOW_OR_HI(vhsubq_u8(avg, TL_1)));                     \
461*b2055c35SXin Li   /* Compute the sum with avg and saturate. */                                 \
462*b2055c35SXin Li   const int16x8_t avg_16 = vreinterpretq_s16_u16(vmovl_u8(LOW_OR_HI(avg)));    \
463*b2055c35SXin Li   const uint8x8_t delta = vqmovun_s16(vaddw_s8(avg_16, diff_avg));             \
464*b2055c35SXin Li   const uint8x8_t res = vadd_u8(LOW_OR_HI(src), delta);                        \
465*b2055c35SXin Li   const uint8x16_t res2 = vcombine_u8(res, res);                               \
466*b2055c35SXin Li   vst1_lane_u32(&out[i + (LANE)], vreinterpret_u32_u8(res), (LANE) & 1);       \
467*b2055c35SXin Li   L = ROTATE32_LEFT(res2);                                                     \
468*b2055c35SXin Li } while (0)
469*b2055c35SXin Li 
PredictorAdd13_NEON(const uint32_t * in,const uint32_t * upper,int num_pixels,uint32_t * out)470*b2055c35SXin Li static void PredictorAdd13_NEON(const uint32_t* in, const uint32_t* upper,
471*b2055c35SXin Li                                 int num_pixels, uint32_t* out) {
472*b2055c35SXin Li   int i;
473*b2055c35SXin Li   uint8x16_t L = LOADQ_U32_AS_U8(out[-1]);
474*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
475*b2055c35SXin Li     const uint8x16_t src = LOADQ_U32P_AS_U8(&in[i]);
476*b2055c35SXin Li     const uint8x16_t T = LOADQ_U32P_AS_U8(&upper[i]);
477*b2055c35SXin Li     const uint8x16_t TL = LOADQ_U32P_AS_U8(&upper[i - 1]);
478*b2055c35SXin Li     DO_PRED13(0, vget_low_u8);
479*b2055c35SXin Li     DO_PRED13(1, vget_low_u8);
480*b2055c35SXin Li     DO_PRED13(2, vget_high_u8);
481*b2055c35SXin Li     DO_PRED13(3, vget_high_u8);
482*b2055c35SXin Li   }
483*b2055c35SXin Li   VP8LPredictorsAdd_C[13](in + i, upper + i, num_pixels - i, out + i);
484*b2055c35SXin Li }
485*b2055c35SXin Li #undef DO_PRED13
486*b2055c35SXin Li 
487*b2055c35SXin Li #undef LOAD_U32_AS_U8
488*b2055c35SXin Li #undef LOAD_U32P_AS_U8
489*b2055c35SXin Li #undef LOADQ_U32_AS_U8
490*b2055c35SXin Li #undef LOADQ_U32P_AS_U8
491*b2055c35SXin Li #undef GET_U8_AS_U32
492*b2055c35SXin Li #undef GETQ_U8_AS_U32
493*b2055c35SXin Li #undef STOREQ_U8_AS_U32P
494*b2055c35SXin Li #undef ROTATE32_LEFT
495*b2055c35SXin Li 
496*b2055c35SXin Li //------------------------------------------------------------------------------
497*b2055c35SXin Li // Subtract-Green Transform
498*b2055c35SXin Li 
499*b2055c35SXin Li // vtbl?_u8 are marked unavailable for iOS arm64 with Xcode < 6.3, use
500*b2055c35SXin Li // non-standard versions there.
501*b2055c35SXin Li #if defined(__APPLE__) && WEBP_AARCH64 && \
502*b2055c35SXin Li     defined(__apple_build_version__) && (__apple_build_version__< 6020037)
503*b2055c35SXin Li #define USE_VTBLQ
504*b2055c35SXin Li #endif
505*b2055c35SXin Li 
506*b2055c35SXin Li #ifdef USE_VTBLQ
507*b2055c35SXin Li // 255 = byte will be zeroed
508*b2055c35SXin Li static const uint8_t kGreenShuffle[16] = {
509*b2055c35SXin Li   1, 255, 1, 255, 5, 255, 5, 255, 9, 255, 9, 255, 13, 255, 13, 255
510*b2055c35SXin Li };
511*b2055c35SXin Li 
DoGreenShuffle_NEON(const uint8x16_t argb,const uint8x16_t shuffle)512*b2055c35SXin Li static WEBP_INLINE uint8x16_t DoGreenShuffle_NEON(const uint8x16_t argb,
513*b2055c35SXin Li                                                   const uint8x16_t shuffle) {
514*b2055c35SXin Li   return vcombine_u8(vtbl1q_u8(argb, vget_low_u8(shuffle)),
515*b2055c35SXin Li                      vtbl1q_u8(argb, vget_high_u8(shuffle)));
516*b2055c35SXin Li }
517*b2055c35SXin Li #else  // !USE_VTBLQ
518*b2055c35SXin Li // 255 = byte will be zeroed
519*b2055c35SXin Li static const uint8_t kGreenShuffle[8] = { 1, 255, 1, 255, 5, 255, 5, 255  };
520*b2055c35SXin Li 
DoGreenShuffle_NEON(const uint8x16_t argb,const uint8x8_t shuffle)521*b2055c35SXin Li static WEBP_INLINE uint8x16_t DoGreenShuffle_NEON(const uint8x16_t argb,
522*b2055c35SXin Li                                                   const uint8x8_t shuffle) {
523*b2055c35SXin Li   return vcombine_u8(vtbl1_u8(vget_low_u8(argb), shuffle),
524*b2055c35SXin Li                      vtbl1_u8(vget_high_u8(argb), shuffle));
525*b2055c35SXin Li }
526*b2055c35SXin Li #endif  // USE_VTBLQ
527*b2055c35SXin Li 
AddGreenToBlueAndRed_NEON(const uint32_t * src,int num_pixels,uint32_t * dst)528*b2055c35SXin Li static void AddGreenToBlueAndRed_NEON(const uint32_t* src, int num_pixels,
529*b2055c35SXin Li                                       uint32_t* dst) {
530*b2055c35SXin Li   const uint32_t* const end = src + (num_pixels & ~3);
531*b2055c35SXin Li #ifdef USE_VTBLQ
532*b2055c35SXin Li   const uint8x16_t shuffle = vld1q_u8(kGreenShuffle);
533*b2055c35SXin Li #else
534*b2055c35SXin Li   const uint8x8_t shuffle = vld1_u8(kGreenShuffle);
535*b2055c35SXin Li #endif
536*b2055c35SXin Li   for (; src < end; src += 4, dst += 4) {
537*b2055c35SXin Li     const uint8x16_t argb = vld1q_u8((const uint8_t*)src);
538*b2055c35SXin Li     const uint8x16_t greens = DoGreenShuffle_NEON(argb, shuffle);
539*b2055c35SXin Li     vst1q_u8((uint8_t*)dst, vaddq_u8(argb, greens));
540*b2055c35SXin Li   }
541*b2055c35SXin Li   // fallthrough and finish off with plain-C
542*b2055c35SXin Li   VP8LAddGreenToBlueAndRed_C(src, num_pixels & 3, dst);
543*b2055c35SXin Li }
544*b2055c35SXin Li 
545*b2055c35SXin Li //------------------------------------------------------------------------------
546*b2055c35SXin Li // Color Transform
547*b2055c35SXin Li 
TransformColorInverse_NEON(const VP8LMultipliers * const m,const uint32_t * const src,int num_pixels,uint32_t * dst)548*b2055c35SXin Li static void TransformColorInverse_NEON(const VP8LMultipliers* const m,
549*b2055c35SXin Li                                        const uint32_t* const src,
550*b2055c35SXin Li                                        int num_pixels, uint32_t* dst) {
551*b2055c35SXin Li // sign-extended multiplying constants, pre-shifted by 6.
552*b2055c35SXin Li #define CST(X)  (((int16_t)(m->X << 8)) >> 6)
553*b2055c35SXin Li   const int16_t rb[8] = {
554*b2055c35SXin Li     CST(green_to_blue_), CST(green_to_red_),
555*b2055c35SXin Li     CST(green_to_blue_), CST(green_to_red_),
556*b2055c35SXin Li     CST(green_to_blue_), CST(green_to_red_),
557*b2055c35SXin Li     CST(green_to_blue_), CST(green_to_red_)
558*b2055c35SXin Li   };
559*b2055c35SXin Li   const int16x8_t mults_rb = vld1q_s16(rb);
560*b2055c35SXin Li   const int16_t b2[8] = {
561*b2055c35SXin Li     0, CST(red_to_blue_), 0, CST(red_to_blue_),
562*b2055c35SXin Li     0, CST(red_to_blue_), 0, CST(red_to_blue_),
563*b2055c35SXin Li   };
564*b2055c35SXin Li   const int16x8_t mults_b2 = vld1q_s16(b2);
565*b2055c35SXin Li #undef CST
566*b2055c35SXin Li #ifdef USE_VTBLQ
567*b2055c35SXin Li   static const uint8_t kg0g0[16] = {
568*b2055c35SXin Li     255, 1, 255, 1, 255, 5, 255, 5, 255, 9, 255, 9, 255, 13, 255, 13
569*b2055c35SXin Li   };
570*b2055c35SXin Li   const uint8x16_t shuffle = vld1q_u8(kg0g0);
571*b2055c35SXin Li #else
572*b2055c35SXin Li   static const uint8_t k0g0g[8] = { 255, 1, 255, 1, 255, 5, 255, 5 };
573*b2055c35SXin Li   const uint8x8_t shuffle = vld1_u8(k0g0g);
574*b2055c35SXin Li #endif
575*b2055c35SXin Li   const uint32x4_t mask_ag = vdupq_n_u32(0xff00ff00u);
576*b2055c35SXin Li   int i;
577*b2055c35SXin Li   for (i = 0; i + 4 <= num_pixels; i += 4) {
578*b2055c35SXin Li     const uint8x16_t in = vld1q_u8((const uint8_t*)(src + i));
579*b2055c35SXin Li     const uint32x4_t a0g0 = vandq_u32(vreinterpretq_u32_u8(in), mask_ag);
580*b2055c35SXin Li     // 0 g 0 g
581*b2055c35SXin Li     const uint8x16_t greens = DoGreenShuffle_NEON(in, shuffle);
582*b2055c35SXin Li     // x dr  x db1
583*b2055c35SXin Li     const int16x8_t A = vqdmulhq_s16(vreinterpretq_s16_u8(greens), mults_rb);
584*b2055c35SXin Li     // x r'  x   b'
585*b2055c35SXin Li     const int8x16_t B = vaddq_s8(vreinterpretq_s8_u8(in),
586*b2055c35SXin Li                                  vreinterpretq_s8_s16(A));
587*b2055c35SXin Li     // r' 0   b' 0
588*b2055c35SXin Li     const int16x8_t C = vshlq_n_s16(vreinterpretq_s16_s8(B), 8);
589*b2055c35SXin Li     // x db2  0  0
590*b2055c35SXin Li     const int16x8_t D = vqdmulhq_s16(C, mults_b2);
591*b2055c35SXin Li     // 0  x db2  0
592*b2055c35SXin Li     const uint32x4_t E = vshrq_n_u32(vreinterpretq_u32_s16(D), 8);
593*b2055c35SXin Li     // r' x  b'' 0
594*b2055c35SXin Li     const int8x16_t F = vaddq_s8(vreinterpretq_s8_u32(E),
595*b2055c35SXin Li                                  vreinterpretq_s8_s16(C));
596*b2055c35SXin Li     // 0  r'  0  b''
597*b2055c35SXin Li     const uint16x8_t G = vshrq_n_u16(vreinterpretq_u16_s8(F), 8);
598*b2055c35SXin Li     const uint32x4_t out = vorrq_u32(vreinterpretq_u32_u16(G), a0g0);
599*b2055c35SXin Li     vst1q_u32(dst + i, out);
600*b2055c35SXin Li   }
601*b2055c35SXin Li   // Fall-back to C-version for left-overs.
602*b2055c35SXin Li   VP8LTransformColorInverse_C(m, src + i, num_pixels - i, dst + i);
603*b2055c35SXin Li }
604*b2055c35SXin Li 
605*b2055c35SXin Li #undef USE_VTBLQ
606*b2055c35SXin Li 
607*b2055c35SXin Li //------------------------------------------------------------------------------
608*b2055c35SXin Li // Entry point
609*b2055c35SXin Li 
610*b2055c35SXin Li extern void VP8LDspInitNEON(void);
611*b2055c35SXin Li 
VP8LDspInitNEON(void)612*b2055c35SXin Li WEBP_TSAN_IGNORE_FUNCTION void VP8LDspInitNEON(void) {
613*b2055c35SXin Li   VP8LPredictors[5] = Predictor5_NEON;
614*b2055c35SXin Li   VP8LPredictors[6] = Predictor6_NEON;
615*b2055c35SXin Li   VP8LPredictors[7] = Predictor7_NEON;
616*b2055c35SXin Li   VP8LPredictors[13] = Predictor13_NEON;
617*b2055c35SXin Li 
618*b2055c35SXin Li   VP8LPredictorsAdd[0] = PredictorAdd0_NEON;
619*b2055c35SXin Li   VP8LPredictorsAdd[1] = PredictorAdd1_NEON;
620*b2055c35SXin Li   VP8LPredictorsAdd[2] = PredictorAdd2_NEON;
621*b2055c35SXin Li   VP8LPredictorsAdd[3] = PredictorAdd3_NEON;
622*b2055c35SXin Li   VP8LPredictorsAdd[4] = PredictorAdd4_NEON;
623*b2055c35SXin Li   VP8LPredictorsAdd[5] = PredictorAdd5_NEON;
624*b2055c35SXin Li   VP8LPredictorsAdd[6] = PredictorAdd6_NEON;
625*b2055c35SXin Li   VP8LPredictorsAdd[7] = PredictorAdd7_NEON;
626*b2055c35SXin Li   VP8LPredictorsAdd[8] = PredictorAdd8_NEON;
627*b2055c35SXin Li   VP8LPredictorsAdd[9] = PredictorAdd9_NEON;
628*b2055c35SXin Li   VP8LPredictorsAdd[10] = PredictorAdd10_NEON;
629*b2055c35SXin Li   VP8LPredictorsAdd[11] = PredictorAdd11_NEON;
630*b2055c35SXin Li   VP8LPredictorsAdd[12] = PredictorAdd12_NEON;
631*b2055c35SXin Li   VP8LPredictorsAdd[13] = PredictorAdd13_NEON;
632*b2055c35SXin Li 
633*b2055c35SXin Li   VP8LConvertBGRAToRGBA = ConvertBGRAToRGBA_NEON;
634*b2055c35SXin Li   VP8LConvertBGRAToBGR = ConvertBGRAToBGR_NEON;
635*b2055c35SXin Li   VP8LConvertBGRAToRGB = ConvertBGRAToRGB_NEON;
636*b2055c35SXin Li 
637*b2055c35SXin Li   VP8LAddGreenToBlueAndRed = AddGreenToBlueAndRed_NEON;
638*b2055c35SXin Li   VP8LTransformColorInverse = TransformColorInverse_NEON;
639*b2055c35SXin Li }
640*b2055c35SXin Li 
641*b2055c35SXin Li #else  // !WEBP_USE_NEON
642*b2055c35SXin Li 
643*b2055c35SXin Li WEBP_DSP_INIT_STUB(VP8LDspInitNEON)
644*b2055c35SXin Li 
645*b2055c35SXin Li #endif  // WEBP_USE_NEON
646